text
stringlengths
0
128k
Error in eval(predvars, data, env): object 'V1' not found I'm trying to use the linear regression's tools: data=read.table("http://users.stat.ufl.edu/~winner/data/pgalpga2008.dat", check.names=FALSE) Then I have to select only female golfers so I did: datF <- subset(data, V3==1, select=1:2) Where V3 is the label of the third column. Now I should apply the linear regression method so I did: datF.lm=lm(V1~V2) Where V1 and V2 are the label of the other two column but I came through this error and I can't understand why. datF.lm <- lm(V1~V2, data = datF) Yes, as slava-kohut states, the lm() function requires a reference to the dataframe as one of its arguments. If you don't specify data = datF, lm() doesn't know where to look for the variables V1 and V2.
Double-acting diaphragm pump Dec. 29, 1959 c. v. Z lEG ETAL 2,918,878 DOUBLE-ACTING DIAPHRAGM PUMP Filed Dec. 5, 1957 4 Sheets-Sheet l 1Z4 INVENTOR 140 Wk 4.0a; ATTORNEY Dec. 29, 1959 c, v, z ET AL 2,918,878 DOUBLE-ACTING DIAPHRAGM PUMP Filed Dec. 5, 1957 4 sheets-sheet z ATTORNEY EAZTOR I w. 7 Dec. 29, 1959 c. v. ZIEG ETAL 2,918,878 - DOUBLE-ACTING DIAPHRAGM PUMP Filed Dec. 5, 1957 4 Sheets-Sheet 3 BY i/aqa ATTORNEY Dec. 29, 1959 c, v, z ETAL DOUBLE-ACTING DIAPHRAGM PUMP 4 Sheets-Sheet 4 Filed Dec. 5, 1957 United States Patent 7 2,918,878 D U L -AC NG DIA HRAGM U Clifford V. Zieg, Fort Wayne, Ind.,=John C.'-Lee, Mount Prospect, 111., and Robert -'D. Lambert, Fort Wayne, .Ind., .assignors, by,mes ne assignments, to Syrnington- Wayne Corporation, Salisbury, ;Md ..-a corporation of Maryland I Application December, 195."lyserialfNok700,841 1;3-.Cl.aims. .l(Cl-::10.3-1.5. "Th D lfi .elatcs t r c pro ca ing ou hl ract n pump including a pair of flexible diaphragms, -,and more particularly relates to a pur np.ofthis type which can be operated by manual power. Since pumps of most types include a connection from a prime mover to a. movable pumping-element, .a pr ob- 1cm usually exists of sealing the. space between ,a moving motion-transmitting element and the aperture ,through which it passes into the pumping chamber. ,Double acting pumps utilizing flexible diaphragms are subject to this sealing problem especially when the chambers are placed side-by-side to conserve space and to simplify the structure. An object of this invention istoprovide a simple, .com pact and economical structure for ,a double-acting diaphragm pump that does mot requi e the use .of a fluid seal about a moving element. In accordance with this invention a pair ofdiaphragms are secured to .both sides ,of a partitioning structure. These diaphragms are tightly secured at a-pre determined longitudinal spacing upon a shaft which recipro cates along a portion of its length disposed between these diaphragms within a bearing in a wall of the partitioning structure. An actuating means, for example a lever mechanism, is connected to a portion of the shaft disposed outside of the diaphragms for recipro cating the shaft along its longitudinal axis to impart a pumping action to the diaphragrns. A.clear a nce is permitted to exist .between the shaft and this bearing which permits a slight flow of fluid between the chambers, which flow-is not sufficiently great to interfere with adequate pumping efiiciency. Since this leakage occurs entirely within elements of the pump sealed within the diaphragms, partitioning structure, and shaft, there is no necessity for fluid seals to be used about any moving part. Novel features and advantages of the present-invention will become apparent toone skilled in the art from a reading of the following description in conjunction with the accompanying drawings wherein similar reference characters refer to similar parts and in which: Fig. 1 is a front view in elevation, partially in cross section of one embodiment of this invention; Fig. 2 is a cross-sectional view taken through Fig. 1 along the line '2 2 and looking in the direction of the arrows; Fig. 3 is a cross-sectional view taken through .Fig. .1 along the line 3-3 and looking in the direction off the arrows; Fig. 4 is an enlarged .view of a portion of Fig. .2; Figs. 5, 6 and 7 are cross-sectional views respectively taken through Fig. 1 along 'the lines 5-5, .6.6, and 7-7, looking in the direction of the arrows; Fig. 8 is a view in elevation partially in cross section of another embodiment of this invention; Fig. 9- is a plan view of a portion of the embodiment shown in Fig.8; Fig. 10 is a side view in elevation of the portion shown inxFig. 9; of the. driving *bearing hub 50 in cover plate Fig. .1 1 is' an enlarged plan \view of a portion .of .the embodiment shown in Figs. .8 and 9; Fig. 12 is .a cross-sectional' view taken through Fig. 11 along the line 12:12.an.d looking in the.direction.of 14. A, pair. of.fiexible.diaphragms 20, :for; example, made of relatively strong'rubber including suitable reinforce. ments (later described inz 'detail) .aresealed :to :both sides ofypartitioningtstructure 12 by means of ring 1incorporatingelements 22 and .23. ,Ring incorporating elements. 22. and23are, ifor, example, -a;pair of. dishedicovertplates 22 ;and. 23 .whichare clamped, for example, :to flanges or rim. elements24 .which, for example, extend from :the ends of wall'14.at obtuse angles relative to-the substantially flat .central I portion 7 of 1 wall -1-4 As shown :in. 'Figs. land 2, wall .14 is substantially circular and, rims .or flanges ,24 extend :from its g periphery to ,form substantially .circular flanges extending therefrom with .a .U- shaped or A l-shaped space 26 formed between iflanges 24. The peripheral edges 28 of diaphragms :20 are enlarged .to .lock. them within corresponding recesses 330 in the 'peripheral edges-of covers 22 and .123 :to securely seal .peripheral .edges 28 of diaphragms20 sides of the central partitioning structure 12. Covers 22 and 23-.are maintained -in engagement with the central partitioning structure 12, for exan1ple,,by means of screwandboltfasteners 32. Diaphragms 20 incorporate, for example, circular J inserts 34 made, forexample, of plate steel for-stiffening their central portions to facilitate transmission'of .a recipro catingmotion thereto. The outer .regions 10f diaphragms20which are subject to considerable flexing are strengthened by incorporating, for example, a .web .of fabric 36 made, for example, of woven nylon. The central portions of diaphragms 20 are'secured in a fluid tight .manner at ,a pre determined longitudinal spacing upon a shaft 38 which recipro cates through. a portion of its length disposed' means of a screw and washer 42; and the diaphragm .20 and insert 34 .atthe intermediate position along shaft 38 are fastened within groove 40 by, for example, the threaded engagement of spacer shaft portion .44 disposed between diaphragms .20 with a threaded projection -46 portion 48 of the shaft 38, which driving portion 48 is disposed outside-of the space between diaphragm'20. The diameter of shaft 38 and the aperture in bearing 16 are manufactured to standard tolerances to provide, a for example, between 0.001 and 0.003 inch. The slight leakage of fluid through this clear ance is insufiicient to interfere with adequaute pumping. action. The driving portion 48 of .shaft 38 extends through a 23 to provide a means for connecting an actuating mechanism, for example, a hand operated lever 52 thereto. The frictionalresistance between shaft 38 and bearing hub 50 is minimized, for example, by means of a nylon bearing insert 54 mounted within hub 50. Lever mechanism 52 includes, for ex against *both' ample, a hand operated lever 56 connected to react against partitioning structure 12 through cover 23, for example, by means of connecting linkage 58 rotatably connected at one end thereof by means of pin 60 to cover 23. The other end of connecting linkage 58 is rotatably connected to a lower extension of handle 56 by means of a pin 62. An intermediate portion of handle 64, for example, disposed at short distance above pivot 62 is rotatably connected to driving portion 48 of shaft 38 by means of a pin 66. Handle 52 is conveniently fabricated in a channel form with the various enumerated connections disposed within the arms of the channel. Aperture d extensions 68 and 70 are respectively provided upon handle 56 and casing 23 to permit handle 56 to be attached to casing 23 by means of a pin inserted therethrough when the pump is inoperative. This, for example, provides a convenient means for carrying the disconnected pump or for locking it to prevent unauthorized use. A convenient material for fabricating the partitioning structure 12 and cover plates 22 and 23 is, for example, an aluminum alloy suitable for die casting. This facilitates provision of the various functional shapes in their proper relationship by an inexpensive manufacturing method. All non-moving parts including a cover plate 72 upon a discharge chamber 74 and an inlet conduit unit 76, connected to dual inlet chambers 78, may also be conveniently fabricated of this same aluminum alloy. However, shaft 38, connecting link 58 and handle 52 are, for example, made of forms of steel to provide adequate strength. Shaft 38 is, for example, made of stainless steel to resist corrosion by the fluids being pumped, and connecting link 58 and handle 52 are, for example, made of carbon steel suitably hardened at parts subject to wear, and coated to resist atmospheric corrosion, for example, by cadmium plating. As shown in Fig. 1, pairs of separating walls 80 and 82 are respectively disposed between rim flanges 24 extending from central wall 14 or partitioning structure 12, to segregate respective discharge chamber 74 and dual inlet chamber 78. An intermediate separating wall 84 is disposed between separating walls 82 to provide a dual inlet chamber 78. These separating walls are, for example, conveniently formed during the die-casting operation which forms partitioning structure 12. Discharge chamber 74 is, for example, connected to pumping chambers 86 and 88 formed between diaphragms 20 and opposite sides of partitioning structure 12, by means of aligned passageways 90 extending through flanges 24. A strip of resilient valve material 92 is disposed within the angular space between flanges 24 and within discharge chamber 74 with its central portion 94 anchored to the end of central wall 14, for example, by a forced fit engagement of projections 96 of central wall 14 which extend into discharge chamber 74. The outer portions 98 of resilient valve element 92 are thereby maintained in firm resilient engagement with the inner surface of flanges 24 in position to cover passageway 90. The outer portions 98, therefore, act as discharge check valves which permit a flow of fluid from pumping chambers 86 and 88 and prevent a backflow from discharge chamber 74 thereto. Outer portions 28 are stiffened, for example, by reinforcing ribs 100 molded, for example, in its non-sealing upper surface. Discharge chamber 74 also includes an open end or discharge connection 75 which is internally threaded, for example, for connection to discharge piping (not shown). Cover plate 72 includes, for example, a bleeder valve means 102 which permits air to enter discharge chamber 74 when the pressure therein drops below atmospheric pressure. Bleeder valve 102 includes, for example, an elongated aperture 104 within which a resilient closure element 106, made of rubber for example, is inserted. Resilient closure element 106 incorporates, for example, a stem 108 inserted through aperture 104 and a flexible head 110 which is large enough to completely cover aperture 104. Head 110 is disposed substantially perpendicularly to stem 108 and lies within discharge chamber 74. Stem 108 is recessed at 112 adjacent head 110 to provide a shoulder 114 for locking closure element 106 within aperture 104 between head 110 and itself. The outer portion or skirt 116 of head 110 is inclined at an angle towards plate 72 to lock element 106 in resilient contact with plate 72. Resilient closure element 106 is constructed and arranged to permit the head to be flexed away from the wall to facilitate flow of air through aperture 104 into discharge chamber 74 when the pressure within chamber 74 falls a slight pre determined amount below atmospheric pressure. This flexing is facilitated by a cutout 118 extending axially through the closure element from head 110 a distance into stem 108 along its axis. Inlet conduit unit 76 is connected to dual inlet chamber 78 by means of respective flanges 120 and 122 connected together, for example, by cap screws 124. A specially shaped and formed strip of resilient material 126 is disposed between flanges 120 and 122 to simultaneously provide a sealing gasket and for providing inlet check valve elements. This gasket is formed by the frame portion 128 of strip 126, and the inlet check valve elements are formed by a pair of valve flaps 130 which extend from a central web 132 which connects the sides of frame 128. Flaps 130 are stiffened by ribs 133 formed upon their upper surfaces as shown in Fig. 6. Strip 126 is, for example, made of rubber of adequate stiflness to act as valve elements. Inlet conduit 76 includes a pair of sloped rims or extensions 134 extending a short distance into inlet chambers 78 as shown in Figs. 1 and 5. These sloped rims 134 are constructed and arranged to extend within the inside edge of frame 128 of strip 126 and lie under flaps 130. When frame 128 is securely squeezed between flanges 120 and 122, flaps 130 are maintained in firm resilient contact with the upper surfaces 136 of sloped rims 134 to provide separate check valve elements for dual inlet chambers 78. The lower portion of inlet conduit 76 includes a tubular extension 138 which is internally threaded at 140 and externally threaded at 142 for respective connection, for example, to a suction standpipe and supply tank (not shown). Slot 141 and small hole 143 in flange 120 admit air into the tank, for example, through annular space 145 between threads 140 and 142 to permit withdrawal of fluid through threaded hole 140. A strainer 144, for example, made of wire screen is inserted within the central flow passage of conduit 76 to prevent large particles from entering the pump. As shown in Figs. 1, 2 and 7, flow passageways from dual inlet chamber 78 into pumping chambers 86 and 88 are, for example, provided by a passageway 146 extending from one inlet chamber 78 to pumping chamber 88 and a second inlet passageway 148 extending through flange rim 24 into pumping chamber 86. Passageways 146 and 148 are staggered from each other to provide separate access means to chambers 86 and 88. Cover plates 22 and 23 are herein represented as substantially continuous plates with the exception of apertures 150 therein which permit air to enter into the space between diaphragms 20 and plates 22 and 23. This flow of air prevents a vacuum from being formed between the diaphragms and the cover plates when the pump is actuated. Cover plates 22 and 23, however, might be formed of ribbed or perforated elements so long as their outer peripheries are substantially unbroken to seal the peripheries of diaphragms 20 against partitioning structure 12. Cover plate 22 also includes a recessed protuberance formed at 152 to receive the head of screw 42 to permit full flexure of the diaphragm up to the limit provided by the cover plate. This limit of travel for the diaphragms provided by cover plates 22 and 23 prevents them LfIOIIl a being .overstressed during .the pumping operation. In-Figs. 8-10..is shown'a-discharge chamber;cover assembly -154 which is constructed and arranged to include elements and features which supplement the features of cover plate 72 shown in Figs. 2 and 4. Discharge cover assembly 154 includes a superimposed outer shell 160 which forms a horizontally disposed pocket or housing 162 for receiving a discharge nozzle 164 which,.for example, is attached to a hose 166 connected to discharge connection 75a-of pump 10a. The lower flange 168 of assembly 154 extends past discharge chamber 74a at 170. An aperture 172'is provided rin-extension170 for receiving a lug 174 formed in the lower surface of nozzle 164. Aperture 172 is, for example, rectangular. This permit sa nozzle 164 v to be conveniently housed within pocket 162 when the'pump is not in use. Assembly 154 is also constructed and arranged to receive a controllable bleeder-valve 103 (later described in detail). Discharge chamber 74a-continues upwardly a short distance 'above walls 24a into a space provided between the lower wall .178 of pocket 162 and the vertically dis- POSCdzWflll .180 which contains controllable bleeder valve 103. This makes valve 103 accessible for convenient manipulation (later described in detail) and places it in a substantially vertically disposed wall with a projecting ridge .182 extending outwardly from wall 180 to shield aperture200 in wall 180, for-example, from rain. A small projection 186 extends from wall 180 a short v distance below and to the side of valve 103 to help locate its on and off positions. .Details of valve 103 are shown in Figs. 11 and 12. Valve 103 is made of a resilient material, for example, 30-40 durometer-rubber of the Buna N variety. Valve 103-includes a flexible head 188, a stern 190, and a'knob or enlarged end 192. Stem 190 isinsertedthrough aperture 200in wall 180 in substantially the samemannerdescribed in conjunction with valve 102 shown in Fig. 4. Skirt 188 includes, for example, a relatively thick section 194 disposed opposite nose 196 of control knob 192. This thicker section 194 helps maintain a seal between the lower surface 198 of nose 19.6 and the ,bleederslot later described in conjunction withlFig. 10. In Fig. 10 are shown details of the aperture 200 formed in .wall 180 of assembly 72a. Aperture 200includes a single. slot 202 similarto dual slots 104 shown in Fig. 3. Extension or nose 196 of bleeder valve 103 is constructed and arranged to cover slot extension 202 when it is rotated in a position to directly overlie it. .However, when knob 192 is turned at the position shown in Fig. .8, extension 196 extends downward leaving slot 202 unobstructed. This permits the zbleeder valve to be optionally turned on or off to permit optional admission to or exclusion from a flow of air into discharge chamber 74a between each discharge stroke. Bleeder valve 103 in the on position allows air to enter discharge chamber 74a through slot 202 to prevent siphoning whenever'the pump outlet is situated below the level of the pump. This vacuum-breaking feature is required, for example, if hose nozzle 164 attached to flexible hose 166 were dropped on the floor while pumping liquid from a barrel or overhead tank (not shown). However, the siphoning-preventing or vacuum-breaking feature is not required, for example, when pumping from an underground tank or when a discharge spout is directly connected to the pump outlet. In such instances the bleeder .valve is turned to the off position by turning knob 192 to place its nose or extension 196 directly over slot 202. Furthermore, the off position is advantageous for the purpose of excluding air from the chamber 74a when more volatile fluids are being pumped. In Fig. 13 a modified cover assembly 156 is shown which includes a controllable bleeder valve 105 disposed in the horizontal lower flange 168b in a manner similar to that. shown in Fig. 4. Assembly 156 includes an upper roof .or shell .6 Q r tain ng e t. -p 20 ahnzzle .(-not shown) in a manner similar to that indicated'in Fig. -18 with-the lower projection 174 of the;nozzle1(not;shown disposed within aperture or slot 172k. Shell 160b, accordingly, performs the dual functions of shielding thebleeder valve and aperture from rain and of providing a housing for the nozzle. Roof 16% is supported by side walls 1561b, mounted upon flange 168b. Projections 182b ;an d 186b which are similar to projections ;182; and 18.6 in Fig. 10 extend from flange168b adjacenttheknob192b of valve 105. The structure of cover assembly 156, shown in Fig. 13 is slightly .simpler than thatzof cover assembly 154, shown in Figs. 8-10; .t;the:positioning of the control valve ,105 of cover assembly ,156 ,under roof 16% makes it slightly less convliifint 10.-operate than the more exposed control .valve 103. Operation When lever 156-is ractuated back :and forth, it recipro cates diaphragms 20 with'it:to alternately'compress and expand -pumping-chambers86 and 88 against partitioning structure 12. .In the position shown in Fig. 2,;pumping chamber 86 has been expanded to substantially its full capacity which opens the check :valve flap element to. permit a flow of fluidtobedrawnthrough-inlet chamber 78 and passageway 1'485into pumping chamber '86. At the same time, pumping chamber88 has been com pressed to its minimum volume which discharges a 'flow of fluid through discharge check-valve element 98 to discharge chamber'74. From discharge chamber 74, the fluid is carried away through -a discharge conduit not shown, connected to the open end75 of discharge'74. [At the same time, a slight flow of fluid leaks through the clear ance between shaft 38 and bearing 16 into expanded pump chamber 86. However, this does not interfere materially with the adequate pumping operation because this clear ance-or leakage is insignificant in comparison to the relatively 'large flow of fluid through discharge passageway 90. This leakage is'maintained within sealed portions of the pump and it mingles -with the fluid being drawn into:pumping chamber 86 through inletpassageway 148. This leakage is also insignificant in comparison with the relatively large amount of fluid drawn into the inlet chamber and, therefore, does not materially decrease the amount of fluid :drawn into the pump. .Since this leakage is contained entirely-within portions of the pump sealed .betweendiaphragms '20, partitioning structure 12 and shaft 38, there is no-necessity'to provide a fluid-tight seal about any moving portions of the pump. This'permits an embodiment of this invention to be-constructed in an unusually economical manner. When handleJ56 is moved to the limit of travel indicated in broken outline, :the diaphragms 20-move'to the positions shown in broken outline to reverse the conditionslof pump chambers 86 and 88. Fluid is then'discharged from pumping chamber 86 into discharge chamber 74 and drawn from respective inlet chamber 78 into pumping chamber 88 through an inlet passageway 146. A steady stream of fluid is, therefore, 'drawn into and discharged from the pump with each stroke of handle '56. ,If many time,'the:pressure within discharge chamber 74 falls below atmospheric pressure, bleeder valve 102 permits an to enter chamber 74 Which'prevents a reduced pressure from existing in chamber 74 which might interfere with proper operation of. discharge check valve flaps or outer portions 98. The controllable bleeder valves 103 and 1 05'shown-in Figs. 813 can be optionally turned -off or on! When these valves are turned .on,- noses or extensions 196 and 19617 are positioned-remote'from keyhole slots 202 and 20212 which freestheseslotsfor the passage of aire when the pressure in the discharge chamberjdrops below atmospheric pressure. These bleeder valves "are, accordingly, , placed .on the "on,position when the pumps .1 are "com a nected to fixed piping to prevent any siphoning from the discharge lines when the pressure in the discharge chamber falls below atmospheric pressure between discharge strokes. However, when the pump discharges to the atmosphere through a hose 166 and nozzle 164 this siphoning problem does not exist, and the bleeder valve 103 or 105 is turned off by positioning nose 196 or 19Gb respectively over keyhole slots 202 or 2112b. This prevents any air from entering the discharge chamber which sealing is advantageous for handling certain type of liquids, for example, light and volatile liquids. What is claimed is: l. A double-acting diaphragm pump for fluids comprising a partitioning structure including a substantially fiat central plate and a pair of flanges extending away from each other about its periphery at obtuse angles relative to said central plate, a bearing means disposed in said central plate, a pair of diaphragms disposed on both sides of said partitioning structure, cover elements including ring means sealing the peripheries of said diaphragms to outer surface of said flanges to form a pumping chamber of variable volume on both sides of said partitioning structure, said flanges being cut out to provide inlet and outlet passageways into said chambers, said inlet passageways being provided with inlet check valves, a recipro cating shaft means, said diaphragms being secured in a fluid-tight manner at a pre determined longitudinal spacing upon said shaft means, said shaft means being inserted to recipro cate within said bearing means along a portion of said shaft disposed between said diaphragms, a clear ance existing between said shaft and said bearing means which permits a slight flow of fluid between said chambers which is not sufliciently great to interfere with adequate pumping efliciency, a lever means mounted upon the outside of one of said cover means and operatively connected to a portion of said shaft disposed outside of said cover means for recipro cating said shaft along its longitudinal axis whereby a pumping action relative to said chambers is imparted to said diaphragms, separating walls being disposed within the substantially angular space formed between adjacent portions of said flanges on both sides of said inlet and outlet passageways to provide inlet and outlet chambers connected thereto, and the end of said substantially flat central plate being disposed substantially midway between a pair of said passageways through said flanges, a strip of resilient valve material being disposed within the angular space formed between said flanges within one of said chambers with the central portion of said strip disposed at the junction of said flanges and its outer portions disposed in resilient contact with one pair of said cutouts in said flanges, and securing means attached to said end of said central plate and extending through said central portion of said strip for anchoring said strip within said chamber with its outer portions maintained in firm resilient contact with said flanges to act as discharge check valves relative to said cutouts whereby said one chamber is constituted as a discharge chamber. 2. A pump as set forth in claim 1 wherein said securing means includes projections extending from said end of said central wall into said angular space between said flanges. 3. A pump as set forth in claim 1 wherein an additional separating wall is positioned between a pair of said separating walls to provide dual inlet chambers therebetween, and one of said cutouts is disposed Within each of said duel chambers to provide separate passageways into each of said pumping chambers. 4. A pump as set forth in claim 3 wherein an inlet conduit means is connected to said dual inlet chambers by means of a pair of flanges connected respectively thereto, said conduit means including a pair of inlet passageways aligned with said dual inlet chambers, a pair of sloped rims extend from said conduit into said dual chambers, a strip of resilient material including a frame and a pair of flaps is disposed between said conduit means and said dual inlet chambers with said frame held between said flanges to simultaneously act as a sealing gasket and to maintain said flaps in resilient contact with said sloped rims to provide inlet check valve elements for said dual inlet chambers. 5. A pump as set forth in claim 1 wherein a bleeder valve means is incorporated in a wall of said discharge chamber for allowing air to enter said chamber when the pressure therein drops below atmospheric pressure, said wall being secured to the outer edges of said flanges and said separating walls of said outlet chamber, said wall of said discharge chamber cut out in the form of an elongated aperture, a resilient closure element being mounted within said elongated aperture, said resilient closure element including a stem inserted within said aperture and a flexible head constructed and arranged to cover said aperture and disposed substantially perpendicularly to said shaft within said discharge chamber, said stem being recessed adjacent said head to form an enlarged end which provides a shoulder for locking said element within said aperture between said head and said shoulder, the outer portion of said head adjacent said wall being inclined at an angle towards said wall to maintain the outer edge of said head locked in resilient contact with said wall, an external portion of said elongated aperture extending past said enlarged end of said stem, and said resilient closure element being constructed and arranged to permit said head to be flexed away from said wall to facilitate a flow of air through said portion of said aperture into said discharge chamber when the pressure in said discharge chamber falls a slight pre deter-- mined amount below atmospheric pressure. 6. A pump as set forth in claim 5 wherein an axial portion of said closure element extending through said head into said stem is cut out to facilitate said flexing of said head, and the end of said stem remote from said head is conical ly shaped to facilitate insertion of said stem into said aperture. 7. A pump as set forth in claim 5 wherein the larger portion of said stem adjacent said shoulder is constructed and arranged to provide a nose for optionally covering and sealing said external portion of said elongated aperture, and said stem and main portion of said aperture are constructed and arranged to permit said bleeder valve to be manually rotated to optionally cover and uncover said external portion of said aperture. 8. A pump as set forth in claim 7 wherein a portion of the upper wall of said outlet chamber is inclined upwardly away from said outlet chamber to provide a continuation of said outlet chamber, and said bleeder valve means is positioned within said inclined wall of said continuation to provide convenient access to said bleeder valve means. 9. A pump as set forth in claim 8 wherein projections extend from said inclined wall of said continuation of said outlet chamber above and below said nose to provide a shield over said aperture and indexing projections for locating the on and ofl positions of said nose. 10. A pump as set forth in claim 7 wherein the portion of said skirt adjacent said nose is thickened to help seal said external portion of said aperture between said nose and said skirt. 11. A pump as set forth in claim 1 wherein a cover assembly is disposed upon said outlet chamber, a hose and nozzle are attached to said outlet chamber, said cover assembly including extending walls which are constructed and arranged to form a pocket for receiving the tip of said nozzle, said cover assembly including an extending flange, and said nozzle and said flange including interlocking projection and recess means for engaging said nozzle with said flange when the tip of said nozzle is disposed within said pocket. 12. A pump as set forth in claim 11 wherein a portion 10 of the wall between said cover assembly and said outlet References Cited in the file of this patent chamber is extended upwardly to provide a continuation of said outlet chamber, and a controllable bleeder valve UNITED STATES PATENTS means is disposed within said inclined wall of said con- 677,474 Russell July 2, 1901 tinuation. 5 1,087,671 Loud Feb. 17, 1914 13. A pump as set forth in claim 11 wherein said cover 1,920,014 Horton et a1. July 25, 1933 assembly includes a lower flange sealing a side of said 2,307,566 Brown Jan. 5, 1943 outlet chamber, a controllable bleeder valve is disposed in said lower flange, and said extending walls of said FOREIGN PATENTS pocket adjacent said bleeder valve are cut out to provide 10 34,296 Austria Sept. 10, 1908 access to said controllable bleeder valve. 214,210 Germany Oct. 12, 1909
// // CenterPopupViewController.swift // Xpert // // Created by Darius on 23/08/2018. // Copyright © 2018. All rights reserved. // import UIKit class CenterPopupViewController: UIViewController { let animator = CustomPopupAnimation() let backgroundView : UIView = { let view = UIView() view.backgroundColor = UIColor(white: 0.1, alpha: 0.1) return view }() let contentView : UIView = { let view = UIView() view.backgroundColor = .white view.layer.cornerRadius = 5 return view }() let titleLabel : UILabel = { let label = UILabel() label.font = UIFont(name: "Muli-Regular", size: 26) label.textColor = UIColor(red: 28/255, green: 41/255, blue: 51/255, alpha: 1) label.numberOfLines = 0 label.lineBreakMode = .byWordWrapping label.textAlignment = .center return label }() let subtitleLabel : UILabel = { let label = UILabel() label.font = UIFont(name: "Muli-Regular", size: 17) label.textColor = UIColor(red: 162/255, green: 174/255, blue: 183/255, alpha: 1) label.numberOfLines = 0 label.lineBreakMode = .byWordWrapping label.textAlignment = .center return label }() let iconImageView : UIImageView = { let imageView = UIImageView() imageView.contentMode = .scaleAspectFit return imageView }() let closeButton : SquareButton = { let button = SquareButton(title: Localization.shared.data_ok_got_it.uppercased(), image: nil, backgroundColor: UIColor.primaryButtonColor, textColor: UIColor.white) button.titleLabel?.font = UIFont(name: "Muli-ExtraBold", size: 10) return button }() init(icon : UIImage, title : String, subtitle : String) { titleLabel.text = title subtitleLabel.text = subtitle iconImageView.image = icon super.init(nibName: nil, bundle: nil) } required init?(coder aDecoder: NSCoder) { fatalError("init(coder:) has not been implemented") } override func viewDidLoad() { super.viewDidLoad() setupView() setupContent() } func setupView() { view.backgroundColor = .clear view.addSubview(backgroundView) view.addConstraintsWithFormat(format: "H:|[v0]|", views: backgroundView) view.addConstraintsWithFormat(format: "V:|[v0]|", views: backgroundView) view.addSubview(contentView) view.addConstraintsWithFormat(format: "H:|-(>=15,==15@900)-[v0(<=400)]-(>=15,==15@900)-|", views: contentView) NSLayoutConstraint(item: contentView, attribute: .centerX, relatedBy: .equal, toItem: view, attribute: .centerX, multiplier: 1, constant: 0).isActive = true NSLayoutConstraint(item: contentView, attribute: .centerY, relatedBy: .equal, toItem: view, attribute: .centerY, multiplier: 1, constant: 0).isActive = true } func setupContent() { contentView.addSubview(iconImageView) contentView.addSubview(titleLabel) contentView.addSubview(subtitleLabel) contentView.addSubview(closeButton) contentView.addConstraintsWithFormat(format: "H:[v0(95)]", views: iconImageView) contentView.addConstraintsWithFormat(format: "V:|-30-[v0(95)]-20-[v1]-25-[v2]-25-[v3(26)]-15-|", views: iconImageView, titleLabel, subtitleLabel, closeButton) NSLayoutConstraint(item: iconImageView, attribute: .centerX, relatedBy: .equal, toItem: contentView, attribute: .centerX, multiplier: 1, constant: 0).isActive = true contentView.addConstraintsWithFormat(format: "H:|-33-[v0]-33-|", views: titleLabel) contentView.addConstraintsWithFormat(format: "H:|-10-[v0]-10-|", views: subtitleLabel) contentView.addConstraintsWithFormat(format: "H:|-15-[v0]-15-|", views: closeButton) closeButton.addTarget(self, action: #selector(closeViewController), for: .touchUpInside) } var closeCompletion : (()->())? @objc func closeViewController() { self.dismiss(animated: true, completion: closeCompletion) } }
class ATM(): def __init__(self, name, atm_card_number, pin, amount): self.name = name self.atm_card_number = atm_card_number self.pin_number = pin self.amount = amount def cash_withdrawl(self, amount): self.amount = self.amount + amount print(self.amount) option = input( "Do you want to check your balance: (yes or no): ") if option == "yes": self.balance_enquiry() else: exit def balance_enquiry(self): print(self.amount) option = input( "Do you want to add more money: (yes or no): ") if option == "yes": amount = int(input("Enter amount: ")) self.cash_withdrawl(amount) else: exit name = input("Your name: ") atm_card_number = input("Your atm card number: ") pin = input("Your pin: ") amount = 0 atm = ATM(name, atm_card_number, pin, amount) option = input("Enter CW for cash withdrawl or BE for balance enquiry: ") if option == "CW": amount = int(input("Enter amount: ")) atm.cash_withdrawl(amount) elif option == "BE": atm.balance_enquiry() else: exit
<?php namespace Generator\Generators\Api; use Api\Info\IApiInfo; use Cli\Tools\CommandUtils; use Helper\ApiXsd\Schema\Api; use Nette\PhpGenerator\ClassType; use Nette\PhpGenerator\PhpNamespace; /** * Generates the ApiInfo.php file which can be found in each Api module. * Class ApiInfo * @package Generator\Api\modules */ final class ApiInfo { final public function fromXml(Api $oApi) { echo "Generating class " . $oApi->getApiNamespace() . '\\Generated\\ApiInfo' . PHP_EOL; $oNamespace = new PhpNamespace($oApi->getApiNamespace() . '\\Generated'); $oClass = new ClassType("ApiInfo"); $oClass->addImplement(IApiInfo::class); $sComment = [ 'This class is automatically generated by the build script.', 'Do not change it as your changes will be lost after the next build', 'If you need modifications please extend' ]; $oClass->setComment(join(PHP_EOL, $sComment)); $aEmail = $oApi->getContacts('TECHNICAL'); $aMethodMap = [ "getTitle" => ["return" => $oApi->getTitle()], "getDescription" => ["return" => $oApi->getDescription()], "getOrganisation" => ["return" => $oApi->getOrganisation()], "getApiDir" => ["return" => $oApi->getApi_dir()], "getEmail" => ["return" => $aEmail[0]['email']], "getServers" => ["return" => $oApi->getServers()], "getContacts" => ["return" => $oApi->getContactArray()], "getEndpointUrl" => ["return" => $oApi->getEndpoint_url()], "getDocumentationUrl" => ["return" => $oApi->getDocumentation_url()], "getAuthorisationModel" => ["return" => $oApi->getAuthorization_model()], "getCaCert" => ["return" => $oApi->getCa_cert_path()], ]; $aMethods = []; foreach ($aMethodMap as $sMethodName => $aImplementation) { $aMethods[$sMethodName] = $oClass->addMethod($sMethodName); $aMethods[$sMethodName]->setFinal(true); $aMethods[$sMethodName]->setVisibility('public'); if (is_array($aImplementation['return'])) { $aMethods[$sMethodName]->setReturnType('array'); $aBody = ['$aReturn = [];']; foreach ($aImplementation['return'] as $iRow => $aProperties) { if (is_array($aProperties)) { foreach ($aProperties as $sKey => $sValue) { $aBody[] = '$aReturn[' . $iRow . ']["' . $sKey . '"] = "' . $sValue . '";'; } } else { $aBody[] = '$aReturn[' . $iRow . '] = "' . $aProperties . '";'; } } $aBody[] = "return \$aReturn;"; $aMethods[$sMethodName]->setBody(join(PHP_EOL, $aBody)); } else { $aMethods[$sMethodName]->setReturnType('string'); $aMethods[$sMethodName]->setBody('return "' . $aImplementation['return'] . '";'); } } $oClass->setFinal(true); /* $oApi->getTitle(); <?php namespace Api\Info; interface IApiInfo { public function getTitle():string; public function getDescription():string; public function getOrganisation():string; public function getApiDir():string; public function getEmail():string; public function getServers():array; public function getContacts():array; public function getEndpointUrl():string; public function getDocumentationUrl():string; public function getAuthorisationModel():string; public function getCaCert():string; } */ $oNamespace->add($oClass); $sTemplate = (string) $oNamespace; $sFileLocation = CommandUtils::getRoot() . "/public_html/" . $oApi->getApi_dir() . "/modules/Generated/ApiInfo.php"; if (!is_dir(dirname($sFileLocation))) { echo "Crate dir $sFileLocation" . PHP_EOL; mkdir(dirname($sFileLocation)); } echo "Write file: " . $sFileLocation . PHP_EOL; file_put_contents($sFileLocation, '<?php' . PHP_EOL . $sTemplate); } }
""" Script to measure the performance """ import timeit import sys # number of iterations iterations = 1000 if len(sys.argv) > 1: iterations = int(sys.argv[1]) # code snippet to be executed only once setup = """ import pandas as pd import association_measures.frequencies as fq import association_measures.measures as am df = pd.read_csv('tests/data/brown.csv') df = fq.observed_frequencies(df) df = fq.expected_frequencies(df, observed=True) """ # code snippet whose execution time is to be measured codes = [ { 'name': 'contingency_table', 'code': 'fq.observed_frequencies(df)' }, { 'name': 'expected_frequencies', 'code': 'fq.expected_frequencies(df)' }, # asymptotic hypothesis tests { 'name': 'z_score', 'code': 'am.z_score(df)' }, { 'name': 't_score', 'code': 'am.t_score(df)' }, { 'name': 'log_likelihood', 'code': 'am.log_likelihood(df)' }, { 'name': 'simple_ll', 'code': 'am.simple_ll(df)' }, # point estimates of association strength { 'name': 'min_sensitivity', 'code': 'am.min_sensitivity(df)' }, { 'name': 'liddell', 'code': 'am.liddell(df)' }, { 'name': 'dice', 'code': 'am.dice(df)' }, { 'name': 'log_ratio', 'code': 'am.log_ratio(df)' }, # likelihood measures # ~2.5s for a ~25,000 rows on 8 threads # { # 'name': 'hypergeometric_likelihood', # 'code': 'am.hypergeometric_likelihood(df)' # }, { 'name': 'binomial_likelihood', 'code': 'am.binomial_likelihood(df)' }, # conservative estimates { 'name': 'conservative_log_ratio', 'code': 'am.conservative_log_ratio(df)' }, # ~1.5s for a ~25,000 rows on 8 threads # { # 'name': 'conservative_log_ratio_poisson', # 'code': 'am.conservative_log_ratio(df, boundary="poisson")' # }, # information theory { 'name': 'mutual_information', 'code': 'am.mutual_information(df)' }, { 'name': 'local_mutual_information', 'code': 'am.local_mutual_information(df)' }, ] for code in codes: res = timeit.timeit(setup=setup, stmt=code['code'], number=iterations) print('Calculate {func} (iterations={iter}, df_size={size}): {res}'.format( iter=iterations, size=24168, res=res, func=code['name'] ))
in theory. The original fertility of the soil has been for generations sent into the towns in the diurnal milk-can, and the poor savings of the business put into the banks of deposit to be loaned to those who build up towns, or used to construct rail roads that bring butter from Iowa to Boston, cheaper than it can be carried thither from Worcester or Franklin. Mean while, the moss-grown pasture, robbed of fertility, grows up to a tangled maze of bushes. Briers, spoon-wood, hardback, wood-waxen huckleberry, and similar usurpers take the place of grasses : hence the constantly recurring question of improvement. In other countries pastures are not exclusively used for neat-stock and horses. In Europe and the British Islands sheep and hogs are pastured ; and these animals maintain a high degree of fertility at the same time that a profitable husbandry is carried on. We are not accustomed to pasture pigs ; and many would, without thought, deny that the pig is a grazing animal. Pigs can be very profitably grown in pasture ; and when herds of them are kept in small compass, and fed with grain, wonderful results in the way of fertilization can be achieved upon worn-out soils. Sheep-husbandry is a sure means of restoring and main taining fertility ; but the present relation of sheep and dogs renders it useless for our benefit. If the matter is brought to the cold light of figures, it would probably appear that the greater part of our pasture land is really not worth renovating. Much of it is poor, stony, and distant. Fencing is a great expense. The land is in use only three or four months in the year. Our best arable fields are not well tilled ; the fault of our farming being in spreading over a wide expanse of land, instead of concentrating labor upon fewer acres. Progressive farmers incline less to pastures, and more to cultivation. Instead of wandering over barren tracts of inhospitable pasture, their cows are kept in convenient yards at home, and fed from green crops raised, in rotation, upon land that is never idle, and not run to waste. Important information in regard to this method of feeding will be found in subsequent pages.
// // HomeViewController.swift // Jad's CV // // Created by Jad Osseiran on 16/07/2014. // Copyright (c) 2014 Jad. All rights reserved. // import UIKit class HomeViewController: UIViewController { //MARK:- Properties /// The model object of the résumé holder. let résuméHolder = RésuméHolder.résuméHolder() let sectionsPageViewController = RefereesCollectionViewController() var homeView: HomeView! { return view as? HomeView } //MARK:- Init required init(coder aDecoder: NSCoder!) { super.init(coder: aDecoder) } override init(nibName nibNameOrNil: String!, bundle nibBundleOrNil: NSBundle!) { super.init(nibName: nibNameOrNil, bundle: nibBundleOrNil) self.addChildViewController(self.sectionsPageViewController) } convenience override init() { self.init(nibName: nil, bundle: nil) } //MARK:- View lifecycle override func loadView() { view = HomeView(sectionsPageView: sectionsPageViewController.view) } override func viewDidLoad() { setupProfileView() } override func viewDidAppear(animated: Bool) { super.viewDidAppear(animated) UIView.animateWithDuration(Animations.Durations.Short.toRaw()) { self.setNeedsStatusBarAppearanceUpdate() } } //MARK:- Status bar override func preferredStatusBarStyle() -> UIStatusBarStyle { return homeView.profileView.expanded ? .Default : .LightContent } //MARK:- Actions func profileInfoAction(sender: AnyObject) { homeView.handleProfileViewFocus() UIView.animateWithDuration(kProfileViewAnimationDuration) { self.setNeedsStatusBarAppearanceUpdate() } } func emailAction(sender: AnyObject) { Contactor.sharedContactor.email(reciepients: [résuméHolder.email], fromController: self) } func phoneAction(sender: AnyObject) { Contactor.call(number: résuméHolder.phoneNumber) } //MARK:- Logic func setupProfileView() { let profileView = homeView.profileView profileView.infoButton.addTarget(self, action: "profileInfoAction:", forControlEvents: .TouchUpInside) profileView.nameLabel.text = résuméHolder.fullName profileView.descriptionLabel.text = résuméHolder.location profileView.profilePictureImageView.image = résuméHolder.picture profileView.backgroundImageView.image = résuméHolder.backgroundImage profileView.textView.text = résuméHolder.detailDescription profileView.emailButton.setTitle(résuméHolder.email, forState: .Normal) profileView.phoneButton.setTitle(résuméHolder.phoneNumber, forState: .Normal) profileView.emailButton.addTarget(self, action: "emailAction:", forControlEvents: .TouchUpInside) profileView.phoneButton.addTarget(self, action: "phoneAction:", forControlEvents: .TouchUpInside) profileView.emailButton.enabled = UIDevice.canEmail() profileView.phoneButton.enabled = UIDevice.canCall() } }
/* * $Source: /project/arl/arlcvs/cvsroot/wu_arl/wusrc/wulib/stats.c,v $ * $Author: fredk $ * $Date: 2007/06/06 21:06:38 $ * $Revision: 1.12 $ * * Author: Fred Kuhns * [email protected] * Organization: Applied Research Laboratory * Washington UNiversity in St. Louis * */ /* Standard include files */ #include <wulib/kern.h> #include <math.h> #include <sys/time.h> #include <wulib/wulog.h> #include <wulib/stats.h> #include <wulib/timer.h> static FILE *open_fp(char *name, char *ext); /* Returns File pointer on success or NULL on error */ static FILE * open_fp(char *fname, char *ext) { char filenm[FILENAME_MAX]; FILE *fp; if (fname == NULL) return NULL; if (ext) sprintf(filenm, "%s.%s", fname, ext); else sprintf(filenm, "%s", fname); if ((fp = fopen(filenm, "w")) == NULL) { wulog(wulogStat, wulogError, "open_fp: Unable to open dist file %s!\n", filenm); return (FILE *)NULL; } return fp; } /* Hold Measurement Parameters and Stats */ void print_stats (stats_t *stats) { int64_t tmp; int i; FILE *sumfp=0, *cdffp=0, *freqfp=0, *sampfp=0; char *units; if (stats == NULL) return; if (stats->scale == STATS_NSEC_SCALE) units = STATS_NSEC_TXT; else if (stats->scale == STATS_USEC_SCALE) units = STATS_USEC_TXT; else if (stats->scale == STATS_MSEC_SCALE) units = STATS_MSEC_TXT; else if (stats->scale == STATS_SEC_SCALE) units = STATS_SEC_TXT; else units = NULL; /* sumfp -- summary data, * freqfp distribution data (for histogram), * sampfp sample data * */ printf("\nMeasurement Results for %i samples, units are %s: \ \n\tSample Mean = %.2f, Measurement Overhead = %lld \ \n\tMax Value = %lld, Max sample number = %d, \ \n\tMin Value = %lld, Min sample number = %d, \ \n\tStandard Deviation = %.2f, \ \n\tStandard Error = %.2f, \ \n\tStdDev as a percent of Mean (StdDev/Mean * 100) = %.2f%%\n\n", stats->indx, (units ? units : "Unknown"), stats->mean, stats->ohead, stats->max_value, stats->max_index, stats->min_value, stats->min_index, stats->sdev, stats->serr, (stats->sdev/stats->mean)*100); if (stats->datafile) { if ((freqfp = open_fp(stats->datafile, STATS_FREQ_EXT)) == NULL || (sumfp = open_fp(stats->datafile, STATS_SUM_EXT)) == NULL || (cdffp = open_fp(stats->datafile, STATS_CDF_EXT)) == NULL || (sampfp = open_fp(stats->datafile, STATS_SAMPLE_EXT)) == NULL) { free (stats->datafile); if (freqfp) fclose (freqfp); if (sumfp) fclose (sumfp); if (cdffp) fclose (cdffp); if (sampfp) fclose (sampfp); return; } tmp = stats->min_value + stats->window/2; printf("Ndist = %d\n", stats->ndist); for (i = 0 ; i < stats->ndist ; i++) { fprintf(freqfp, "%lld\t%d\n", tmp, stats->Dist[i]); tmp += stats->window; } tmp = stats->min_value - 1; for (i = 0; i < stats->ncdf ; i++) { fprintf(cdffp, "%lld\t%.4f\n", tmp, stats->cdf[i]); tmp += stats->window; } if (stats->tstamps) { for (i = 0; i < stats->indx; i++) fprintf(sampfp, "%lld\t%lld\n", stats->tstamps[i], stats->samples[i]); } else { for (i = 0; i < stats->indx; i++) fprintf(sampfp, "%lld\n", stats->samples[i]); } fprintf(sumfp, "Latency Measurement Record:"); fprintf(sumfp,"\nMeasurement Results for %i samples (%s): \ \n\tSample Mean = %.2f, Measurement Overhead = %lld \ \n\tMax Value = %lld, Max sample number = %d, \ \n\tMin Value = %lld, Min sample number = %d, \ \n\tStandard Deviation = %.2f, \ \n\tStandard Error = %.2f, \ \n\tStdDev as a percent of Mean (StdDev/Mean * 100) = %.2f%%\n\n", stats->indx, (units ? units : "Unknown"), stats->mean, stats->ohead, stats->max_value, stats->max_index, stats->min_value, stats->min_index, stats->sdev, stats->serr, (stats->sdev/stats->mean)*100); fclose (freqfp); fclose (sumfp); fclose (sampfp); } return; } stats_t * reuse_stats (stats_t *stats) { if (stats == NULL) return NULL; if (stats->Dist) { free(stats->Dist); stats->Dist = NULL; stats->ndist = 0; } if (stats->cdf) { free(stats->cdf); stats->cdf = NULL; stats->ncdf = 0; } stats->indx = 0; // stats->ohead = 0; /* */ stats->sum = 0; stats->mean = 0.0; stats->sdev = 0.0; stats->serr = 0.0; /* */ // stats->window = 0; /* */ stats->max_value = 0; stats->max_index = 0; stats->min_value = 0; stats->min_index = 0; /* */ /* */ // stats->scale = 0; memset((void *)stats, 0, sizeof(stats_t)); return stats; } void release_stats (stats_t *stats) { if (stats == NULL) return; /* this free memory */ reuse_stats (stats); if (stats->samples) { free(stats->samples); stats->samples = NULL; } if (stats->tstamps) { free(stats->tstamps); stats->tstamps = NULL; } if (stats->datafile) { free(stats->datafile); stats->datafile = NULL; } free(stats); return; } /* * nsamples - the maximum number of samples that will be collected. * window - the bin size in usecs for histogram data * fname - file prefix used for output data. If null then data is * not written to a file. * scale - 0, STATS_NSEC_SCALE, STATS_USEC_SCALE, STATS_MSEC_SCALE, * STATS_SEC_SCALE */ stats_t * init_mstats (int nsamples, int window, const char *fname, int scale) { stats_t *stats; #ifdef __i386__ if (cpuspeed == 0) { if (init_timers () < 0) { wulog(wulogStat, wulogError, "init_mstats: Unable to initialize timers\n"); return NULL; } } #endif if (scale == 0) { scale = STATS_NSEC_SCALE; } else if (scale != STATS_NSEC_SCALE && scale != STATS_USEC_SCALE && scale != STATS_MSEC_SCALE && scale != STATS_SEC_SCALE) { wulog( wulogStat, wulogError, "init_mstats: Invalid scale value (%d)\n", scale); return NULL; } /* First make sure we can allocate a stats struct */ stats = (stats_t *)malloc(sizeof(stats_t)); if (stats == NULL) { wulog(wulogStat, wulogError, "init_mstats: Unable to allocate memory for stats struct\n"); return NULL; } memset((void *)stats, 0, sizeof(stats_t)); /* Do they want the results written to a file? */ if (fname != NULL) { stats->datafile = (char *)calloc((strlen(fname) + 1), sizeof(char)); if (stats->datafile == NULL) { wulog(wulogStat, wulogError, "init_mstats: Unable to allocate memory file file name\n"); return NULL; } strcpy(stats->datafile, fname); } else stats->datafile = NULL; stats->samples = (int64_t *)calloc(nsamples, sizeof(int64_t)); if (stats->samples == NULL) { wulog(wulogStat, wulogError, "init_mstats: Unable to allocate memory for sample array\n"); if (stats->datafile) free(stats->datafile); free(stats); return NULL; } /* * Note, calloc will zero fill the allocated memory so no need to deal with * fields that have default values of zero ... but I do it anyway. * */ stats->nsamples = nsamples; stats->indx = 0; stats->ohead = 0; /* */ stats->sum = 0; stats->mean = 0.0; stats->sdev = 0.0; stats->serr = 0.0; /* */ stats->window = window; /* */ stats->max_value = INT64_MIN; stats->max_index = 0; stats->min_value = INT64_MAX; stats->min_index = 0; /* */ stats->Dist = NULL; stats->ndist = 0; stats->ncdf = 0; stats->cdf = NULL; /* */ stats->scale = scale; return stats; } void do_figuring (stats_t *stats) { int64_t window = 0; int d, i; if (stats == NULL || stats->indx <= 0) return; /* we use index in case there really aren't nsamples samples */ stats->mean = ((double)stats->sum)/((double)stats->indx); if (stats->window) { window = stats->window; /* calculate number of buckets: * window = w = size of bin (range of values) * * bucket number for sample i (Si) = bi. * w = window size in "units" * if Si = min, bi = 0 (bucket number 0) * if Si = max, bi = n-1, (bucket number n-1, there are n buckets) * * Let S = S' + min, then S' = S - min * * b = 0, for 0*w <= S' < 1*w or 0 <= S'/w < 1 * b = 1, for 1*w <= S' < 2*w or 1 <= S'/w < 2 * b = 2, for 2*w <= S' < 3*w or 2 <= S'/w < 3 * ... ... * b = n-1, for (n-1)*w <= S' < n*w or (n-1) <= S'/w < n * * But (n-1)w <= (max - min) < nw * * This is true for n = floor((max - min)/w) + 1; * let ndist = n, see below * */ stats->ndist = ((int)(stats->max_value - stats->min_value) / window) + 1; stats->ncdf = stats->ndist + 1; stats->Dist = (int *)calloc(stats->ndist, sizeof(int)); stats->cdf = (double *)calloc(stats->ncdf, sizeof(double)); if (stats->Dist == NULL || stats->cdf == NULL) { wulog(wulogStat, wulogError, "do_figuring: Unable to allocate Dist/CDF array\n"); if (stats->Dist) free(stats->Dist); if (stats->cdf) free(stats->cdf); stats->Dist = NULL; stats->cdf = NULL; window = 0; } } /* Standard Deviation = * SQRT(SUM((sample - mean)**2))/(number_samples - 1) * here we get the SUM part. * * Standard Error = * Standard_Deviation/SQRT(number_samples) */ for (i = 0; i < stats->indx ; i++ ) { stats->sdev += ((double)stats->samples[i] - stats->mean) * ((double)stats->samples[i] - stats->mean); /* get index into distribution array */ if (window) { d = (int)((stats->samples[i] - stats->min_value)/window); if (d < 0 || d > stats->ndist) { wulog(wulogStat, wulogError, "do_figuring: Error indexing into dist array %d (%d)\n\n", d, stats->ndist); window = 0; free(stats->Dist); free(stats->cdf); stats->Dist = NULL; stats->cdf = NULL; continue; } stats->Dist[d] += 1; /* increment bucket count for this sample */ } } /* for all samples */ if (window) { stats->cdf[0] = 0.0; // X <= min_value - 1 for (i = 1; i < stats->ncdf; i++) { // P(X <= min_value + i*window - 1) stats->cdf[i] = stats->cdf[i-1] + stats->Dist[i-1]; } for (i = 0; i < stats->ncdf; i++) stats->cdf[i] = stats->cdf[i] / (double)stats->indx; } if (stats->indx > 1) stats->sdev = (double) sqrt(stats->sdev/(double)(stats->indx - 1)); else stats->sdev = (double)0; stats->serr = (double) stats->sdev/sqrt((double)stats->indx); return; }
Velocity: First Argument was not a property map I am using Velocity to do some animations (without jQuery), and I am having an issue where I get the following error: Velocity: First argument ([object HTMLDivElement]) was not a property map, a known action, or a registered redirect. Aborting. I have looked at other examples, and they seem to do it the same way I do, and thiers works. Why is mine not working? When I dump out the value of rev.elem I get a HTMLDivElement Velocity(rev.elem, 'slideUp', { duration: 225, queue: false, easing: 'easeInOutQuad' }); This guy seems to use it the same way and it works for him: http://jasonweaver.name/blog/velocity-js-vertical-slide-without-jquery I got it to work by importing both velocity and velocity ui: import Velocity from 'velocity-animate'; import 'velocity-animate/velocity.ui'; This did not work for me. If you're using the beta Velocity V2 then this won't work yet - I removed the shortcuts such as this in favour of an extensible API - I'm working on sequences right now and will return to this answer when they are in there (within the next few weeks). I have got the example of this up and running... The example is based on Jason Weaver's original example but has been updated to the latest version of Velocity and has the PUG syntactical error fixed. var el = document.querySelector(element), // cache click target con = document.querySelector(container), // cache container target up = 'slideUp', // store up command down = 'slideDown'; // store down command el.addEventListener('click', function(event) { var active = el.classList.contains('active'); // store active state el.classList.toggle('active'); // toggle click target active Velocity(con, active ? up : down, { // test and init command duration: duration, // duration set in function call params easing: easing // easing set in function call params }); event.preventDefault(); }); You need to import both "velocity" and "velocity ui": <script<EMAIL_ADDRESS><script<EMAIL_ADDRESS>
Camera system and lens barrel ABSTRACT A lens barrel of the present invention includes: a first barrel; a second barrel capable of relative rotation around an optical axis with respect to the first barrel; a focusing lens; a drive portion that drives the focusing lens; a rotational operation member that can move in the optical axis direction to a first position, and a second position; and an engagement portion that, when the rotational operation member is at the second position, causes the second barrel and the rotational operation member to engage with each other and rotates the second barrel accompanying rotation of the rotational operation member, and when the rotational operation member is at the first position, disengages the second barrel and the rotational operation member from each other. CROSS REFERENCE TO RELATED APPLICATION This application is a continuation application of PCT/JP2012/060441 filed on Apr. 18, 2012 and claims benefit of Japanese Application No. 2011-128617 filed in Japan on Jun. 8, 2011, the entire contents of which are incorporated herein by this reference. BACKGROUND OF THE INVENTION 1. Field of the Invention The present invention relates to a camera system in which an autofocus operation is possible, and a lens barrel that is used in the camera system. 2. Description of the Related Art Camera systems in which a manual focus operation and an autofocus operation are possible mainly adopt a configuration that includes a rotatable rotational operation member on a lens barrel, and in which a focusing distance is changed in accordance with rotation of the rotational operation member at the time of a manual focus operation, and the focusing distance is changed by a motor at the time of an autofocus operation. For example, Japanese Patent Application Laid-Open Publication No. 6-11643 discloses a lens barrel in which it is possible that a rotational operation member does not rotate at a time of autofocusing, and a focusing operation is performed in accordance with rotation of the rotational operation member at a time of manual focusing. Furthermore, the lens barrel disclosed in Japanese Patent Application Laid-Open Publication No. 6-11643 has a configuration that includes a rotatable display member on which a distance scale that displays focusing distances is provided, and in which the display member is rotated so that a distance displayed by the distance scale and a focusing distance correspond. SUMMARY OF THE INVENTION A lens barrel according to one aspect of the present invention includes: a first barrel; a second barrel that is capable of relative rotation around an optical axis with respect to the first barrel; a drive source; a focusing lens; a drive portion that is driven by the drive source, and that drives the focusing lens in an optical axis direction; a rotational operation member that is a member operated for driving the focusing lens in the optical axis direction, and that can move in the optical axis direction to a first position at which the rotational operation member covers the second barrel and to a second position at which the rotational operation member causes the second barrel to be exposed to outside, and which is configured to be capable of rotation around the optical axis at the first position and the second position, respectively; and engagement means that, when the rotational operation member is at the second position, causes the second barrel and the rotational operation member to engage with each other and rotates the second barrel accompanying rotation of the rotational operation member, and when the rotational operation member is at the first position, disengages the second barrel and the rotational operation member from each other; wherein when the rotational operation member is at the first position, the drive portion drives the focusing lens in accordance with rotation of the rotational operation member, and when the rotational operation member is at the second position, the drive portion drives the focusing lens in accordance with rotation of the second barrel that is engaged with the rotational operation member. A camera system according to one aspect of the present invention includes: a first barrel; a second barrel that is capable of relative rotation around an optical axis with respect to the first barrel; a drive source; a focusing lens; a drive portion that is driven by the drive source, and that drives the focusing lens in an optical axis direction; a rotational operation member that is a member operated for driving the focusing lens in the optical axis direction, and that can move in the optical axis direction to a first position at which the rotational operation member covers the second barrel and to a second position at which the rotational operation member causes the second barrel to be exposed to outside, and which is configured to be capable of rotation around the optical axis at the first position and the second position, respectively; and engagement means that, when the rotational operation member is at the second position, causes the second barrel and the rotational operation member to engage with each other and rotates the second barrel accompanying rotation of the rotational operation member, and when the rotational operation member is at the first position, disengages the second barrel and the rotational operation member from each other; wherein when the rotational operation member is at the first position, the drive portion drives the focusing lens in accordance with rotation of the rotational operation member, and when the rotational operation member is at the second position, the drive portion drives the focusing lens in accordance with rotation of the second barrel that is engaged with the rotational operation member. BRIEF DESCRIPTION OF THE DRAWINGS FIG. 1 is a perspective view showing a front face side of a camera constituting a camera system; FIG. 2 is a top face view of a lens barrel in a state where a rotational operation member is positioned at a first position; FIG. 3 is a top face view of the lens barrel in a state where the rotational operation member is positioned at a second position; FIG. 4 is a cross-sectional view of the lens barrel; FIG. 5 is a cross-sectional view of the lens barrel in a state where the rotational operation member is positioned at the first position; FIG. 6 is a cross-sectional view of the lens barrel in a state where the rotational operation member is positioned at the second position; FIG. 7 is an exploded perspective view illustrating a state where engagement between the rotational operation member and a display member has been released; FIG. 8 is an exploded perspective view illustrating a state where the rotational operation member and the display member are engaged; FIG. 9 is an exploded perspective view illustrating a configuration of rotation restriction means; FIG. 10 is an exploded perspective view illustrating another configuration of the rotation restriction means; FIG. 11 is a perspective view illustrating a configuration of a second encoder portion; FIG. 12 is a block diagram of an electric circuit of a configuration involved in focusing operations of the camera system; FIG. 13 is a flowchart of a focus operation mode switching subroutine; FIG. 14 is a flowchart of an autofocus operation subroutine; FIG. 15 is a flowchart of a manual focus operation subroutine; FIG. 16 is a flowchart of a distance-specification focus operation; FIG. 17 is a perspective view showing a front face side of a camera constituting a camera system according to a second embodiment; FIG. 18 is a block diagram of an electric circuit of a configuration involved in focusing operations of the camera system of the second embodiment; FIG. 19 is a cross-sectional view of a lens barrel in a state where a rotational operation member is positioned at a first position according to a third embodiment; FIG. 20 is a cross-sectional view of the lens barrel in a state where the rotational operation member is positioned at a second position according to the third embodiment; FIG. 21 is a partial enlarged cross-sectional view of a lens barrel in a state where a rotational operation member is positioned at a first position according to a fourth embodiment; FIG. 22 is a partial enlarged cross-sectional view of the lens barrel in a state where the rotational operation member is positioned at a second position according to the fourth embodiment; FIG. 23 is a flowchart of a focus operation mode switching subroutine of a camera system according to a fifth embodiment; FIG. 24 is a view illustrating another form of display means that includes an indicator and a distance scale; and FIG. 25 is a view illustrating another form of display means that includes an indicator and a distance scale. DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS Preferred embodiments of the present invention are described hereunder with reference to the drawings. It should be noted that each of the components in the drawings referred to in the following description is displayed in a different contraction scale so as to be shown in a size that is recognizable in the drawings. Further, the present invention is not limited to only the quantity of components, the shapes of components, the ratios between the sizes of components, and the relative positional relationship between the respective components described in the drawings. First Embodiment As shown in FIG. 1, a camera system 1 of the present embodiment includes a camera body 2 and a lens barrel 10 as an interchangeable lens. The lens barrel 10 holds an optical system member 11 for forming an object image. As one example according to the present embodiment, the camera system 1 has a form in which the camera body 2 and the lens barrel 10 can be detachably attached to each other. Note that the camera system 1 may also have a form in which the camera body and the lens barrel are integrated. Further, as one example according to the present embodiment, the camera system 1 has a configuration referred to as a so-called “electronic camera” or “digital camera” or the like, in which an image pickup device 9 is provided in the camera body 2 and which electronically picks up an object image and records the image. The image pickup device 9 outputs, at a predetermined timing, an electrical signal in accordance with light incident on a light-receiving surface (pixel region). For example, the image pickup device 9 has the form of a charge coupled device (CCD), a CMOS (complementary metal-oxide semiconductor) sensor or the like. The camera system 1 is configured so as to enable an autofocus operation (automatic focus operation). An autofocus sensor portion that is used for an autofocus operation is arranged in the camera body 2. As one example according to the present embodiment, the camera system 1 is configured so as to perform an autofocus operation according to a so-called “contrast detection method” in which a contrast value of an object image is detected based on a signal that is outputted from the image pickup device 9, and focusing control of the optical system member 11 is performed so that the contrast value becomes the maximum value. That is, in the camera system 1 of the present embodiment, the image pickup device 9 is an autofocus sensor portion. Note that the camera system 1 may also be configured to perform an autofocus operation according to a so-called “phase difference detection method”. In this case, a sensor that detects a phase difference of an object image that is arranged in the camera body 2 serves as an autofocus sensor portion. Further, the autofocus sensor portion may be a distance measuring sensor of a different form or the like. A release switch 3 that a user uses to input an instruction to perform an image pickup operation, and a power switch 4 that a user uses to input an instruction to perform an operation to turn the power supply of the camera body 2 on and off are arranged on the top face portion of the camera body 2. According to the present embodiment, the release switch 3 is a push-button type switch that includes two release switches, namely, a first release switch 3 a and a second release switch 3 b, that enter an “on” state in response to two different stroke amounts (depression amounts). When a so-called “half-stroke operation” is performed in which the release switch 3 is depressed by an amount that is a partial amount of a full stroke amount, the first release switch 3 a enters an “on” state. When a so-called “full-stroke operation” is performed in which the release switch 3 is depressed further than in the half-stroke operation, the second release switch 3 b enters an “on” state. When the second release switch 3 b enters an “on” state, the camera system 1 executes an image pickup operation and stores the image. Note that the release switch 3 may be of a form in which the first release switch 3 a and the second release switch 3 b are arranged at separated positions. Further, the release switch 3 is not limited to the form of a push-button type switch, and may be a switch of another form such as a touch sensor. Although not shown in the drawings, a focus mode switching operation portion 5 for inputting an instruction to switch a mode of a focusing operation of the camera system 1 is arranged on the camera body 2. By operating the focus mode switching operation portion 5, a user selects either one of an autofocus operation mode that performs an autofocus operation and a manual focus operation mode that performs a manual focus operation as the focus operation mode of the camera system 1. Operations of the camera system 1 in the autofocus operation mode and in the manual focus operation mode are described later. Note that the focus mode switching operation portion 5 is not limited to the form of a button switch, and may be a touch sensor or a dial switch or the like. The camera body 2 also includes an image display apparatus, and a form may also be adopted in which a user switches the focus operation mode by selecting, through a button switch or a touch sensor, a menu displayed on the image display apparatus. A form may also be adopted in which the focus mode switching operation portion 5 is arranged on the lens barrel 10. Although not shown in the drawings, a battery housing portion that houses a primary battery or a secondary battery for supplying power to the camera system 1, and a storage media housing portion that houses a flash memory for storing images are provided in the camera body 2. According to the present embodiment, the camera body 2 and the lens barrel 10 can be detachably attached to each other by means of an engagement mechanism that is generally referred to as a “bayonet mount”. Note that, in the camera system 1, a configuration that makes the camera body 2 and the lens barrel 10 detachably attachable to each other is not limited to the present embodiment and, for example, a configuration that is generally referred to as a “screw-type mount” that uses a screw mechanism may be adopted. Further, a configuration may be adopted that makes the camera body 2 and the lens barrel 10 detachably attachable to each other by means of a mechanism that fits the camera body 2 and the lens barrel 10 together or uses a magnet or the like. The lens barrel 10 includes a base portion 12, a fixed barrel (first barrel) 14, a focusing barrel 13, a drive portion 15, an indicator display barrel 16, a display member (second barrel, distance display ring, distance display means) 18, and a rotational operation member (operation member, rotational operation ring, rotational operation means, rotational operation portion) 17. Note that the display member 18 is mainly constituted by a first cylindrical member that has a cylindrical shape, and also includes a thin second cylindrical member that is shorter than the first cylindrical member in the axial direction that is fitted to the outer circumference thereof and fixed thereto. A distance scale 18 a that is described later is displayed on the outer circumference of the second cylindrical member. Hereunder, the display member 18 is described as a member in which the first cylindrical member and the second cylindrical member are integrated. As shown in FIG. 2, the base portion 12 has a bayonet portion 12 a that engages with the camera body 2. The base portion 12 is fixed to the camera body 2 by the bayonet portion (bayonet claw) 12 a engaging with the camera body 2. As shown in FIG. 4, the lens barrel 10 has a mechanism that holds an optical system member 11 that includes a plurality of lens and the like as an image pickup optical system with respect to the base portion 12. Note that the form of the optical system member 11 that the lens barrel 10 holds may be a form that includes a diaphragm, a prism, a mirror, a filter or the like in addition to lenses. A configuration is adopted so that the focusing distance changes in accordance with relative movement of part or all of the optical system member (optical system elements and image pickup optical system) 11 in an optical axis O direction of the optical system member 11 with respect to the base portion 12. Here, the term “focusing distance change” refers to changing a distance to an object that an attempt is being made to focus on. The term also refers to changing a focus position on which the optical system member 11 is focused. Hereunder, among the elements of the optical system member 11, an element that moves in the optical axis O direction when changing the focusing distance is referred to as a focusing lens (focusing optical element, focusing optical system member) 11 a. Specifically, the optical system member 11 of the present embodiment includes a plurality of lenses and a diaphragm mechanism portion 19 that are arranged along the optical axis O. According to the present embodiment, a lens that is arranged at the most rearward position (image side) among the plurality of lens of the optical system member 11 is the focusing lens 11 a. The focusing lens 11 a is held by the focusing barrel 13 that is arranged so as to be moveable forward and rearward in the optical axis O direction relative to the base portion 12. The focusing barrel 13 is driven in the optical axis O direction by the drive portion 15. Although the configuration of the drive portion 15 is not particularly limited, according to the present embodiment, the drive portion 15 includes a screw 15 a arranged substantially parallel to the optical axis O, a drive source (motor) 15 c that rotates the screw 15 a, and a nut 15 b that is screwed together with the screw 15 a. According to the present embodiment, the drive source 15 c is a stepping motor. The rotation around the screw 15 a of the nut 15 b is restricted, and the nut 15 b moves substantially parallel to the optical axis O accompanying rotation of the screw 15 a. The focusing barrel 13 is engaged with the nut 15 b so as to follow the nut 15 b. The drive portion 15 drives the focusing barrel 15 in the optical axis O direction by rotating the screw 15 a by means of the drive source 15 c. Note that the configuration of the drive portion 15 is not limited to the present embodiment, and may be another form such as a linear motor. In addition, the drive portion 15, for example, may have a configuration in which some constituent members, such as the drive source, are arranged inside the camera body 2. Elements other than the focusing lens 11 a of the optical system member 11 are held by a fixed barrel 14 as a first barrel whose position with respect to the base portion 12 is fixed. Note that, although elements other than the focusing lens 11 a of the optical system member 11 are held by the fixed barrel 14 because the lens barrel 10 of the present embodiment has the form of a so-called “fixed focal length lens” in which the focal distance is fixed, it goes without saying that in a case where the lens barrel 10 is a so-called “collapsible lens barrel” in which the entire length can be expanded and contracted or a case where the lens barrel 10 is a so-called “zoom lens” or “varifocal lens” in which a focal distance can be changed, members other than the focusing lens 11 a of the optical system member 11 are also held by a barrel member that moves relatively with respect to the base portion 12. The rotational operation member 17, the indicator display barrel 16 and the display member 18 are arranged on the outer circumferential portion of the lens barrel 10. The rotational operation member 17 is a substantially cylindrical member that is arranged so as to be rotatable around the optical axis O of the optical system member 11 on the outer circumferential portion of the lens barrel 10, and also to be movable forward and rearward in the optical axis O direction. At least one part of the rotational operation member 17 is exposed on the outer circumferential face of the lens barrel 10, and is arranged so that a finger of a user of the camera system 1 engages therewith. Specifically, according to the present embodiment, as shown in the cross-sectional views of FIG. 5 and FIG. 6 and in the perspective views of FIG. 7 and FIG. 8, the rotational operation member 17 includes two substantially cylindrical areas, namely, a substantially cylindrical operation portion 17 a that is exposed on the outer circumferential face of the lens barrel 10 and in which projections and depressions are provided on an outer circumferential portion so that a finger of a user engages therewith and an inside cylindrical portion (engagement barrel) 17 b that is substantially cylindrical and is arranged on the inner side of the operation portion 17 a with a predetermined clearance therebetween. Note that, according to the present embodiment illustrated in the drawings, although the rotational operation member 17 is constituted by the operation portion 17 a and the inside cylindrical portion 17 b that are separate members which are, for example, fixed by screws or an adhesive, a form may also be adopted in which the operation portion 17 a and the inside cylindrical portion 17 b are formed integrally with each other. The rotational operation member 17 rotates around the optical axis O by means of a force applied to the operation portion 17 a by fingers of the user. As shown in FIG. 2 and FIG. 3, the rotational operation member 17 is moveable within a predetermined range in the optical axis O direction, and is arranged so as to be selectively positioned at either one of two ends of the moving range in a state in which a force is not being applied thereto from outside. The rotational operation member 17 can be moved back and forth from one end to the other end of the moving range and from the other end to the one end thereof in the optical axis O direction by a force applied to the operation portion 17 a by fingers of the user. Hereunder, among the two positions at which the rotational operation member 17 is selectively positioned, an end portion on the front side (object side) of the moving range in the optical axis O direction is referred to as a “first position” and an end portion on the rear side (image side) of the moving range is referred to as a “second position”. That is, FIG. 2, FIG. 4, FIG. 5 and FIG. 7 illustrate a state in which the rotational operation member 17 is positioned at the first position, while FIG. 3, FIG. 6 and FIG. 8 illustrate a state in which the rotational operation member 17 is positioned at the second position. As shown in FIG. 5 and FIG. 6, on an inner circumferential face of the rotational operation member 17, a convex portion 17 c that protrudes to the inner side in the diameter direction and has a cross-section that is a substantially triangular shape is formed across the entire circumference in the circumferential direction. The convex portion 17 c includes a first inclined face portion 17 d and a second inclined face portion 17 e. The first inclined face portion 17 d is formed so that the inner diameter thereof decreases in accordance with proximity to the front along the optical axis O and reaches the apex of the triangular shape. The second inclined face portion 17 e is formed so the inner diameter thereof increases in accordance with proximity to the front along the optical axis O from the apex of the triangular shape on the front side of the first inclined face portion 17 d. A through-hole 14 a is formed at a location facing the convex portion 17 c in the fixed barrel 14 that is arranged on the inner side of the rotational operation member 17. A ball 14 b is loosely fitted inside the through-hole 14 a. The ball 14 b can protrude further to the outside than the outer circumferential face of the fixed barrel 14, and is urged towards the outer side in the diameter direction of the fixed barrel 14 by an urging member 14 c that is a plate spring. A spring portion (plate spring) of the urging member 14 c is disposed on the inner circumferential face of the fixed barrel 14. The urging member 14 c and the ball 14 b are arranged at a plurality of places in the circumferential direction. When the rotational operation member 17 is positioned at the first position, the ball 14 b contacts against the first inclined face portion 17 d of the convex portion 17 c, and when the rotational operation member 17 is positioned at the second position, the ball 14 b contacts against the second inclined face portion 17 e of the convex portion 17 c. Because the convex portion 17 c has a substantially triangular cross-section, irrespective of the position of the rotational operation member 17, the ball 14 b is always contacting against either one of the first inclined face portion 17 d and the second inclined face portion 17 e. Accordingly, when the rotational operation member 17 is situated towards the front of the moveable range, the rotational operation member 17 is urged front ward by the ball 14 b that contacts against the first inclined face portion 17 d, and contacts against one end of the moveable range at the first position and is thereby positioned at the first position. On the other hand, when the rotational operation member 17 is situated towards the rear of the moveable range, the rotational operation member 17 is urged rearward by the ball 14 b that contacts against the second inclined face portion 17 e, and contacts against the other end of the moveable range at the second position and is thereby positioned at the second position. Therefore, in the lens barrel 10 of the present embodiment, when an external force is not being applied to the rotational operation member 17, the position in the optical axis O direction of the rotational operation member 17 is either one of the first position and the second position. For example, in a state where the rotational operation member 17 is positioned at the first position, if an external force in the rearward direction that is applied to the rotational operation member 17 is weaker than an urging force that the urging member 14 c and the ball 14 b generate, when the external force disappears, the rotational operation member 17 returns to the first position. Further, for example, in a state where the rotational operation member 17 is positioned at the first position, if an external force in the rearward direction that is applied to the rotational operation member 17 is stronger than an urging force that the urging member 14 c and the ball 14 b generate, the rotational operation member 17 moves to the second position. A first encoder portion (operation portion rotation detection portion, rotation detection means) 21 is also arranged in the lens barrel 10. The first encoder portion 21 is first encoder means that, at least when the rotational operation member 17 is positioned at the first position, detects a rotational direction and a rotational amount around the optical axis O of the rotational operation member 17. Further, an operation member position detection portion (position detection means) 22 that detects which one of the first position and the second position in the optical axis O direction the rotational operation member 17 is positioned at is arranged in the lens barrel 10. The first encoder portion 21 detects, with a pair of photo-interrupters, passage of a plurality of slits (notches) 17 n that are provided in the rotational operation member 17 at predetermined intervals in the circumferential direction. The first encoder portion 21 detects the rotational direction and the rotational amount of the rotational operation member 17 based on an output signal of the pair of photo-interrupters. The first encoder portion 21 and the slits 17 n provided in the rotational operation member 17 of the present embodiment have the same form as a so-called “incremental-type rotary encoder”. More specifically, according to the present embodiment, as shown in FIG. 7, the slits 17 n are formed in an end portion on a front ward side of the inside cylindrical portion 17 b. As shown in FIG. 5 and FIG. 6, the pair of photo-interrupters constituting the first encoder portion 21 are fixed to the fixed barrel 14. The end portion on the front ward side of the inside cylindrical portion 17 b in which the slits 17 n are formed is positioned within the detection range of the pair of photo-interrupters only when the rotational operation member 17 is positioned at the first position. Accordingly, the first encoder portion 21 of the present embodiment can detect the rotational direction and the rotational amount around the optical axis O of the rotational operation member 17 only in a case where the rotational operation member 17 is positioned at the first position. Note that the form of the first encoder portion 21 is not limited to the present embodiment, and the first encoder portion 21 may be of any form that can detect the rotational direction and the rotational amount around the optical axis O of the rotational operation member 17 in at least a case where the rotational operation member 17 is positioned at the first position. For example, the first encoder portion 21 may have the form of a magnetic rotary encoder. The operation member position detection portion 22 is constituted by a photo-interrupter that is fixed to the base portion 12 or the fixed barrel 14. The operation member position detection portion 22 is fixed to a position such that, when the rotational operation member 17 is positioned at the second position, part of the rotational operation member 17 advances to inside a detection range of the photo-interrupter. Note that the form of the operation member position detection portion 22 is not particularly limited as long as the operation member position detection portion 22 has a configuration that can detect a position in the optical axis O direction of the rotational operation member 17. For example, the operation member position detection portion 22 may be a magnetic sensor or the like. Further, engagement convex portions 17 g constituting engagement means (engagement portion) are provided on the rotational operation member 17. The engagement convex portions 17 g are portions for engaging with the display member 18, described later, when the rotational operation member 17 is positioned at the second position, and causing the rotational operation member 17 and the display member 18 to rotate integrally with each other. The engagement convex portions 17 g include a plurality of convex portions that protrude to the outer side in the diameter direction on the outer circumferential face of the inside cylindrical portion 17 b. As shown in FIG. 7 and FIG. 8, the plurality of convex portions constituting the engagement convex portions 17 g are spaced uniformly in the circumferential direction with a fixed gap therebetween. When viewed from the outer side in the diameter direction, the plurality of convex portions have a substantially V shape in which the width of a part on the rear side narrows towards the rear side. The form of engagement between the engagement convex portions 17 g and the display member 18 is described later. The position of the indicator display barrel 16 is fixed with respect to the base portion 12. The indicator display barrel 16 is a barrel member (first barrel) that has a function of a fixed barrel and is part of an exterior member of the lens barrel 10. The indicator display barrel 16 is fixed to the base portion 12 through the fixed barrel 14. The indicator display barrel 16 is arranged at a position that is further to the front side than the operation portion 17 a of the rotational operation member 17 in a state in which the rotational operation member 17 is positioned at the first position. An indicator 16 a that has the shape of a short straight line that is substantially parallel to the optical axis O is provided on the indicator display barrel 16. The indicator 16 a is a member for pointing at the distance scale 18 a that is provided on the display member 18 as described later. A depth-of-field scale 16 b is arranged on the indicator display barrel 16 in a substantially symmetrical manner on both sides of the indicator 16 a so as to interpose the indicator 16 a therebetween. The depth-of-field scale 16 b is a scale for displaying a depth of field that corresponds to an F number (diaphragm value) of the optical system member 11. The depth-of-field scale 16 b is displayed so that characters representing an F number of the same numeric value form a pair in a manner that interposes the indicator 16 a therebetween. The depth-of-field scale 16 b includes a plurality of such pairs, and each pair denotes an F number of a different value to the other pairs. Note that, although in the description of the present embodiment a plurality of numeric values (5.6, 11, 22) are displayed, it is sufficient that at least only one pair of numeric values that include the same numeric value is displayed. The display member 18 that is a second barrel is a substantially cylindrical member that is arranged on the inside of the rotational operation member 17 and is capable of relative rotation around the optical axis O with respect to the base portion 12. In other words, the display member 18 is capable of relative rotation around the optical axis O with respect to the indicator display barrel 16. As shown in FIG. 3, the distance scale 18 a that shows focusing distances of the optical system member 11 is provided on the outer circumferential face of the display member 18. In the distance scale 18 a, numeric values that show distances from the minimum focusing distance of the optical system member 11 to infinity are arrayed in the circumferential direction. The numeric value of the distance scale 18 a that is pointed at by the indicator 16 a changes as a result of the display member 18 being rotated relatively around the optical axis O with respect to the indicator display barrel 16. The rotational range of the display member 18 around the optical axis O is limited, and the display member 18 can only rotate around the optical axis O within a range in which the distance scale 18 a is pointed at by the indicator 16 a. That is, by means of a combination between the distance scale 18 a and the indicator 16 a, the distance scale 18 a always displays a numeric value of a distance between the minimum focusing distance of the optical system member 11 and infinity. A configuration that limits the rotational range around the optical axis O of the display member 18 is not particularly limited. As one example according to the present embodiment, as shown in FIG. 9, a convex portion 18 c that protrudes rearward is formed at an end portion at the rear of the display member 18, and a first abutting portion (abutting wall) 14 x and a second abutting portion (abutting wall) 14 y that are arranged so as to hold the convex portion 18 c therebetween are formed in the fixed barrel 14 in a manner in which the first and second abutting portions 14 x and 14 y are separated by a predetermined angle in the circumferential direction. The first abutting portion 14 x and the second abutting portion 14 y are arranged so as to abut against the convex portion 18 c. Consequently, the rotational range around the optical axis O of the display member 18 is limited to a range between the positions at which the convex portion 18 c collides with the first abutting portion 14 x and the second abutting portion 14 y. Note that a configuration that limits the rotational range around the optical axis O of the display member 18 is not limited to the configuration shown in FIG. 9. For example, as shown in FIG. 10, a configuration may be adopted in which a convex portion 14 z is formed in the fixed barrel 14, and a first abutting portion 18 x and a second abutting portion 18 y that are arranged so as to hold the convex portion 14 z therebetween are formed in the display member 18 in a manner in which the first and second abutting portions 18 x and 18 y are separated by a predetermined angle in the circumferential direction. That is, rotation restriction means is provided in which a convex portion is provided in one of the fixed barrel 14 or first barrel to which the indicator display barrel 16 is fixed and the display member 18 (second barrel) that is rotatable with respect to the first barrel, and which restricts a range of the relative rotation of the second barrel with respect to the first barrel by interference between the convex portion and the other of the first barrel and the second barrel. According to the present embodiment, as shown in FIG. 2, when the rotational operation member 17 is positioned at the first position, a state is entered in which the distance scale 18 a of the display member 18 can not be seen from outside of the lens barrel 10. In contrast, as shown in FIG. 3, when the rotational operation member 17 is positioned at the second position, a state is entered in which the distance scale 18 a of the display member 18 can be seen from outside of the lens barrel 10. Specifically, the display member 18 is arranged between the operation portion 17 a and the inside cylindrical portion 17 b of the rotational operation member 17. In other words, a cylindrical portion that is the operation portion 17 a of the rotational operation member 17 is arranged at a location that is further on the outer side in the diameter direction than the outer circumferential face of the display member 18. When the rotational operation member 17 is positioned at the first position, the operation portion 17 a advances over the distance scale 18 a and covers and conceals the distance scale 18 a. Further, when the rotational operation member 17 is positioned at the second position, the operation portion 17 a withdraws from the position over the distance scale 18 a and thus the distance scale 18 a is exposed to the outside and displayed. In other words, when the rotational operation member 17 is positioned at the second position, a state is entered in which the display member 18 is exposed to outside. According to the lens barrel 10 of the present embodiment, the display member 18 is configured so as to rotate around the optical axis O together with the rotational operation member 17 only in a case where the rotational operation member 17 is positioned at the second position. Further, when the rotational operation member 17 is positioned at the first position, the rotational operation member 17 can rotate independently of the display member 18. Specifically, one engagement pin 18 b constituting engagement means is provided so as to protrude to the inner side in the diameter direction on the inner circumferential portion (inner circumferential face) of the display member 18. According to the present embodiment, the engagement pin 18 b is a separate member from the display member 18 and is fixed to the display member 18 by press-fitting or an adhesive. Note that the engagement pin 18 b may also be integrally formed with the display member 18. Since the display member 18 is arranged between the operation portion 17 a and the inside cylindrical portion 17 b of the rotational operation member 17, the engagement pin 18 b protrudes in the direction of the inside cylindrical portion 17 b that is arranged on the inner side of the display member 18. In other words, the engagement pin 18 b of the display member 18 and the engagement convex portions 17 g of the rotational operation member 17 protrude in a direction in which they face each other. The external diameter of the engagement pin 18 b is smaller than a clearance between adjacent engagement convex portions 17 g, and the engagement pin 18 b has a shape that fits in a loose fitting state between adjacent engagement convex portions 17 g. In addition, the engagement pin 18 b has a cross-sectional shape in which the front ward side when viewed from the inner side in the diameter direction is a substantially V shape. As shown in FIG. 5 and FIG. 7, when the rotational operation member 17 is positioned at the first position, the engagement pin 18 b is arranged at a position that is further on the rear side than the engagement convex portions 17 g of the rotational operation member 17 and that is a position such that the engagement pin 18 b does not interfere with the engagement convex portions 17 g even if the rotational operation member 17 rotates around the optical axis O. As shown in FIG. 6 and FIG. 8, when the rotational operation member 17 is positioned at the second position, the engagement pin 18 b is positioned in the same circumferential direction as the engagement convex portions 17 g. In other words, when the rotational operation member 17 is positioned at the second position, the engagement pin 18 b is arranged at a position that overlaps with the engagement convex portions 17 g in the optical axis O direction. That is, when the rotational operation member 17 moves relatively from the first position to the second position, the engagement pin 18 b fits between adjacent engagement convex portions 17 g. Conversely, when the rotational operation member 17 moves relatively from the second position to the first position, the fitting state between the engagement pin 18 b and the engagement convex portions 17 g is released. By means of the engagement means including the engagement convex portions 17 g and the engagement pin 18 b having the above described configuration, when the rotational operation member 17 is positioned at the second position, the display member 18 rotates around the optical axis O together with the rotational operation member 17, and when the rotational operation member 17 is positioned at the first position, even if the rotational operation member 17 rotates around the optical axis O, the display member 18 does not rotate and remains stopped. Note that, as described above, the rearward side of each of the plurality of convex portions of the engagement convex portions 17 g has a substantially V shape as viewed from the outer side in the diameter direction, and the front ward side of the engagement pin 18 b has a substantially V shape as viewed from the inner side in the diameter direction. Hence, when the rotational operation member 17 moves from the first position to the second position, the display member 18 slightly rotates due to the substantially V-shaped areas provided in the rotational operation member 17 and the display member 18 following each other, and consequently the engagement pin 18 b and the engagement convex portions 17 g smoothly engage. Therefore, a catching action does not occur when an operation is performed to move the rotational operation member 17 in the optical axis O direction, and movement of the rotational operation member 17 can be rapidly performed. Further, as described above, the rotational range of the display member 18 is limited to a range in which a combination of the distance scale 18 a and the indicator 16 a displays a numeric value of a distance between the minimum focusing distance of the optical system member 11 and infinity. Consequently, when the rotational operation member 17 is positioned at the second position and engaged with the display member 18, it is possible for the rotational operation member 17 to rotate only within the same rotational range as the display member 18. That is, when the rotational operation member 17 is positioned at the second position, a limitation is applied to the rotational range of the rotational operation member 17. As described above, when the rotational operation member 17 is positioned at the second position, the range of rotation of the rotational operation member 17 is also restricted to a predetermined range by the rotation restriction means that restricts the range of the relative rotation of the display member 18 (second barrel) with respect to the fixed first barrel. On the other hand, when the rotational operation member 17 is positioned at the first position, since the rotational operation member 17 does not interfere with the display member 18, there is no limitation on the rotational range of the rotational operation member 17. That is, when the rotational operation member 17 is positioned at the first position, the rotational operation member 17 can rotate without any limitation, without the range of rotation being restricted by the rotation restriction means. A second encoder portion 23 (display member rotation detection portion, rotation position detection means) that is second encoder means that detects an absolute rotational position around the optical axis O of the display member 18 with respect to the base portion 12 is also arranged in the lens barrel 10 of the present embodiment. As shown in FIG. 11, as one example according to the present embodiment, the second encoder portion 23 is configured to have the form of a so-called “absolute rotary encoder”. The second encoder portion 23 includes a code pattern 23 a of a predetermined number of bits constituted by a conductor, and a contact portion 23 b constituted by a conductor that slides over the code pattern 23 a. The code pattern 23 a is arranged on the outer circumferential portion of the fixed barrel 14. The contact portion 23 b is arranged on the display member 18. The position of the code pattern 23 a that the contact portion 23 b contacts changes according to the rotational position around the optical axis O of the display member 18. Although not shown in the drawings, the second encoder portion 23 has an electric circuit that detects a state of contact between the code pattern 23 a and the contact portion 23 b, and the absolute rotational position around the optical axis O of the display member 18 with respect to the base portion 12 can be calculated based on the detection result. Note that the configuration of the second encoder portion 23 is not limited to the present embodiment as long as a form is adopted that enables detection of an absolute rotational position around the optical axis O with respect to the base portion 12. For example, the second encoder portion 23 may be an optical or magnetic absolute rotary encoder, and a form may also be adopted that has a configuration similar to a so-called “potentiometer”, in which a resistance value changes in accordance with the rotational position around the optical axis O of the display member 18. Next, the electrical configuration of portions involved in focusing operations of the camera system 1 is described referring to FIG. 12. As described in the foregoing, the release switch 3, the power switch 4, the focus mode switching operation portion 5 and the image pickup device 9 that is an autofocus sensor portion are arranged in the camera body 2. In addition, a camera body control portion 6, an image processing portion 7, a communication portion 8, an image display circuit 30, and an image display apparatus 31 are arranged in the camera body 2. The camera body control portion (control means) 6 that constitutes a control portion includes a processing apparatus (CPU), a storage apparatus (RAM), an input/output apparatus, a power control apparatus and the like, and controls operations of the camera body 2 based on a predetermined program. The image processing portion 7 is an electronic circuit portion for performing image processing. The image processing portion 7 can calculate a contrast value of an object image based on a signal that is outputted from the image pickup device 9. Note that a form of implementing the image processing portion 7 in the camera body 2 may be a hardware form in which computational hardware for image processing is mounted in the camera body 2, or may be a software form in which the processing apparatus of the camera body control portion 6 performs image processing based on a predetermined image processing program. The communication portion 8 is a portion for performing communication with a lens barrel control portion 24 through a communication portion 25 provided in the lens barrel 10 by wire communication or wireless communication. The image processing circuit 30 is a circuit for displaying an object image from the image pickup device 9 on the image display apparatus 31, and for example, displays an object image during a focusing operation on the image display apparatus 31. As described above, the drive portion 15, the operation member position detection portion 22, the first encoder portion 21 and the second encoder portion 23 are arranged in the lens barrel 10. The lens barrel control portion 24 and the communication portion 25 are also arranged in the lens barrel 10. The lens barrel control portion (control means) 24 constituting a control portion includes a processing apparatus, a storage apparatus, an input/output apparatus and the like, and controls operations of the lens barrel 10 based on a predetermined program in conjunction with the camera body control portion 6. Further, the communication portion 25 is a portion for performing communication with the camera body control portion 6 through the communication portion 8 provided in the camera body 2 by wire communication or wireless communication. Next, operations of the camera system 1 that has the above described configuration are described. In the camera system 1 of the present embodiment, when the camera body is in a state in which the power is turned on, a focus operation mode switching subroutine shown in FIG. 13 is repeatedly performed at a predetermined cycle by the camera body control portion 6. The focus operation mode switching subroutine is a subroutine for judging a focus operation mode that is selected by a user by input of an instruction from among a plurality of focus operation modes of the camera system 1, and switching the focus operation mode of the camera system 1 in accordance with the inputted instruction. Note that the focus operation mode switching subroutine described hereunder is appropriately incorporated into a main routine for causing the camera system 1 to perform photographing operations. The main routine for causing the camera system 1 to perform photographing operations is the same as known technology, and hence a description thereof is omitted. According to the focus operation mode switching subroutine, first, in step S01, the camera body control portion 6 judges whether or not a focus operation mode that the user selected through the focus mode switching operation portion 5 is the autofocus operation mode. If the result of the judgment in step S01 is that the autofocus operation mode is selected, the process shifts to step S03 to switch the operation of the camera system 1 so as to perform an autofocus operation. In step S03, the camera system 1 performs an autofocus operation illustrated in FIG. 14. Although the timing at which the camera system 1 performs an autofocus operation is not particularly limited, as one example according to the present embodiment, as shown in the flowchart of FIG. 14, it is assumed that an autofocus operation is performed when the first release switch 3 a enters an “on” state. When the camera system 1 performs an autofocus operation, the camera body control portion 6 detects a contrast value of an object image based on a signal that is outputted from the image pickup device 9 that is an autofocus sensor portion, and drives the focusing barrel 13 so that the contrast value becomes a maximum value. For example, in the case of the camera system 1 that has a configuration in which the camera body 2 and the lens barrel 10 can be detachably attached to each other as in the present embodiment, at the time of an autofocus operation the camera body control portion 6 issues a command for driving the focusing barrel 13 to perform focusing control based on a signal outputted from the image pickup device 9. In this case, the lens barrel control portion 24 of the lens barrel 10 drives the focusing barrel 13 by means of the drive portion 15 in accordance with the command from the camera body control portion 6 that is inputted through the communication portion 25. Since the focus operation mode switching subroutine is ended by execution of step S03, according to the camera system 1 of the present embodiment, in a case where the autofocus operation mode is selected, an autofocus operation is performed irrespective of whether the rotational operation member 17 is positioned at the first position or the second position. Accordingly, in an autofocus operation, even when the rotational operation member 17 is positioned at the second position and the distance scale 18 a is exposed to outside, the focusing barrel 13 is driven and focusing control is performed irrespective of the numeric value of the distance on the distance scale 18 a that the indicator 16 a indicates. Further, although in this case a configuration is adopted so that “even when the rotational operation member 17 is positioned at the second position and the distance scale 18 a is exposed to outside, the focusing barrel 13 is driven and focusing control is performed irrespective of the numeric value of the distance on the distance scale 18 a that the indicator 16 a indicates”, as a separate embodiment, a configuration may also be adopted so that when the rotational operation member 17 is positioned at the second position and the distance scale 18 a is exposed to outside, the focusing barrel 13 is driven to a position that corresponds to the distance on the distance scale 18 a that the indicator 16 a indicates. In contrast, if the result of the judgment in step S01 is that the autofocus operation mode is not selected, the process shifts to step S02. In step S02, the camera body control portion 6 judges whether or not the rotational operation member 17 is present at the first position based on an output signal of the operation member position detection portion 22. If the result of the judgment in step S02 is that the rotational operation member 17 is present at the first position, the process shifts to step S04. In step S04, the operation of the camera system 1 is switched so as to perform a manual focus operation for focusing control. In step S04, the camera system 1 performs a manual focus operation illustrated in FIG. 15. When performing the manual focus operation, if rotation of the rotational operation member 17 is detected by the first encoder portion 21, the camera system 1 drives the focusing barrel 13 in accordance with the rotational direction and the rotational amount of the rotational operation member 17. Note that at the time of a manual focus operation, the user can observe the focusing state by observing the state of an object image displayed on the image display apparatus 31, and the user operates the rotational operation member 17 based on the observation result. Naturally, also at the time of another focusing operation, the user can similarly observe the focusing state by observing the state of an object image displayed on the image display apparatus 31. For example, in a case where it is detected that the rotational operation member 17 rotated in the clockwise direction when viewing the lens barrel 10 from the rear, the camera system 1 moves the focusing barrel 13 in a direction in which the focusing distance of the optical system member 11 becomes shorter. Further, for example, in a case where it is detected that the rotational operation member 17 rotated in the counterclockwise direction when viewing the lens barrel 10 from the rear, the camera system 1 moves the focusing barrel 13 in a direction in which the focusing distance of the optical system member 11 becomes longer. The movement distance and speed on the focusing barrel 13 in a manual focus operation is determined in accordance with a rotational amount (rotation angle) and a rotational speed (angular speed of rotation) of the rotational operation member 17 detected by the first encoder portion 21. The manual focus operation shown in FIG. 15 is repeatedly executed until a focusing operation other than a manual focus operation is selected by execution of the focus operation mode switching subroutine. That is, the manual focus operation shown in FIG. 15 is repeatedly executed until the autofocus operation mode is selected by means of the focus mode switching operation portion 5 or the rotational operation member 17 is moved from the first position to the second position. In contrast, if the result of the judgment in step S02 is that the rotational operation member 17 is not present at the first position, that is, that the rotational operation member 17 is present at the second position, the process moves to step S05. In step S05, the operation of the camera system 1 is switched so as to perform a distance-specification focus operation for focusing control. In step S05, the camera system 1 performs a distance-specification focus operation illustrated in FIG. 16. When performing the distance-specification focus operation, the camera system 1 drives the focusing barrel 13 to a position that corresponds to a numeric value of a distance on the distance scale 18 a that the indicator 16 a indicates, the distance scale 18 a being calculated based on an output signal from the second encoder portion 23. That is, the focusing barrel 13 is driven by the drive portion 15 to a position that is in accordance with a relative position between the first barrel that is the fixed barrel 14 and indicator display barrel 16 and the second barrel that is the display member 18. Specifically, in the distance-specification focus operation, first, in step S30, the lens barrel control portion 24 reads an output value of the second encoder portion 23. The output value of the second encoder portion 23 is a value that represents an absolute rotational position around the optical axis O of the display member 18 with respect to the base portion 12. Next, in step S31, based on the output value of the second encoder portion 23, the lens barrel control portion 24 calculates a numeric value of the distance on the distance scale 18 a that the indicator 16 a indicates. Conversion to the numeric value of the distance on the distance scale 18 a that the indicator 16 a indicates on the basis of the output value of the second encoder portion 23 is performed, for example, by the lens barrel control portion 24 based on a conversion table that is previously stored. Next, in step S32, the lens barrel control portion 24 moves the focusing barrel 13 so that, based on the output value of the second encoder portion 23, the numeric value of the distance on the distance scale 18 a that the indicator 16 a indicates and a focusing distance of the optical system member 11 match or are in proximity to each other. For example, if the numeric value of the distance on the distance scale 18 a that the indicator 16 a indicates is 3 meters, the lens barrel control portion 24 moves the focusing barrel 13 so that the focusing distance of the optical system member 11 becomes 3 meters. Note that, although it is ideal that the numeric value of the distance on the distance scale 18 a that the indicator 16 a indicates and the focusing distance of the optical system member 11 match, since it is difficult to match the aforementioned numeric value and focusing distance in a case where the resolution of the second encoder portion 23 is low, the focusing barrel 13 is moved so that the numeric value and focusing distance are in proximity to each other. The lens barrel control portion 24 outputs the calculated numeric value of the distance on the distance scale 18 a that the indicator 16 a indicates to the camera body control portion 6 of the camera body 2 through the communication portion 25. The camera body control portion 6 determines an exposure value in accordance with the numeric value that is received, and adds the numeric value to a photographed image as metadata. The distance-specification focus operation shown in FIG. 16 is repeatedly executed until a focusing operation other than the distance-specification focus operation is selected by execution of the focus operation mode switching subroutine. That is, the distance-specification focus operation shown in FIG. 16 is repeatedly executed until the autofocus operation mode is selected by means of the focus mode switching operation portion 5 or the rotational operation member 17 is moved from the second position to the first position. As described above, the camera system 1 of the present embodiment includes the indicator 16 a that is arranged at a fixed position with respect to the base portion 12 that is fixed to the camera body 2, the display member 18 having the distance scale 18 a that is arranged so as to be rotatable with respect to the base portion 12, the second encoder portion 23 that detects a rotational position of the display member 18 with respect to the base portion 12, and the rotational operation member 17 that rotates together with the display member 18. At the time of an autofocus operation, the camera system 1 of the present embodiment performs focusing control based on an output of the image pickup device 9 that is a focusing sensor portion, regardless of the rotational position of the display member 18. Further, at the time of a distance-specification focus operation, the camera system 1 drives the focusing barrel 13 to a position that corresponds to a numeric value of a distance on the distance scale 18 a that the indicator 16 a indicates, the distance scale 18 a being calculated based on an output signal of the second encoder portion 23. That is, in the camera system 1, manual focusing the distance scale is possible. Further, the user can switch between an autofocus operation and a distance-specification focus operation in the camera system 1 by operating the focus mode switching operation portion 5 and the rotational operation member 17. According to the lens barrel 10 of the camera system 1 having the above described configuration, the only member that the drive portion 15 drives is the focusing barrel 13 that holds the focusing lens 11 a, and thus the camera system 1 can be provided in which the number of members driven by the drive portion 15 is decreased and which is lightweight. Thus, according to the present embodiment, the drive portion 15 can be made a small member with a small output, and the lens barrel 10 can be miniaturized. Further, focus operations can be performed rapidly. In addition, according to the present embodiment, a mechanism that transmits a motive force for driving the display member 18 is not required, and a motive force can be transmitted from the drive portion 15 to the focusing barrel 13 as a member to be driven with a simple configuration that has few component parts. Consequently, it is easy to reduce the volume of sound generated when driving the focusing barrel 13 to perform focusing control. Suppressing the sound volume generated at a time of focusing control is preferable, for example, when shooting a moving image. Further, according to the present embodiment, in a case where the manual focus operation mode is selected by the focus mode switching operation portion 5 and the rotational operation member 17 is positioned at the second position, manual focus using the distance scale is possible. Therefore, according to the present embodiment, a photographing technique can be executed in which the focusing distance is previously set manually to a predetermined focusing distance using the distance scale 18 a and photographing is performed rapidly without performing a focusing control operation. For example, according to the present embodiment, because the depth-of-field scale 16 b is provided on the indicator display barrel 16, a user can immediately confirm a focusing object distance by looking at the distance scale 18 a, the indicator 16 a and the depth-of-field scale 16 b that are exposed on the outer circumferential portion of the lens barrel 10. On the other hand, according to the present embodiment, in a case where the manual focus operation mode is selected by means of the focus mode switching operation portion 5 and the rotational operation member 17 is positioned at the first position, it is possible to perform manual focusing as illustrated in FIG. 15 in which focusing control is performed in accordance with rotation of the rotational operation member 17. In the manual focus operation, since the rotational operation member 17 is not engaged with the display member 18, the rotational operation member 17 can rotate without limitation around the optical axis O, and focusing control of the optical system member 11 is performed in accordance with the rotational amount of the rotational operation member 17. Therefore, in a manual focus operation of the camera system 1 of the present embodiment, focusing control can be performed that is finer than in a distance-specification focus operation using the distance scale. As described above, according to the camera system 1 of the present embodiment it is possible to selectively execute a distance-specification focus operation in which rapid photographing is realized as a result of the focusing distance and the depth of field being clearly indicated by means of the indicator 16 a and the distance scale 18 a as rough focusing control (coarse control) through the drive portion 15, and a manual focus operation that performs fine focusing control (fine control) through the drive portion 15 according to the rotational amount of the rotational operation member 17. Further, since it is possible to switch between a distance-specification focus operation and a manual focus operation by merely moving the rotational operation member 17 back and forth, the switching can be executed rapidly and explicitly. Note that fine focusing control of course can also be performed even in a distance-specification focus operation if the resolution of the second encoder portion 23 is sufficiently high. Further, a distance-specification focus operation may be referred to as an “absolute focus operation” and a manual focus operation may be referred to as a “relative focus operation”. Furthermore, according to the present embodiment, in a case where the rotational operation member 17 is positioned at the first position, since the rotational operation member 17 and the display member 18 are not engaged, the display member 18 does not rotate even if the rotational operation member 17 rotates. Therefore, a focusing distance that has been set at the time of a distance-specification focus operation is not changed at the time of a manual focus operation. For example, after the user rotates the display member 18 so that the focusing distance becomes 3 meters at the time of a distance-specification focus operation, even if the rotational operation member 17 is moved to the first position and focusing control by a manual focus operation is performed, the display member 18 does not rotate. Hence, in a case where, thereafter, the user moves the rotational operation member 17 to the second position to perform a distance-specification focus operation, the focusing distance of the optical system member 11 returns to 3 meters. Accordingly, by previously setting the focusing distance to a value that is desired at the time of a distance-specification focus operation, photographing with a desired focusing distance can be quickly performed by merely moving the rotational operation member 17 rearward from a state in which a manual focus operation is being performed. As described above, according to the present embodiment it is possible to perform autofocusing and manual focusing the distance scale, and a small camera system as well as a lens barrel that is used in the camera system can be realized. Second Embodiment Hereunder, a second embodiment of the present invention is described. In the following, only differences with respect to the first embodiment are described, and components that are the same as in the first embodiment are denoted by the same reference symbols and descriptions of such components are appropriately omitted. As shown in FIG. 17, the camera system 1 of the present embodiment includes the camera body 2 and the lens barrel 10 that are integrally formed and can not be detached and attached with respect to each other. In the camera system 1 in which the camera body 2 and the lens barrel 10 are integrated into a single body as in the present embodiment, it is not necessary to arrange a control portion in each of the camera body 2 and the lens barrel 10, and a configuration can be adopted that includes only a single rotational direction and rotational amount control portion. As shown in FIG. 18, the only control portion provided according to the present embodiment is the camera body control portion 6 that is arranged inside the camera body 2. The camera body control portion 6 is configured to be capable of additionally executing the control that the lens barrel control portion performs in the first embodiment. The remaining configuration and operations of the camera system 1 of the present embodiment are the same as in the first embodiment. Accordingly, the camera system 1 of the present embodiment can obtain similar advantageous effects as those of the first embodiment. Third Embodiment Hereunder, a third embodiment of the present invention is described. The third embodiment differs from the first embodiment only with respect to the form of engagement between the rotational operation member 17 and the display member 18. In the following, only differences with respect to the first embodiment are described, and components that are the same as in the first embodiment are denoted by the same reference symbols and descriptions of such components are appropriately omitted. FIG. 19 is a cross-sectional view of the lens barrel 10 in a state where the rotational operation member 17 is positioned at the first position and the rotational operation member 17 and the display member 18 are not engaged. FIG. 20 is a cross-sectional view of the lens barrel 10 in a state where the rotational operation member 17 is positioned at the second position and the rotational operation member 17 and the display member 18 are engaged. According to the present embodiment, engagement between the rotational operation member 17 and the display member 18 is performed by means of friction between the two members. Specifically, in a state where the rotational operation member 17 is positioned at the second position, the rotational operation member 17 is urged relatively rearward with respect to the display member 18 by the urging member 14 c and the ball 14 b. According to the present embodiment, a configuration is adopted so that friction force is generated between the rotational operation member 17 and the display member 18 by pressing the rotational operation member 17 against the display member 18 by means of the aforementioned urging force, and consequently the rotational operation member 17 and the display member 18 rotate together. More specifically, according to the present embodiment, a flange portion 17 h that protrudes to the outer side in the diameter direction is provided on the outer circumferential portion of the inside cylindrical portion 17 b of the rotational operation member 17. On the other hand, in the display member 18 that is arranged on the outer circumference of the inside cylindrical portion 17 b, a flange portion 18 e is provided that protrudes to the inner side in the diameter direction at a position that is further to the rear than the flange portion 17 h. The flange portion 17 h and the flange portion 18 e are arranged at positions that face each other in a separated state in the optical axis O direction when the rotational operation member 17 is positioned at the first position, and that abut when the rotational operation member 17 is positioned at the second position. According to this configuration, in the present embodiment, when the rotational operation member 17 is positioned at the second position, the rotational operation member 17 and the display member 18 engage by means of friction between the flange portion 17 h and the flange portion 18 e. Note that preferably a member or surface treatment that raises a frictional coefficient between the flange portion 17 h and the flange portion 18 e is provided on at least one of the faces of the flange portion 17 h and the flange portion 18 e that face each other. As one example according to the present embodiment, a high friction material 17 i is attached onto a face that faces towards the rear side of the flange portion 17 h. The high friction material 17 i is constituted, for example, by a resin material such as rubber, and is a member in which a frictional coefficient with respect to the flange portion 18 e is higher than that of the flange portion 17 h. In addition, for example, the high friction material 17 i may be a material in which microscopic protrusions are provided that are made of cemented carbide on stainless steel. According to the present embodiment, by using friction to cause the rotational operation member 17 and the display member 18 to engage, the amount by which the rotational position of the display member 18 changes when the rotational operation member 17 is moved between the first position and the second position can be suppressed. The remaining configuration and operations of the camera system 1 of the present embodiment are the same as in the first embodiment or the second embodiment. Accordingly, the camera system of the present embodiment can obtain similar advantageous effects as those of the first embodiment or the second embodiment. Fourth Embodiment Hereunder, a fourth embodiment of the present invention is described. The fourth embodiment differs from the third embodiment only with respect to the form of engagement between the rotational operation member 17 and the display member 18. In the following, only differences with respect to the third embodiment are described, and components that are the same as in the third embodiment are denoted by the same reference symbols and descriptions of such components are appropriately omitted. FIG. 21 is a partial cross-sectional view of the lens barrel 10 in a state where the rotational operation member 17 is positioned at the first position and the rotational operation member 17 and the display member 18 are not engaged. FIG. 22 is a partial cross-sectional view of the lens barrel 10 in a state where the rotational operation member 17 is positioned at the second position and the rotational operation member 17 and the display member 18 are engaged. According to the present embodiment, similarly to the third embodiment, a configuration is adopted so that, in a state where the rotational operation member 17 is positioned at the second position, by pressing the rotational operation member 17 against the display member 18 by means of an urging force that urges the rotational operation member 17 rearward, a frictional force is generated between the rotational operation member 17 and the display member 18, and consequently the rotational operation member 17 and the display member 18 rotate together. According to the present embodiment, a groove portion 17 m is carved in the circumferential direction in an outer circumferential portion of the inside cylindrical portion 17 b of the rotational operation member 17. A C-shaped ring 17 k is fitted into the groove portion 17 m so as to surround the optical axis O. The C-shaped ring 17 k is a substantially C-shaped member having a form in which a part of an annular ring has been notched. The C-shaped ring 17 k has an inner diameter that fits together with a bottom face of the groove portion 17 m with a predetermined clearance therebetween. A tapered portion 17 j is provided in the outer circumferential portion of the C-shaped ring 17 k. The outer diameter of the tapered portion 17 j decreases in the rearward direction in a state in which the C-shaped ring 17 k is fitted in the groove portion 17 m. On the other hand, in an inner circumferential portion of the display member 18, a tapered portion 18 d whose inner diameter decreases in the rearward direction is provided at a position that is further to the rear than the C-shaped ring 17 k. The tapered portion 17 j of the C-shaped ring 17 k and the tapered portion 18 d of the display member 18 are arranged so as to face each other in a separated state in the optical axis O direction when the rotational operation member 17 is positioned at the first position, and so as to abut against each other when the rotational operation member 17 is positioned at the second position. By means of this configuration, according to the present embodiment, when the rotational operation member 17 is positioned at the second position, since the C-shaped ring 17 k is urged rearward in a state in which the tapered portion 17 j and tapered portion 18 d abut against each other, the C-shaped ring 17 k is compressed towards the inner side in the diameter direction so that the inner diameter thereof decreases. Further, the shape of the inner circumferential face of the C-shaped ring 17 k is changed to the direction of the bottom face of the groove portion 17 m by elasticity. Therefore, according to the present embodiment, the rotational operation member 17 and the display member 18 engage by means of friction between the tapered portion 17 j and the tapered portion 18 d and friction between the inner circumferential face of the C-shaped ring 17 k and the bottom face of the groove portion 17 m. According to the present embodiment, a contact area for contact between the tapered portions is large and a strong fastening force is obtained. Furthermore, since a part in the circumferential direction of the C-shaped ring 17 k is notched, release of fastening can be performed with ease. Note that the configuration may also be one in which, to restrict relative rotation of the C-shaped ring 17 k with respect to the inside cylindrical portion 17 b, a key is provided in one of the C-shaped ring 17 k and the inside cylindrical portion 17 b, and a key groove that engages with the key is provided in the other of the C-shaped ring 17 k and the inside cylindrical portion 17 b. According to the present embodiment, by engaging the rotational operation member 17 and the display member 18 by means of friction therebetween, the amount by which the rotational position of the display member 18 changes when the rotational operation member 17 is moved between the first position and the second position can be suppressed. The remaining configuration and operations of the camera system 1 of the present embodiment are the same as in the first embodiment or the second embodiment. Accordingly, the camera system 1 of the present embodiment can obtain similar advantageous effects as those of the first embodiment or the second embodiment. Fifth Embodiment Hereunder, a fifth embodiment of the present invention is described. The fifth embodiment differs from the first embodiment only with respect to the focus operation mode switching subroutine. In the following, only differences with respect to the first embodiment are described, and components that are the same as in the first embodiment are denoted by the same reference symbols and descriptions of such components are appropriately omitted. FIG. 23 is a flowchart of a focus operation mode switching subroutine of the camera system according to the fifth embodiment. Note that the focus operation mode switching subroutine described hereunder is appropriately incorporated into a main routine for causing the camera system 1 to perform photographing operations. The main routine for causing the camera system 1 to perform photographing operations is the same as known technology, and hence a description thereof is omitted. According to the present embodiment, first, in step S51, it is judged whether or not the rotational operation member 17 is present at the first position based on an output signal of the operation member position detection portion 22. If the result of the judgment in step S51 is that the rotational operation member 17 is not present at the first position, the process shifts to step S55, and the operation of the camera system 1 is switched so as to perform a distance-specification focus operation as shown in FIG. 16. In contrast, if the result of the judgment in step S51 is that the rotational operation member 17 is present at the first position, the process shifts to step S52. In step S52, it is judged whether or not the focus operation mode that the user selected through the focus mode switching operation portion 5 is the autofocus operation mode. If the result of the judgment in step S52 is that the autofocus operation mode is selected, the process shifts to step S53, and the operation of the camera system 1 is switched so as to perform an autofocus operation. In contrast, if the result of the judgment in step S52 is that the autofocus operation mode has not been selected, the operation of the camera system 1 is switched so as to perform a manual focus operation as shown in FIG. 15. As described above, according to the focus operation mode switching subroutine of the present embodiment, first, a judgment is made to determine which position among the first position and the second position the rotational operation member 17 is positioned at. Consequently, if the rotational operation member 17 is positioned at the second position, the focus operation mode switching subroutine ends by execution of step S55. Therefore, according to the camera system 1 of the present embodiment, regardless of whether a focus operation mode that is selected by the user through the focus mode switching operation portion 5 is the autofocus operation mode or the manual focus operation mode, if the rotational operation member 17 is positioned at the second position a distance-specification focus operation is performed. That is, according to the present embodiment, when the user moves the rotational operation member 17 to the second position, the camera system 1 always executes a distance-specification focus operation. Therefore, according to the present embodiment, it is possible to immediately switch from a state in which an autofocus operation is being performed to a distance-specification focus operation that uses the distance scale. Further, as described in the above embodiments, when the rotational operation member 17 is positioned at the first position, the rotational operation member 17 and the display member 18 are not engaged, and the display member 18 does not rotate even if the rotational operation member 17 rotates. That is, a focusing distance that was set at the time of a distance-specification focus operation does not change at the time of a manual focus operation. For example, if the focusing distance is set to 3 meters when the rotational operation member 17 is positioned at the second position, and thereafter the rotational operation member 17 is temporarily moved to the first position and subsequently returned to the second position, the focusing distance remains 3 meters irrespective of whether or not the rotational operation member 17 was rotated and whether or not an autofocus operation was performed at the first position. Therefore, in a case where the rotational operation member 17 is positioned at the second position, based on the display of the depth-of-field scale 16 b, if the focusing distance is set in advance so that a photographing technique that is referred to as a so-called “pan-focus” can be performed, it is possible to rapidly switch from a state in which an autofocus operation or a manual focus operation is being performed to a state of performing pan-focus photographing by merely moving the rotational operation member 17 to the second position. Thus, the convenience for the user is enhanced. Conversely, it is also possible to rapidly switch from a state in which pan-focus photographing is being performed at the time of a distance-specification focus operation to a state of performing an autofocus operation or a manual focus operation by merely moving the rotational operation member 17 to the first position. The configuration other than the above described focus operation mode switching subroutine of the present embodiment is the same as in the first embodiment. Accordingly, the camera system of the present embodiment can obtain similar advantageous effects as those of the first to fourth embodiments that are described above. That is, as described in the foregoing embodiments, the number of members driven by the drive portion 15 can be reduced and the camera system can be made lightweight. Thus, according to the present embodiment, the drive portion 15 can be made a small member with a small output, and the lens barrel 10 can be miniaturized. Further, focus operations can be rapidly performed. Note that, although in the above described embodiments the indicator 16 a is provided on the indicator display barrel 16 as a first barrel that is fixed, and the distance scale 18 a is provided on the display member 18 as a second barrel that is rotatable, the present invention is not limited to this form. For example, as shown in FIG. 24 and FIG. 25, a form may also be adopted in which a distance scale 16 aa is provided on the fixed first barrel (denoted by reference numeral 16) and an indicator 18 aa is provided on the rotatable second barrel (denoted by reference numeral 18). FIG. 24 illustrates a state in which the rotational operation member 17 is positioned at the first position and a display of diaphragm values cannot be seen. FIG. 25 illustrates a state in which the rotational operation member 17 is positioned at the second position and the display of diaphragm values can be seen. That is, the first barrel that is fixed includes one of the indicator and the distance scale, and the second barrel that is rotatable displays the other of the indicator and the distance scale. In the case of this form also, the lens barrel and the camera system have exactly the same functions as in the above described embodiments. Note that it may be said that a display function that includes a member in which an indicator is provided and a member on which a distance scale is provided is display means. Note that the present invention is not limited to the above described embodiments, but may be suitably changed without departing from the spirit or concept of the invention readable from the appended claims and the entire specification, and a camera system and a lens barrel with such changes are also included in the technical scope of the present invention. What is claimed is: 1. A lens barrel, comprising: a first barrel; a second barrel that is capable of relative rotation around an optical axis with respect to the first barrel; a drive source; a focusing lens; a drive portion that is driven by the drive source, and that drives the focusing lens in an optical axis direction; a rotational operation member that is a member operated for driving the focusing lens in the optical axis direction, and that can move in the optical axis direction to a first position at which the rotational operation member covers the second barrel and to a second position at which the rotational operation member causes the second barrel to be exposed to outside, and which is configured to be capable of rotation around the optical axis at the first position and the second position, respectively; and engagement means that, when the rotational operation member is at the second position, causes the second barrel and the rotational operation member to engage with each other and rotates the second barrel accompanying rotation of the rotational operation member, and when the rotational operation member is at the first position, disengages the second barrel and the rotational operation member from each other; wherein when the rotational operation member is at the first position, the drive portion drives the focusing lens in accordance with rotation of the rotational operation member, and when the rotational operation member is at the second position, the drive portion drives the focusing lens in accordance with rotation of the second barrel that is engaged with the rotational operation member. 2. The lens barrel according to claim 1, further comprising: rotation restriction means configured so as to restrict relative rotation of the second barrel with respect to the first barrel to a predetermined range; wherein when the rotational operation member is at the first position, the rotational operation member is not subjected to rotational restriction because the rotational operation member is not engaged with the second barrel, and when the rotational operation member is at the second position, because the rotational operation member is engaged with the second barrel, rotation of the rotational operation member around the optical axis is restricted to the predetermined range by the rotation restriction means. 3. The lens barrel according to claim 2, further comprising: a first encoder portion configured so as to detect a rotational amount and a rotational speed of the rotational operation member when the rotational operation member is at the first position; a second encoder portion configured so as to detect a rotational position of the second barrel when the rotational operation member is at the second position; and an operation member position detection portion configured so as to detect a position in the optical axis direction of the rotational operation member; wherein when the rotational operation member is at the first position, the focusing lens is driven by the drive portion to an arbitrary position in the optical axis direction in accordance with an output of the first encoder portion, and when the rotational operation member is at the second position, the focusing lens is driven by the drive portion in accordance with an output of the second encoder portion. 4. The lens barrel according to claim 3, wherein one of an indicator and a distance scale is displayed on an outer circumferential face of the first barrel, and the other of the indicator and the distance scale is displayed on an outer circumferential face of the second barrel. 5. The lens barrel according to claim 2, wherein one of an indicator and a distance scale is displayed on an outer circumferential face of the first barrel, and the other of the indicator and the distance scale is displayed on an outer circumferential face of the second barrel. 6. The lens barrel according to claim 1, further comprising: a first encoder portion configured so as to detect a rotational amount and a rotational speed of the rotational operation member when the rotational operation member is at the first position; a second encoder portion configured so as to detect a rotational position of the second barrel when the rotational operation member is at the second position; and an operation member position detection portion configured so as to detect a position in the optical axis direction of the rotational operation member; wherein when the rotational operation member is at the first position, the focusing lens is driven by the drive portion to an arbitrary position in the optical axis direction in accordance with an output of the first encoder portion, and when the rotational operation member is at the second position, the focusing lens is driven by the drive portion in accordance with an output of the second encoder portion. 7. The lens barrel according to claim 6, wherein one of an indicator and a distance scale is displayed on an outer circumferential face of the first barrel, and the other of the indicator and the distance scale is displayed on an outer circumferential face of the second barrel. 8. The lens barrel according to claim 1, wherein one of an indicator and a distance scale is displayed on an outer circumferential face of the first barrel, and the other of the indicator and the distance scale is displayed on an outer circumferential face of the second barrel. 9. The lens barrel according to claim 8, wherein the indicator is displayed in a manner in which the indicator is interposed between at least one pair of diaphragm values displaying a depth of field of an identical numeric value. 10. A camera system, comprising: a first barrel; a second barrel that is capable of relative rotation around an optical axis with respect to the first barrel; a drive source; a focusing lens; a drive portion that is driven by the drive source, and that drives the focusing lens in an optical axis direction; a rotational operation member that is a member operated for driving the focusing lens in the optical axis direction, and that can move in the optical axis direction to a first position at which the rotational operation member covers the second barrel and to a second position at which the rotational operation member causes the second barrel to be exposed to outside, and which is configured to be capable of rotation around the optical axis at the first position and the second position, respectively; and engagement means that, when the rotational operation member is at the second position, causes the second barrel and the rotational operation member to engage with each other and rotates the second barrel accompanying rotation of the rotational operation member, and when the rotational operation member is at the first position, disengages the second barrel and the rotational operation member from each other; wherein, when the rotational operation member is at the first position, the drive portion drives the focusing lens in accordance with rotation of the rotational operation member, and when the rotational operation member is at the second position, the drive portion drives the focusing lens in accordance with rotation of the second barrel that is engaged with the rotational operation member. 11. The camera system according to claim 10, further comprising: rotation restriction means configured so as to restrict relative rotation of the second barrel with respect to the first barrel to a predetermined range; wherein when the rotational operation member is at the first position, the rotational operation member is not subjected to rotational restriction because the rotational operation member is not engaged with the second barrel, and when the rotational operation member is at the second position, because the rotational operation member is engaged with the second barrel, rotation of the rotational operation member around the optical axis is restricted to the predetermined range by the rotation restriction means. 12. The camera system according to claim 11, further comprising: a first encoder portion configured so as to detect a rotational amount and a rotational speed of the rotational operation member when the rotational operation member is at the first position; a second encoder portion configured so as to detect a rotational position of the second barrel when the rotational operation member is at the second position; an operation member position detection portion configured so as to detect a position in the optical axis direction of the rotational operation member; and a control portion configured so as to drive the focusing lens to an arbitrary position in the optical axis direction by means of the drive portion in accordance with an output of the first encoder portion when the rotational operation member is at the first position, and to drive the focusing lens by means of the drive portion in accordance with an output of the second encoder portion when the rotational operation member is at the second position. 13. The camera system according to claim 12, wherein one of an indicator and a distance scale is displayed on an outer circumferential face of the first barrel, and the other of the indicator and the distance scale is displayed on an outer circumferential face of the second barrel. 14. The camera system according to claim 11, wherein one of an indicator and a distance scale is displayed on an outer circumferential face of the first barrel, and the other of the indicator and the distance scale is displayed on an outer circumferential face of the second barrel. 15. The camera system according to claim 10, further comprising: a first encoder portion configured so as to detect a rotational amount and a rotational speed of the rotational operation member when the rotational operation member is at the first position; a second encoder portion configured so as to detect a rotational position of the second barrel when the rotational operation member is at the second position; an operation member position detection portion configured so as to detect a position in the optical axis direction of the rotational operation member; and a control portion configured so as to drive the focusing lens to an arbitrary position in the optical axis direction by means of the drive portion in accordance with an output of the first encoder portion when the rotational operation member is at the first position, and to drive the focusing lens by means of the drive portion in accordance with an output of the second encoder portion when the rotational operation member is at the second position. 16. The camera system according to claim 15, wherein one of an indicator and a distance scale is displayed on an outer circumferential face of the first barrel, and the other of the indicator and the distance scale is displayed on an outer circumferential face of the second barrel. 17. The camera system according to claim 10, wherein one of an indicator and a distance scale is displayed on an outer circumferential face of the first barrel, and the other of the indicator and the distance scale is displayed on an outer circumferential face of the second barrel. 18. The camera system according to claim 17, wherein the indicator is displayed in a manner in which the indicator is interposed between at least one pair of diaphragm values displaying a depth of field of an identical numeric value.
Pirates Cause Climate Change (Hoax) Research Problems Couldn’t resist this story Piracy in the Indian Ocean is hampering the efforts of climate change researchers to the point where they have had to call in the Australian Navy. Pirate activity off the coast of Somalia has increased dramatically over recent years, so much so that a quarter of the Indian Ocean is now considered a “no-go” area. (big snip) The pirates are hampering research that is crucial in the global understanding of weather forecasting. The CSIRO is part of an international research team which uses thousands of robotic floats. They are dropped into the ocean to provide near real time observations and for the last year, scientists have not been able to access the data they need. Fortunately, the US and Aussie navies are wasting time, resources, and money to help the climate wankers gather their data, which well then be “adjusted” to conform to the pre-conceived notions of the True Believers, and they will refuse to show the raw data, which is funded for by tax payers. Dead men tell no tales: climate alarmists feel the same way about their raw data. But, wait a second, if you read the story (and notice the end of the 3rd paragraph in the excerpt), you notice that most of the research is about…..weather forecasting. So, according to the media, the weather is now “climate change.” If you liked my post, feel free to subscribe to my rss feeds. Both comments and trackbacks are currently closed Bad Behavior has blocked 8120 access attempts in the last 7 days.
Windy Starkiller "Windy's nothing but a follower&hellip;" - Biggs Darklighter Biography Windy saw Luke again when he returned briefly to Tatooine when he was trying to hide from X-7. Behind the scenes In the children's book Luke's Fate by Jim Thomas, Windy is depicted as a young girl. Appearances * Luke Skywalker's Walkabout * Adventure in Beggar's Canyon * Luke's Fate * Star Wars: Empire 8: Darklighter, Part 1 * Star Wars radio drama * Star Wars Journal: The Fight for Justice * Star Wars Episode IV: A New Hope novel * Star Wars: Episode IV A New Hope * Rebel Force: Renegade * Star Wars 17: Crucible * Star Wars: Rogue Squadron III: Rebel Strike * The Life and Legend of Obi-Wan Kenobi * A New Hope: The Life of Luke Skywalker * The Life and Legend of Obi-Wan Kenobi * A New Hope: The Life of Luke Skywalker
Template:Table graph/doc Usage This template is a wrapper to Module:Table graph, which itself is a wrapper to call Module:Chart and Module:Graph using wikitables as input data.
## TextDrawable Fork of the TextDrawable library originally found [here](https://github.com/amulyakhare/TextDrawable). It has been reworked to be more simplified as well as had several features. ### Features - Oval, Rectangle, and Rounded Rectangle shapes - Icon support - Highly customizable (border color, border thickness, type face, text size, text color, and much more) - Ability to create a Bitmap object from the Drawable - Kotlin support - Builder pattern for Java ### Using TextDrawable To use TextDrawable, simply create an object with the desired parameters. Each value has a default value, so you can supply as many or as few as you want by using kotlins [named arguements](https://www.programiz.com/kotlin-programming/default-named-arguments). At a minimum, a `text` or an `icon` should be specified. ```kotlin val textDrawable = TextDrawable(shape = TextDrawable.DRAWABLE_SHAPE_OVAL, desiredHeight = 250, desiredWidth = 250, color = Color.BLUE, textColor = Color.RED, text = "A") ``` If using java, you can use the [TextDrawableBuilder](https://github.com/Kennyc1012/TextDrawable/blob/master/library/src/main/java/com/kennyc/textdrawable/TextDrawableBuilder.kt) class which used the Builder pattern for easy creation. ```java TextDrawable drawable = new TextDrawableBuilder(TextDrawable.DRAWABLE_SHAPE_OVAL) .setHeight(250) .setWidth(250) .setColor(Color.BLUE) .setTextColor(Color.RED) .setText("A") .build() ``` ### Customizable fields You can customize the follow attributes of a TextDrawable to fit your needs. Each value has a default value. ```kotlin // The Shape the drawable should take @DrawableShape val shape: Int = TextDrawable.DRAWABLE_SHAPE_RECT, // The solid fill color of the drawable @ColorInt var color: Int = Color.GRAY, // The color of the text for the drawable. Will be ignored if an icon is set @ColorInt var textColor: Int = Color.WHITE, // The corner radius for the drawable. Will be ignored if the shape is not a DRAWABLE_SHAPE_ROUND_RECT var cornerRadius: Float = 0F, // The text size for the text of the drawable. Will be ignored if an icon is set var textSize: Float = 0f, // The desired height of the drawable var desiredHeight: Int = -1, // The desired width of the drawable var desiredWidth: Int = -1, // The border thickness of the drawable var borderThickness: Float = 0F, // The color of the border. Will be ignored if borderThickness is <= 0 @ColorInt var borderColor: Int = color, // The typeface to use for the text of the drawable. Will be ignored if an icon is set var typeFace: Typeface = Typeface.DEFAULT, // The text to use for the drawable. Will be ignored if an icon is set var text: String? = null, // The icon to use for the drawable. Will override any text that may have been set var icon: Bitmap? = null ``` ### Including in your project Add repository ```groovy allprojects { repositories { ... maven { url "https://jitpack.io" } } } ``` Add dependency ```groovy dependencies { implementation 'com.github.Kennyc1012:TextDrawable:2.0.1' } ```
PRODUCTS, ADVENTITIOUS. viscid, and free from solid particles of any kind. And the evolution of all kinds of blas tema proceeds in the same manner, as far as hitherto ascertained, until the formation of cells is effected. The successive steps may in general terms be stated to be increase of viscidity, formation of granules, of nuclei, and of cells.* And these various steps cannot be accomplished except in blastema in contact with living animal structure, a blastema loses its potentiality either by the death of the textures amid which it is evolved, or by its removal from among them. We place no confidence in the experiments upon which a statement of certain exceptions to this law has been founded, f No matter what be the ultimate destiny of the blastema, the process of its evolution is conducted, then, on the same principle ; but the development of cells sets a limit to this identity of process. The vital qualities of the cells differ, and affect the function and end of these, in three principal modes ; and, accord ing as each of these prevails in any given blastema, will the product generated (solid or semi-solid) present peculiar characters. These three modes of cell may be described as fol lows. First, the cells once developed may be alto gether inapt for life, incapable either of under going such changes in physical, chemical, and vital constitution as shall qualify them for sus taining a permanent existence, or of generating the elements of new cells previous to their own destruction. They are consequently acted upon physically and chemically by the sur rounding materials : they are either dissolved by the fluid with which they are associated ; or, disintegrated and broken down into non productive granular matter, they lose all trace of the attributes of organization. Cells of this kind may be termed evanescent and retro grading. Secondly, the cells may be deficient in the faculty of permanency requisite for the forma tion of tissue ; while, on the other hand, they possess the power of generating the elements of new cells (or of causing indirectly the generation of those elements) previously to their own disappearance, cells endowed in turn with a similar generative force. These cells consequently present the characters of the formative stages of evolution, never those of perfectly evolved structure. To this kind of cells the title non-permanent and vegetative may be applied. Thirdly, the cells may possess an inherent force, qualifying them to pass through the necessary steps towards the formation of structure more or less closely resembling the natural tissues, and in this evolved condition they are destined permanently to remain. These cells appear likewise to be destitute of * The production of fibres without the interven tion of a cell-stage (an exceptional phenomenon, if the entire system be taken in view) will find its place elsewhere. the power either of generating or of indirectly causing the generation of the elements of new cells similar to themselves ; they may there fore be termed permanent and non-vegetative cells. From cells of one or other of these three kinds all blastemal formations are produced. Formations produced from the evanescent cell are non-stromal, and may be termed deposits ; those from the vegetative cell are stromal, and may be termed groivths ; those from the permanent cell are stromal, and may be termed pseudo-tissues, ORDER I. DEPOSITS. Deposits are deficient in the characters of texture ; they possess neither permanent fibre, nor definite arrangement of parts, septa, nor loculi ; and are insusceptible of vascularization. They tend to produce eliminatory action and ulceration in the seats they occupy ; and are prone to appear, mainly through the influence of so-called " diathesis," in several parts of the frame simultaneously or consecutively. The substance of all deposits is per sc non-inocu lable ; we say all, because, though the point has not been, to our knowledge, experimentally tested in regard of the typhous and diphtheritic species, there can be little doubt that the proposition applies to them as to the others. But certain varieties of one genus of deposit (pus) are, on the contrary, readily inoculable through the agency of certain associated principles called viruses (see Pus). In such coses, be it observed, the propagation of the disease in no wise depends on the cell of the fluid. In the order Deposits (constituting transi tion products from the non-plastic protein precipitates to formations of higher attributes) we place the following genera : typhous, tuber culous, purulent, melanic, and diphtheritic pro ducts. 1. TYPHOUS DEPOSIT. In the form of continued fever anatomically characterized by alteration of structure in the glandular textures of the small intestine, a peculiar substance of new formation (as first accurately described by M. Louis*) is dis covered in the cellular membrane between the mucous and muscular coats of the patches of agminated glands of Peyer. The propor tion of cases of continued fever of intestinal type in which this deposit occurs, has been differently estimated from less than one third of the cases to nearly the entire number. We have found this matter homogeneous in aspect, of pinkish or yellowish hue (the former acci dental), and from a sixth to a quarter of an inch thick ; we have always seen it more or less firm and tenacious, and never succeeded in catching it in its earlier stage of fluid blastema. Examined under the microscope by Bohmf it appeared utterly destitute of * Roedercr and "Wagler (De morbo mucoso, p. 332.) first noticed this siibstance thus : " ne semel tamen elevatos [folliculos coagmentatos] et materia mucosa obscure cinerea rcpertos vidimus."
Webview editor z-index Testing #85991 Can the webview be forced beneath the rest of the workbench? MacOS, correct? Any repo steps? Looks like same root cause as #82653 Unfortunately we have to keep webview elements outside of the main dom layout to avoid reparenting them
Page:The Cambridge History of American Literature, v2.djvu/182 i66 Magazines, Annuals, and Gift-books gratuitously for two years, reluctantly resigned the editorship, and Emerson as reluctantly took it up, noting in his diary: "I wish it to live, but I do not wish to be its life. Neither do I like to put it into the hands of the Humanity and Reform Men, because they trample on letters and poetry; nor in the hands of the scholars, for they are dead and dry." After spending much time and some money Emerson too felt forced to abandon the undertaking, and The Dial came to an end with the close of the fourth volume. Among contributors other than those already noted were C. P. Cranch, George Ripley, William H. Channing, William EUery Channing, Theodore Parker, James Freeman Clarke, James Russell Lowell, Charles A. Dana, and Jones Very. In its own day The Dial was re- garded reverently by a few, but by the great mass of readers it was ignored or taken as a joke. A later generation still finds many things in its pages amusing but has come to re- cognize it as the best single exponent of New England Transcen- dentalism, and of the peculiar aspects of culture that accom- panied that movement. ' Although The Dial was unique, several earlier and later Boston magazines appealed to much the same constituency. In 1838 the Reverend Orestes A. Brownson began to issue The Boston Quarterly Review, and the next year he urged the Transcendentalists to contribute to his journal rather than to found The Dial. After five years The Boston Quarterly Review was merged with The Democratic Review of New York. A more important periodical was Brownson' s Quarterly Review, founded in 1844 after the editor had been converted to the Roman Catholic faith. An immediate successor of The Dial was The Harbinger, established in 1845 by the members of the Brook Farm community as an organ of Fourierism. From 1847 to 1850 the Reverend Theodore Parker, one of the most virile of the Transcendental group, conducted The Massachusetts Quarterly Review, which he humorously characterized as "The Dial with a beard." One of the earliest of the popular New York magazines to attain permanency was The Knickerbocker.^ This first « See also Book II, Chap. viii. the earlier issues was Knickerbacker. * Owing to some whim of Hoffman, the first editor, the spelling adopted for
Aryanism Aryanism is a form of fascism that originated in Germany in the 1930s. Tenets Genocide Post-war efforts to eliminate Aryanism Discrimination of Odinists as a result of Aryanist appropriation of Odinist symbols
推理平台 大佬,您好,请问下BasicVSR++是在哪个平台下获得推理时间的?如果在CPU下,有测过大概时间吗?谢谢! Excuse me, BasicVSR++ is in which platform to get reasoning time? If you're on the CPU, have you measured the approximate time? Thank you very much! Hello, we did not test the time on CPU. I used an V100 GPU to time the model. Closing due to inactivity, please reopen if there are any further problems.
Wikipedia:Featured article review/Chelsea F.C./archive1 * The following is an archived discussion of a featured article review. Please do not modify it. Further comments should be made on the article's talk page or at Wikipedia talk:Featured article review). No further edits should be made to this page. The article was kept by User:Marskell 12:11, 13 October 2008. Chelsea F.C. * Notified WikiProject Football. Marskell (talk) 10:16, 30 September 2008 (UTC) I think that this article shouldn't be neither a featured one nor a good one because: the history section suffers of recentism, the lead section is too short, the crest section has a non-free image gallery, and so on. Hadrianos1990 (talk) 07:55, 27 September 2008 (UTC) * Looking at the archive, this passed FAC way too easily. But I'm not really surprised since it really is happening way too often, based on popularity etc. Anyway, I'm amongst the biggest supporters of Chelsea, but this article is indeed poory structured. The lead is way too short, and some other raised issues stand predominantly. I will do my brief best to fix up any issues, but I really can't promise anything. Domiy (talk) 09:03, 27 September 2008 (UTC) * Not liking the FAC is not a reason for removal, and assumes bad faith on the part of the reviewers. Oldelpaso (talk) 10:34, 27 September 2008 (UTC) * Please notify significant contributors and the FAC nominator of this FAR. User:SteveO does a lot of work on this article, for example. Oldelpaso (talk) 10:34, 27 September 2008 (UTC) * Please follow the instructions at WP:FAR, using Chelsea F.C., to do the notifications and post them back to this FAR, as in the sample at Featured article review/Felix the Cat. Sandy Georgia (Talk) 20:30, 27 September 2008 (UTC) Glancing quickly at this article, I see some minor issues, not bad for an article promoted almost two years ago, but I wonder if we aren't seeing some pointy-ness going on in the Football nominations. Was there any attempt to resolve these concerns on the article talk page? Folks, there are some really old and out-of-compliance FAs out there, and clogging up FAR with articles that could be addressed via talk doesn't bode well. If this becomes a trend, we may need to alter FAR instructions to require prior attempts to work things out on talk before bringing articles that could easily be fixed to FAR. Sandy Georgia (Talk) 21:27, 27 September 2008 (UTC) Agree with Sandy about this being in pretty good shape, but not that happy with the non-free gallery. Not sure five old logos are needed, not to mention a second use of the current one. The recentism issue is difficult because Chelsea have acheived most of their success since 1997. Naturally, the history section will be slanted toward this period. A little more on the club from before 1950 wouldn't hurt, though. As for the lead, some more on their history could be included. I said that for Real Madrid and it's only fair for me to repeat that here. But overall there have been many worse articles through here and I think with a little work this can be kept. Giants2008 ( 17-14 ) 00:50, 29 September 2008 (UTC) Per the comments above, I've expanded the lead and tried to reduce the recentism in the history section. As for the images, my suggestion is to remove the two variants of the current crest and the 52-53 initials one, which would leave just the main three. It's possible that the 1905 and 1955 crests are out of copyright anyway. SteveO (talk) 00:49, 30 September 2008 (UTC) The notifications still have not been done. Per the instrutions at the top of WP:FAR, pls notify with Chelsea F.C. and post them back to here as in the sample at Featured article review/Felix the Cat. Sandy Georgia (Talk) 04:05, 30 September 2008 (UTC) * Comment - History section and lead look absolutely fine to me, gallery of crests has gone, and what exactly does "and so on" cover? Looks a lot like a pointy nomination to me..... -- ChrisTheDude (talk) 10:18, 30 September 2008 (UTC) * I felt this FAR was premature, but the non-free image gallery has been remove,, so I am happy. I would give this change a few days to settle as the consensus version before considering closing the review. Fasach Nua (talk) 11:09, 30 September 2008 (UTC) * I have to say I'm not really that keen on the article's "History" section as it stands. There's currently about one paragraph for each of the last six decades (slightly less, actually, as the 70s and 80s seem to be sharing a paragraph), and the paragraphs get progressively longer as they approach more recent events. Even worse, there is a single paragraph that covers the entire period from the club's foundation in 1905 to the start of the 1950s! 45 years covered by one paragraph and then five whole paragraphs for the next 58 years? Reeks of WP:RECENT to me. You could argue that most of Chelsea's history has come in the last five years (and in terms of trophies, it has), but I'm sure there were plenty of noteworthy events that took place between 1905 and 1950 that could be expanded upon. If this was currently an FAC candidate, I would be in opposition. – PeeJay 15:41, 30 September 2008 (UTC) This didn't get comment after the first few days. Normally, it would go to FARC now but I am with the group that views this as premature and probably unnecessary. I don't see a huge issue wrt to recentism and the crests are gone. Keeping. Marskell (talk) 09:16, 13 October 2008 (UTC)
Small Blobs silently stored as ShortBlobs - Need option to force full Blob Maybe there is an option but someone will have to point it out to me: how can I ensure that my Blobs are stored as full Blob objects and not ShortBlobs? The issue for me is that I'm writing entities to the datastore using gcloud-java and then reading them with Objectify. When reading I see the error: Expected class com.google.appengine.api.datastore.Blob, got class com.google.appengine.api.datastore.ShortBlob: <ShortBlob: 247 bytes> I'm inserting my Blobs like this: entityBuilder.set("compressedText", BlobValue.builder(Blob.copyFrom(compressedtext)).build()); where compressedtext is a byte[] Also just FYI I've posed this as a potential Objectify change: https://github.com/objectify/objectify/issues/284 From the Google Cloud Datastore perspective there is only one type of Blob which is a Byte String. I think Objectify need to map/handle ShortBlob in a similar way they do for Blob. BTW, a similar concept should apply to Text and String (Google Cloud Datastore has only one field for it, stringValue). Thanks I suppose I just assumed that ShortBlob was a datastore primitive (this page). Jeff has very graciously agreed to fix this in Objectify. I think we can close this out.
import { CI, map, MemoizedIterator, MI } from "../../src/collection/Iterator.js" declare const StartTest : any StartTest(t => { t.it('Should be able to use chained iterators', t => { const a : Set<string> = new Set([ '1', '12', '123' ]) const iter1 : Iterable<string> = CI(a) const iter2 : Iterable<number> = map(a, el => el.length) t.isDeeply(Array.from(iter1), [ '1', '12', '123' ]) t.isDeeply(Array.from(iter2), [ 1, 2, 3 ]) }) t.it('Should be able to use memoized iterators', t => { const a : Set<string> = new Set([ '1', '12', '123' ]) const iter1 : MemoizedIterator<string> = MI(a) const iter2 : Iterable<number> = iter1.map(el => el.length) t.isDeeply(Array.from(iter1), [ '1', '12', '123' ]) t.isDeeply(Array.from(iter1), [ '1', '12', '123' ]) t.isDeeply(Array.from(iter2), [ 1, 2, 3 ]) }) t.it('Should be able to use iterators, derived from the memoized iterator in any order', t => { const a : Set<string> = new Set([ '1', '12', '123' ]) const iter1 : MemoizedIterator<string> = MI(a) const iter2 : Iterable<number> = iter1.map(el => el.length) const iter3 : Iterable<string> = iter1.map(el => el.repeat(2)) const iterator2 = iter2[ Symbol.iterator ]() const iterator3 = iter3[ Symbol.iterator ]() t.isDeeply( [ iterator2.next().value, iterator3.next().value ], [ 1, '11' ] ) // opposite order t.isDeeply( [ iterator3.next().value, iterator2.next().value ], [ '1212', 2 ] ) t.isDeeply( [ iterator2.next().value, iterator3.next().value ], [ 3, '123123' ] ) t.isDeeply( [ iterator3.next().done, iterator2.next().done ], [ true, true ] ) }) })
package edu.southwestern.evolution.lineage; public class JumpPoint { int objective; double jump; Offspring individual; int generation; long comparisonId; public JumpPoint(int objective, double jump, Offspring individual, int generation, boolean firstParent) { this(objective, jump, individual, generation, firstParent ? individual.parentId1 : individual.parentId2); } public JumpPoint(int objective, double jump, Offspring individual, int generation, long comparisonId) { this.objective = objective; this.jump = jump; this.individual = individual; this.generation = generation; this.comparisonId = comparisonId; } @Override public String toString() { String result = "Obj. " + objective + " is " + jump + " higher in " + individual.offspringId + " than in " + comparisonId + " at Gen. " + generation; return result; } @Override public boolean equals(Object o) { if (o instanceof JumpPoint) { JumpPoint jp = (JumpPoint) o; return jp.objective == objective && jp.individual.offspringId == individual.offspringId && jp.generation == generation && jp.comparisonId == comparisonId; } return false; } @Override public int hashCode() { int hash = 7; hash = 79 * hash + this.objective; hash = 79 * hash + (this.individual != null ? this.individual.hashCode() : 0); hash = 79 * hash + this.generation; hash = 79 * hash + (int) (this.comparisonId ^ (this.comparisonId >>> 32)); return hash; } }
In the Matter of Margaret L. Albanese, Respondent, v Andrew Albanese, Sr., Appellant. [844 NYS2d 150] Spain, J. Appeal from an order of the Family Court of Broome County (Pines, J.), entered July 12, 2006, which granted petitioner’s application, in a proceeding pursuant to Family Ct Act article 6, for custody of the parties’ child. Petitioner and respondent have been married approximately 20 years and are the parents of a son born in 1997 (hereinafter the child). In early 2004, apparently due to the parents’ drug use, the child was removed by the Chenango County Department of Social Services (hereinafter DSS) from their home and placed in the custody of the child’s adult half brother and his wife, Frank Albanese and Barbara Albanese. Respondent reportedly visited the child daily until his incarceration in April 2004, upon his arrest and later conviction for criminal possession of stolen property, for which a 2 to 6-year sentence was imposed (see People v Albanese, 38 AD3d 1015 [2007], lv denied 8 NY3d 981 [2007]). Petitioner reportedly completed drug treatment and parenting services recommended by DSS and, in August 2005, was granted joint custody with the Albaneses, who were given and have exercised visitation with the child every two weeks, with primary physical custody to petitioner. That order is not in the record on appeal. In January 2006, petitioner sought to modify that custody/ visitation order seeking sole, full custody with continued visitation with the Albaneses, to which they stipulated. Respondent opposed the petition, requesting visitation. After a hearing and by decision and order dated April 12, 2006, Family Court relieved the Albaneses of their joint custodial status and directed that custody remain with petitioner pending the hearing on respondent’s opposition to the petition and his request for visitation. At the June 2006 hearing, only the parties testified. Petitioner sought sole custody; respondent, who remained incarcerated, requested visitation, ultimately waiving his earlier request for joint custody. Petitioner testified that her son had lived with her the past year, during which she took him every two weeks to visit the Albaneses. She has not taken the child to visit respondent, who the child has not seen since his incarceration in April 2004, reportedly due to financial constraints and transportation difficulties; she has, however, sent respondent monthly envelopes containing the son’s cards, pictures, report cards, and the like. She confirmed that, prior to his incarceration, the child had enjoyed a “good” relationship with his father, hindered only by respondent’s drug use. She is agreeable to the child visiting respondent, provided it is accomplished by means of free bus transportation that respondent reported was available, to be supervised by his paternal grandmother (or another agreed-upon paternal relative). Petitioner also indicated that she would be willing to undertake the trip with the child depending upon the cost, if any, to her. Respondent testified that he previously enjoyed a close relationship with the child, with whom he does not want to lose contact, and requested monthly visitation. As his financial resources consist only of $5.80 earned every two weeks in prison, he proposed use of the free bus although he did not have any detailed information. At the close of the hearing, petitioner’s counsel requested sole custody and did not oppose visitation; respondent’s counsel requested “some sort of visitation,” and the Law Guardian advocated in favor of sole custody for petitioner and monthly visitation with respondent, with transportation responsibility to alternate monthly between the parties’ families. Ruling from the bench, Family Court stated on the record that the limited issue before it was custody, which was awarded to petitioner, and held that respondent is entitled to “reasonable visitation as the parties may agree to consist solely of mail visitation,” i.e., no in-person visitation. In its subsequent written order, the court directed that respondent “shall have reasonable visitation ... as the parties can agree, and in the minimum he shall have mail contact.” Respondent now appeals, arguing that he was deprived of a fair trial and should have been awarded in-person visitation, contentions also advocated on appeal by the Law Guardian. As an initial matter, we have reviewed the entire record and do not find that respondent was denied a fair trial by either Family Court’s conduct or its remarks. With that said, we do not condone the frequent and unprovoked intemperate and denigrating remarks directed at respondent by the court, which were clearly inappropriate and served only to undermine “public confidence in the integrity, fair-mindedness and impartiality of the judiciary” (Matter of Esworthy, 77 NY2d 280, 282 [1991]). Inasmuch as respondent had an unquestioned, fundamental statutory right to be represented by counsel in these proceedings (see Family Ct Act § 262 [a]; Matter of Wilson v Bennett, 282 AD2d 933, 934-935 [2001]), he should not have been chastised by the court for exercising that right and “wasting [the court’s] time” at the initial hearing. Turning to the denial of visitation, while respondent did not file a formal petition for visitation as he was advised he could do, petitioner’s modification petition clearly reflected that it related to both custody and visitation of the child; respondent at all times appeared in opposition to that petition, requested visitation and, at the outset of every appearance, it was stated that both custody and visitation were in issue. Thus, as all parties were on notice of the foregoing and the issue of in-person visitation was in fact litigated by all, we cannot agree with Family Court’s bench-ruling statement that because only petitioner’s “custody petition” was pending, the only issue before the court was custody. Turning to respondent’s visitation request, under established principles, “ [visitation by a noncustodial parent is presumed to be in the child’s best interest and should be denied only in exceptional situations, such as where substantial evidence reveals that visitation would be detrimental to the welfare of the child” (Matter of Frierson v Goldston, 9 AD3d 612, 614 [2004]; see Matter of Conklin v Hernandez, 41 AD3d 908, 910 [2007]; Matter of Tanner v Tanner, 35 AD3d 1102, 1103 [2006]). No such proof was adduced here, where petitioner’s testimony and the child’s Law Guardian both supported in-person visitation. That presumption exists despite respondent’s incarceration, which should not, by itself, preclude visitation in the absence of any evidence that it would not be in the child’s best interests (see Matter of Edward S. v Moon, 7 AD3d 834, 836 [2004]; Matter of Ellett v Ellett, 265 AD2d 747, 747 [1999]; see also Matter of Conklin v Hernandez, 41 AD3d at 910). Despite the parties’ limited resources and the minimal details offered by respondent regarding his plan to use a reportedly available free bus service, the record supports the conclusion that in-person visitation is viable and workable among these families, and would benefit the child (cf. Matter of Conklin v Hernandez, 41 AD3d at 910-911). As such, the court’s denial of such unopposed in-person visitation is “ ‘lacking a sound basis in the record’ ” (Matter of Edward S. v Moon, 7 AD3d at 836, quoting Matter of Williams v Tillman, 289 AD2d 885, 885 [2001]; cf. Matter of Trombley v Trombley, 301 AD2d 890, 891 [2003]). Accordingly, this matter is remitted to Family Court for further proceedings, as necessary, before a different judge to establish a permanent visitation order within the court’s discretion based, among other factors, on updated information. In the interim, this Court will issue a temporary order granting respondent (1) in-person visitation with his son every other month commencing with November 2007, with respondent bearing the responsibility to make the necessary arrangements and pay any associated expenses, with the child to be accompanied either by petitioner, respondent’s mother, or another adult agreed to by petitioner, and (2) monthly mail communication between the child and respondent through petitioner. Mercure, J.P., Peters, Carpinello and Kane, JJ., concur. Ordered that the order is modified, on the law and facts, without costs, by reversing so much thereof as awarded respondent unspecified minimal mail access and denied him in-person access to the child; matter remitted to the Family Court of Broome County for further proceedings not inconsistent with this Court’s decision, and, pending said proceedings, respondent is temporarily awarded bimonthly in-person access commencing in November 2007 to be arranged by and at the expense of respondent, and monthly mail communication with the child via petitioner; and, as so modified, affirmed. Ordinarily, the failure to include the order sought to be modified in Family Court would result in a finding that an incomplete record has been submitted (see Matter of Pratt v Anthony, 30 AD3d 708 [2006]). Here, there is no dispute that the August 2005 order resulted in petitioner obtaining joint custody with the Albaneses (who also received visitation) and physical custody, and no indication that any award of visitation to respondent was made.
Ralph Hart Tweddell Ralph Hart Tweddell (25 May 1843 – 3 September 1895) was a British mechanical engineer, known particularly for inventing the portable hydraulic riveter, which greatly facilitated the construction of boilers, bridges and ships. Early life Tweddell was born in South Shields in 1843. His father, Marshall Tweddell, was a shipowner. He was educated at Cheltenham College, with the intention of entering the army. However, more interested in engineering, he was apprenticed to R and W Hawthorn, a locomotive manufacturer in Newcastle upon Tyne. During his apprenticeship he took out a patent in 1865 for a portable hydraulic apparatus for fixing the ends of boiler tubes in tube plates. The encouraging results suggested that hydraulic power should be used for machines used in boiler construction. In 1865 he designed a stationary hydraulic riveting machine. This was more effective than the existing mechanical riveting machines, which usually did not make steam-tight joints in the boilers of steamships, which were by then operating under higher pressure than before. The plant, consisting of a pump, an accumulator, and a riveter, was first used by Thompson, Boyd & Co., of Newcastle. The portable riveter In 1871 Tweddell invented a portable riveter, so that the work did not have to be brought to the machine. It was manufactured by Fielding & Platt; an early user was Armstrong, Mitchell and Company in Newcastle. In 1873 it was used in riveting a lattice girder bridge carrying Primrose Street over the Great Eastern railway at Bishopsgate railway station in London. The success of this led to its use in the construction of other bridges. The portable riveter was first used for locomotive work by F. W. Webb at Crewe Works. It was also used for agricultural machinery, for underframes of railway carriages, and by the Italian government for gun-carriages. In France, Tweddell's system was used when in 1874 the French government began to build iron warships in Toulon, and it was later used at the shipyard at Penhoët near Saint-Nazaire, and at Brest. In 1885 he was awarded a gold medal, under the Howard Trust, at the International Inventions Exhibition in South Kensington for "his system of applying hydraulic power to the working of machine tools, and for the rivetting and other machines which he has invented in connection with that system". He wrote the paper "On Machine Tools and Labour-saving Appliances worked by Hydraulic Pressure" for the Institution of Civil Engineers, for which he was awarded the Telford Medal and premium in 1883. He sent three papers to the Institution of Mechanical Engineers including "On the Application of Water Pressure to Shop-tools and Mechanical Engineering Works". In 1890 a paper entitled "The Application of Water Pressure to Machine Tools and Appliances" was awarded a Bessemer Premium by the Society of Engineers. He became a member of the Institution of Mechanical Engineers in 1867. From 1879 he was a member of the Institution of Civil Engineers. He was also a member of the from 1879. In 1875 Tweddell married Hannah Mary Grey. In his spare time he was interested in hunting, shooting and fishing. He died in 1895 at his home near Gravesend in Kent, two years after a riding accident which affected his health.
Gustav Alexander Gustav Alexander (1873 – 12 April 1932) was an Austrian otolaryngologist remembered for describing Alexander's law. He was the director of the Department of Otology of the Wiener Allgemeine Poliklinik from 1917 until his death. He was the father of Leo Alexander. He was assassinated on the street between his home and the Poliklinik by Johann Sokoup, a Czechoslovakian former patient who had tried to assassinate him 22 years earlier.
Digiview Productions TV 2 1. The best of Ernie & Bert 2. The best of elmo 3. Lost at sea (the sponge who could fly) 4. Rock 'N'Roll Clifford/team Clifford (embarrassing moments) 5. Kim Possible the villain fillies (blush) 6. Pooh wishes (balloonatics) 7. The wacky adventures of Ronald McDonald have time will travel 8. Barney's adventure bus 9. Barney's super singing circus 10. Elmo's world food, water and exercise (food) 11. Nautical nonsense (something Smells) 12. SpongeBob goes prehistoric (nature pants) 13. Come on over to Barney's house 14. A boy named Charlie Brown 15. Sing and dance with barney 17. Barney's alphabet zoo 18. Elmo's world head to toe with elmo (jacket) 19. Everybody Loves Clifford/Good Friends, Good Times (Jetta's Tall Tale) 20. Look Out Clifford/Big Fun in the Sun (lights out) 21. The Legend of Grimace Island 22. Barney in Outer Space
namespace Threax.AspNetCore.Halcyon.Ext { /// <summary> /// This interface will generate a title from a given name. /// </summary> public interface IAutoTitleGenerator { string CreateTitle(string name); } }
Evil Hand From Dragon Quest Wiki Evil Hand is a skill that appears exclusively in Dragon Quest Tact. It is one of the Ace of Spades' signature abilities, slamming a giant hand covered in Dark Matter at its enemies. Details[edit] Evil Hand can be learned by Ace of spades at level 54 and costs 84 MP to use. It inflicts 250% potency Zam-type physical damage to all enemies in a plus sign-shaped area of effect in front of it. Upgrading the skill increases its damage and reduces the amount of MP it costs to use. Evil Hand (イービルハンド Ībiruhando) Ability information Evil Hand Role * Type * Element MP cost Attack Physical 60 Range Additional effects Front None Description Deals 250% potency Zam-type physical damage to all enemies in area of effect Naturally learnt by Ace of Spades
package sg.edu.iss.demo; import java.util.List; import org.springframework.beans.factory.annotation.Autowired; import org.springframework.boot.CommandLineRunner; import org.springframework.boot.SpringApplication; import org.springframework.boot.autoconfigure.SpringBootApplication; import org.springframework.context.annotation.Bean; import sg.edu.iss.demo.domain.Student; import sg.edu.iss.demo.repo.StudentRepository; @SpringBootApplication public class DemoApplication { @Autowired StudentRepository srepo; public static void main(String[] args) { SpringApplication.run(DemoApplication.class, args); } @Bean CommandLineRunner runner() { return args -> { Student s1 = new Student("Hou Lu", "Sweet", null, 5000.00, 4.50); Student s2 = new Student("Zerita", "Nice", null, 5000.00, 4.50); Student s3 = new Student("Austin", "Rationlist", null, 5000.00, 4.50); Student s4 = new Student("Benjo", "ExamSmart", null, 5000.00, 5.00); srepo.save(s1);srepo.save(s2);srepo.save(s3);srepo.save(s4); List<Student> list = srepo.findAll(); for (Student student : list) { System.out.println("Record from DB"+student.toString()); } }; } }
type Link = { path: string; label: string; }; export type Downloadable = { name: string; links: Link[]; }; export const getDownloadableItemsList = (fileName: string): Downloadable[] => { const versionPrefix = `latest-${fileName.split('.')[0]}.x`; return [ { name: 'Windows Installer (.msi)', links: [ { path: `https://nodejs.org/dist/${versionPrefix}/node-${fileName}-x86.msi`, label: '32-bit', }, { path: `https://nodejs.org/dist/${versionPrefix}/node-${fileName}-x64.msi`, label: '64-bit', }, ], }, { name: 'Windows Binary (.zip)', links: [ { path: `https://nodejs.org/dist/${versionPrefix}/node-${fileName}-win-x86.zip`, label: '32-bit', }, { path: `https://nodejs.org/dist/${versionPrefix}/node-${fileName}-win-x64.zip`, label: '64-bit', }, ], }, { name: 'macOS Installer (.pkg)', links: [ { path: `https://nodejs.org/dist/${versionPrefix}/node-${fileName}.pkg`, label: '64-bit', }, ], }, { name: 'macOS Binary (.tar.gz)', links: [ { path: `https://nodejs.org/dist/${versionPrefix}/node-${fileName}-darwin-x64.tar.gz`, label: '64-bit', }, ], }, { name: 'Linux (PPC, S390x, AIX, ARMv7, ARMv8)', links: [ { path: `https://nodejs.org/dist/${versionPrefix}/node-${fileName}-linux-x64.tar.xz`, label: '64-bit', }, { path: `https://nodejs.org/dist/${versionPrefix}/node-${fileName}-linux-ppc64le.tar.xz`, label: 'ppc-le-64-bit', }, { path: `https://nodejs.org/dist/${versionPrefix}/node-${fileName}-linux-s390x.tar.xz`, label: 's390x-64-bit', }, { path: `https://nodejs.org/dist/${versionPrefix}/node-${fileName}-aix-ppc64.tar.gz`, label: 'aix-64-bit', }, { path: `https://nodejs.org/dist/${versionPrefix}/node-${fileName}-linux-armv7l.tar.xz`, label: 'armv7-32-bit', }, { path: `https://nodejs.org/dist/${versionPrefix}/node-${fileName}-linux-arm64.tar.xz`, label: 'armv8-64-bit', }, ], }, ]; };
How to configure the Linux kernel/drivers/acpi Howto configure the Linux kernel / drivers / acpi * ACPI Configuration * ACPI Configuration A ceforge.net/projects/acpi > * Linux support for ACPI is based on Intel Corporation's ACPI Component Architecture (ACPI CA). * Linux support for ACPI is based on Intel Corporation's ACPI Component Architecture (ACPI CA). * Option: ACPI_SLEEP_PROC_FS * Kernel Versions: <IP_ADDRESS> ... * (on/off) * depends on ACPI_SLEEP && PROC_FS * default y * Option: ACPI_SLEEP_PROC_SLEEP * Kernel Versions: <IP_ADDRESS> ... * (on/off) /proc/acpi/sleep (deprecated) * depends on ACPI_SLEEP_PROC_FS * default n * Create /proc/acpi/sleep Deprecated by /sys/power/state * Option: ACPI_AC * Kernel Versions: <IP_ADDRESS> ... * (on/off/module) AC Adapter * depends on X86 * default y * Option: ACPI_BATTERY * Kernel Versions: <IP_ADDRESS> ... * (on/off/module) Battery * depends on X86 * default y * Option: ACPI_BUTTON * Kernel Versions: <IP_ADDRESS> ... * (on/off/module) Button * default y * Option: ACPI_VIDEO * Kernel Versions: <IP_ADDRESS> ... * (on/off/module) Video * depends on X86 * default y * Option: ACPI_HOTKEY * Kernel Versions: <IP_ADDRESS> ... * (on/off/module) Generic Hotkey (EXPERIMENTAL) * depends on EXPERIMENTAL * depends on X86 * default n * Experimental consolidated hotkey driver. If you are unsure, say N. * Option: ACPI_FAN * Kernel Versions: <IP_ADDRESS> ... * (on/off/module) Fan * default y * Option: ACPI_PROCESSOR * Kernel Versions: <IP_ADDRESS> ... * (on/off/module) Processor * default y * Option: ACPI_HOTPLUG_CPU * Kernel Versions: <IP_ADDRESS> ... * (on/off) * depends on ACPI_PROCESSOR && HOTPLUG_CPU * select ACPI_CONTAINER * default y * Option: ACPI_THERMAL * Kernel Versions: <IP_ADDRESS> ... * (on/off/module) Thermal Zone * depends on ACPI_PROCESSOR * default y * Option: ACPI_NUMA * Kernel Versions: <IP_ADDRESS> ... * (on/off) NUMA support * depends on NUMA * depends on (IA64 || X86_64) * default y if IA64_GENERIC || IA64_SGI_SN2 * Option: ACPI_ASUS * Kernel Versions: <IP_ADDRESS> ... tristate ASUS/Medion Laptop Extras * depends on X86 * Option: ACPI_IBM * Kernel Versions: <IP_ADDRESS> ... * (on/off/module) IBM ThinkPad Laptop Extras * depends on X86 * If you have an IBM ThinkPad laptop, say Y or M here. * Option: ACPI_TOSHIBA * Kernel Versions: <IP_ADDRESS> ... * (on/off/module) Toshiba Laptop Extras * depends on X86 * default y * More information about this driver is available at . * If you have a legacy free Toshiba laptop (such as the Libretto L1 series), say Y. * Option: ACPI_CUSTOM_DSDT * Kernel Versions: <IP_ADDRESS> ... * (on/off) Include Custom DSDT * depends on !STANDALONE * default n * Thist option is to load a custom ACPI DSDT If you don't know what that is, say N. "Custom DSDT Table file to include" * Option: ACPI_CUSTOM_DSDT_FILE * Kernel Versions: <IP_ADDRESS> ... * depends on ACPI_CUSTOM_DSDT * default "" * Enter the full path name to the file wich includes the AmlCode declaration. "Disable ACPI for systems before Jan 1st this year" if X86 * Option: ACPI_BLACKLIST_YEAR * Kernel Versions: <IP_ADDRESS> ... * default 0 * Option: ACPI_DEBUG * Kernel Versions: <IP_ADDRESS> ... * (on/off) Debug Statements * default n * Option: ACPI_EC * Kernel Versions: <IP_ADDRESS> ... * (on/off) * depends on X86 * default y * Option: ACPI_POWER * Kernel Versions: <IP_ADDRESS> ... * (on/off) * default y * Option: ACPI_SYSTEM * Kernel Versions: <IP_ADDRESS> ... * (on/off) * default y * Option: X86_PM_TIMER * Kernel Versions: <IP_ADDRESS> ... * (on/off) Power Management Timer Support * depends on X86 * depends on !X86_64 * default y * Option: ACPI_CONTAINER * Kernel Versions: <IP_ADDRESS> ... * (on/off/module) ACPI0004,PNP0A05 and PNP0A06 Container Driver (EXPERIMENTAL) * depends on EXPERIMENTAL * If one selects "m", this driver can be loaded with modprobe acpi_container. $>modprobe acpi_memhotplug * Option: ACPI_HOTPLUG_MEMORY * Kernel Versions: <IP_ADDRESS> ... * (on/off/module) Memory Hotplug * depends on ACPI * depends on MEMORY_HOTPLUG * default n * If one selects "m," this driver can be loaded using the following command: * ACPI Linux Kernel Configuration
# Copyright (c) 2019 The Regents of the University of Michigan # All rights reserved. # This software is licensed under the BSD 3-Clause License. """Prints signac statepoints.""" import os import signac def modify(frame, data): source_file = data.attributes["SourceFile"] dirname = os.path.dirname(source_file) job = signac.get_job(dirname) for k, v in job.sp.items(): print(k, v)
What is page name? I often have a case, when I don't know real name of page. For example, if I go to stackoverflow.com, I don't know, do I go to stackoverflow.com/index.php or stackoverflow.com/index.html or some another page. And it can be none of index.*. Is a method to learn real full path and name of page? The real full path is https://stackoverflow.com/. I just toke example. And no, https://stackoverflow.com/ is domain name and usually (especially, if we speak about Apache) domain name points to root directory of site and server automatically redirects client to file with page (index.php, for example). About that I speak — can I know the name of the finish file? The concept of a direct correspondence between urls and disk files belongs to 1995. Yes, about that I speak. Usually, URL gives file. And server gives a file, that's path is pointed in URL (relatively of root of site). If the URL points not to file, but to folder, server looks in that directory for file with especially name. Usually it is index.php, index.html, etc. I want to know, what exactly is this name in a current case. Like I said, you cannot tell by looking at a URL whether or not there is a file one the disk, nor should you try. A URL that looks like https://stackoverflow.com/files/archive.zip may be processed by a controller at https://stackoverflow.com/, without having any files on disk. A URL that looks like https://stackoverflow.com/index.php can be processed by a controller at https://stackoverflow.com and make you think SO is written in PHP, which it isn't.
Diversity+of+Life-Fungi = Diversity of Life = = Kingdon: Fungi = * Amanita phalloides ** * Contents * 1) Taxonomy * 2) Introduction * 3) Morphological Features * 4) Anatomical/Physiological Features * 5) Range of Habitat * 6) Interesting Facts * 7) References || * Taxonomy ** * Kingdom || Fungi || * Phylum || Basidiomycota || * Class || Agaricomyetes || * Order || Agaricales || * Family || Pluteaceae || * Genus || Amanita || * Species || phalloides || * Introduction ** * Morphological Features ** * Anatomical/Physiological Features ** * Range of Habitat ** * Interesting Facts ** * Amanitin phalloides is responsible for over 90% of fatal and serious mushroom poisoning cases. * References ** * 2) "Basidiomycota." Tree of Life Web Project. Web. 15 Nov. 2010. . * 3) "Amanita Phalloides." Wikipedia, the Free Encyclopedia. Web. 15 Nov. 2010. . * 4) "Fungi." Tree of Life Web Project. Web. 15 Nov. 2010. .
package com.baeldung; import java.util.Optional; import org.apache.commons.lang3.ObjectUtils; public class IntegerNullOrZero { private IntegerNullOrZero() { throw new RuntimeException("This class cannot be instantiated."); } public static boolean usingStandardWay(Integer num) { return num == null || num == 0; } public static boolean usingTernaryOperator(Integer num) { return 0 == (num == null ? 0 : num); } public static boolean usingOptional(Integer num) { return Optional.ofNullable(num).orElse(0) == 0; } public static boolean usingObjectUtils(Integer num) { return ObjectUtils.defaultIfNull(num, 0) == 0; } }
Ask Your Question 0 Why do I get the "unable to find a common ring for all elements" error message? asked 2018-04-29 09:17:28 -0500 import hashlib import binascii from sage.crypto.util import ascii_integer bin = BinaryStrings() k1 = hashlib.sha224(b"Nobody inspects the spammish repetition").hexdigest() k2 = hashlib.sha224(b"Nobody inspects").hexdigest() A = bin.encoding(k1) B = bin.encoding(k2) A = list(A) B = list(B) A = Matrix([[A],[B]]) edit retag flag offensive close merge delete 1 answer Sort by » oldest newest most voted 2 answered 2018-04-29 13:24:50 -0500 updated 2018-04-29 16:23:52 -0500 The elements of your lists belong to a monoid, not a ring: sage: type(A[0]) <class 'sage.monoids.string_monoid_element.StringMonoidElement'> sage: A[0].parent() Free binary string monoid So, you can transform them into integer first: sage: A = [ZZ(str(i)) for i in A] sage: B = [ZZ(str(i)) for i in B] Then, you can do: sage: Matrix([A,B]) 2 x 448 dense matrix over Integer Ring (use the '.str()' method to see the entries) (note that you put to many brackets). You can also trnasform 0 and 1 as elements of GF(2) if it makes more sense in your application. edit flag offensive delete link more Your Answer Please start posting anonymously - your entry will be published after you log in or create a new account. Add Answer Question Tools 1 follower Stats Asked: 2018-04-29 09:17:28 -0500 Seen: 37 times Last updated: Apr 29 '18
use wasmi::{MemoryRef}; use crate::error::CryptoError; use crate::key::{Key as KeyTrait, KeyBase}; pub trait Signature { type Base:SignatureBase; fn size() -> usize { Self::Base::SIGNATURE_SIZE } fn sign(memory:&MemoryRef,seed_ptr:u32,seed_size:usize,seckey_ptr:u32,new_ptr:u32) -> Result<(u32,usize),CryptoError> { let seed_data = memory.get(seed_ptr, seed_size).unwrap(); let seckey_slice = memory.get(seckey_ptr, <<<Self::Base as SignatureBase>::Key as KeyTrait >::Base as KeyBase>::SECRET_SIZE).unwrap(); let seckey:<<<Self::Base as SignatureBase>::Key as KeyTrait >::Base as KeyBase>::Secret = <<<Self::Base as SignatureBase>::Key as KeyTrait >::Base as KeyBase>::serialize_secret(&seckey_slice[..])?; let signature = Self::Base::sign(&seed_data[..],&seckey)?; memory.set(new_ptr,signature.as_ref()).unwrap(); Ok((new_ptr,Self::size())) } fn verify(memory:&MemoryRef,seed_ptr:u32,seed_size:usize,pubkey_ptr:u32,signature_ptr:u32) -> Result<bool,CryptoError> { let seed_data = memory.get(seed_ptr, seed_size).unwrap(); let pubkey_slice = memory.get(pubkey_ptr, <<<Self::Base as SignatureBase>::Key as KeyTrait >::Base as KeyBase>::PUBLIC_SIZE).unwrap(); let pubkey:<<<Self::Base as SignatureBase>::Key as KeyTrait >::Base as KeyBase>::Public = <<<Self::Base as SignatureBase>::Key as KeyTrait >::Base as KeyBase>::serialize_public(&pubkey_slice[..])?; let signature_slice = memory.get(signature_ptr, <Self::Base as SignatureBase>::SIGNATURE_SIZE).unwrap(); let signature:<Self::Base as SignatureBase>::Signature = <Self::Base as SignatureBase>::serialize_signature(&signature_slice[..])?; <Self::Base as SignatureBase>::verify(&seed_data[..], &pubkey, &signature) } } pub trait SignatureBase:AsRef<[u8]> { const SIGNATURE_SIZE:usize; type Key:KeyTrait; type Signature:AsRef<[u8]>; //[u8;SIGNATURE_SIZE] fn sign(seed:&[u8],seckey:&<<Self::Key as KeyTrait>::Base as KeyBase>::Secret) -> Result<Self::Signature,CryptoError>; fn verify(seed:&[u8],pubkey:&<<Self::Key as KeyTrait>::Base as KeyBase>::Public, signature:&Self::Signature) -> Result<bool,CryptoError>; fn serialize_signature(data:&[u8]) -> Result<Self::Signature,CryptoError>; }
Weakness * In-game description: ''Condition. While suffering from this Condition, you deal less damage with attacks.'' * style="padding-left: 10px;" | * } Weakness is a condition that reduces the damage of attacks by 66%. Weakness does not apply to damage bonuses gained from attack skills. Weakness now reduces all your Attributes by 1 (Update 1/19/07) Weakness is represented by a white cloud around the feet of the character. Heavy weapon upgrades extend the duration of weakness. Condition Removal Skills See Condition removal quick reference. Condition Transferring Skills See Transfer condition skills quick reference. Other Weakness-related skills * Auspicious Blow (gain energy if foe suffers from Weakness) * Fierce Blow (deal more damage if foe suffers from Weakness) * Heavy Blow (knock down and deal extra damage to foe if foe suffers from Weakness) * Oppressive Gaze (steal Health from target and adjacent foes suffering from Weakness) * Glowstone (gain energy if foe suffers from Weakness) * Stoning (knock down foe if foe suffers from Weakness)
import { MigrationInterface, QueryRunner, TableColumn, TableForeignKey } from 'typeorm'; export class AddColumnFarmIdToProcessTable1622130152632 implements MigrationInterface { private toFarm = new TableForeignKey({ name: 'fk_process_farm', columnNames: ['farm_id'], referencedColumnNames: ['id'], referencedTableName: 'farm', }); public async up(queryRunner: QueryRunner): Promise<void> { await queryRunner.addColumn('process', new TableColumn({ name: 'farm_id', type: 'uuid', isNullable: false, })); await queryRunner.createForeignKey('process', this.toFarm); } public async down(queryRunner: QueryRunner): Promise<void> { await queryRunner.dropColumn('process', 'farm_id'); await queryRunner.dropForeignKey('process', this.toFarm); } }
export interface ClientOptions { isPlaying: boolean; /** The id used to identify the client. */ id: string; /** * The id of the room that the user wants to join if we want to join a specific one. If ommitted, * the user will join the first available room. */ roomToJoin?: string; }
Articulated joint for sinuous chain conveyers H. E. SMITH Sept. 14, 1954 ARTICULATED JOINT FOR SINUOUS CHAIN CONVEYERS 3 Sheets-Sheet 1 Filed Dec. 17, 1952 INVENTOR. fi EE/VAA/E 67w r/-/ Sept. 14, 1954 H. E. SMITH 2,689,034 ARTICULATED JOINT FOR SINUOUS cum: CONVEYERS Filed Dec. 17. 1952 5 Sheets-Sheet 2 E4 IN VEN TOR. Patented Sept. 14, 1954 ARTICULATED JOINT FOR SINUOUS CHAIN CONVEYERS Herman E. Smith, Riverside, Ill., assignor to Goodman Manufacturing Company, Chicago, 111., a corporation of Illinois Application December 17, 1952, Serial No. 326,458 4 Claims. This invention relate to improvements in conveyors and more particularly relates to an improved closure means and flight hold down across the gaps between adjacent ends of the troughs of an articulated conveyor. An object of my invention is to provide a novel and improved form of closure means for closing the gaps between adjacent ends of the troughs of an articulated conveyor, having a retaining guide associated therewith, holding the flights down to the bottom of the conveyor trough in all positions of articulation of the trough sections of the conveyor with respect to each other. Another object of my invention is to provide an articulated trough construction for a sinuous conveyor of the center strand chain and flight type having a simple form of hold down means, holding down the flights as they pass by the articulated joints of the conveyor and maintaining the closure plates at the gaps between the side walls of the conveyor into slidable engagement therewith as the trough sections of the conveyor are pivoted with respect to each other A further object of my invention is to provide a closure means for a sinuous flight conveyor, closing the gaps between adjacent ends of the trough sections of the conveyor and holding down the conveyor flights during movement along the articulated joint thereof, which is of a simpler form and more economical construction than the conventional gap closure and articulated joint structures heretofore used in flexible conveyors of this type. A still further object of my invention is to provide a means for closing the gaps between adjacent side walls of the trough sections of an articulated conveyor at the articulated joint thereof, wherein the closure means also forms a hold down for the ends of the flights at the joints between the trough sections of the conveyor, and is retained in closure engagement with the side walls of the trough sections of the conveyor by slidable engagement of the hold down therewith. These and other objects of my invention will appear from time to time as the following speciflcation proceeds and with reference to the accompanying drawing wherein: Figure l is a top plan view of an articulated trough section of a sinuous conveyor, illustrating one form in which my invention may be embodied; Figure 2 is a view in side elevation of the trough section shown in Figure 1; Figure 3 is an enlarged fragmentary plan view Figure 6 is a fragmentary transverse sectional view taken through the conveyor at the pivot thereof. In the embodiment of my invention illustrated in the drawings, the sinuous conveyor may be made up of a series of conveyor trough units, each including two similar trough sections In and H, connected together at their ends in end to end relation with respect to each other. The trough sections l0 and II are shown as being pivotally connected at their adjacent ends for movement about the axis of an upright pin l2. A laterally flexible endless chain [3 having parallel spaced transverse flights l4, l4 pivotally connected thereto at suitable intervals, passes along the center of the conveyor troughs l0 and H for conveying loose material therealong, the flights being guided at their ends on the insides of the side walls of the troughs of the conveyor. The conveyor chain l3 and flights l4, 14 are shown as being of a conventional laterally flexible construction, including a plurality of links l5, [5 connected together for movement about horizontal axes and having pivotal connection with opposite sides of the flights I4, 14 for pivotal movement about vertical axes, giving the conveyor sufflcient lateral flexibility to conform to the trough sections in and II, when in various laterally extended positions with respect to each other about the axis of the pin 12. The upper run of the conveyor is shown as being movable along the top surface of material carrying pans I6 and ll of the respective trough sections In and II. The return run of the conveyor may, a usual, be guided for movement beneath the pans l6 and I1 along under pans l9, 19 supporting the return run of the conveyor. The trough sections [0 and II are shown as having upstanding generally parallel side walls 20, 20 and 2|, 2| respectively, extending above and below the pans Hi and I1, and suitably secured to opposite sides thereof, and to opposite side of the bottom pans l9, l9 as by welding. At the ends of the side walls 20, 20 and secured thereto, inwardly of the ends thereof, are bifurcated coupling members 23, 23, adapted to receive coupling tongues 24, 24, spaced laterally from and extending forwardly of the adjacent side walls 2!, 2| of the next adjacent trough section. The coupling tongues 28, 24 are adapted to extend within the space between the fur cations of the coupling members 23, 23, and to be pivotally connected thereto as by horizontally extending coupling pins (not shown), connecting said trough sections to conform to rolling ground. Supporting the trough sections I and II on the ground for movement therealong is a pair of wheels 25, 25 rotatably mounted on opposite ends of an axle 25, which may extend beneath the trough section III and be suitably secured thereto. The wheels 25, 25 are mounted intermediate the ends of the trough section It and besides forming a support for the conveyor on the ground also enable a :person to grasp the end of the trough section II as 'by coupling tongues 22, 22 and lift the same and wheel the individual trough sections along the ground as a wheelbarrow, when assembling or disconnecting the conveyor to extend or move the same to a new location. The free end of the bottom pan I7. of the trough section Il may be curved downwardly from the horizontal as indicated by reference character 2! and slidably engaged by an overlapping beveled or arcuate free end of the bottom plate I6, so as to form a continuous trough with no gap therebetween, where the ground may be uneven, and to prevent the spilling of coal at the joints between the individual trough sections and to provide a smooth uninterrupted conveying surface across the joints thereof. The side walls 2|, 2| are likewise shown as having spaced retaining plates 29, 29 secured to opposite sides thereof, and extending forwardly therefrom and having slidable engagement with opposite sides of the side walls 20, 20 of the next adjacent trough section to maintain the gaps between adjacent ends of the trough sections-closed, when one trough section is vertically displaced with respect to .the other. Brackets 30, 30 are shown as extending outwardly from the side walls 20, 20 rearwardly of the wheels .25, 25 and as having rollers 34, 3| mounted thereon for rotation about vertical pivot pins 32, 32. Said rollers may engage an arcuate guide (not shown), which may be positioned about a corner where the mine room communicates with the mine entry, to accommodate the sinuous conveyor to be connected to and extended from a loading machine, or continuous mining machine, out of the working place and along a mine entry extending at right angles to the work ing place and to be advanced with the loading machine and load into a main line conveyor extending along the entry, as in an application Serial No. 203,676, filed by Frank Cartlidge on December 30, 1950 and entitled Portable Conveying Apparatus. The trough section III is shown as having flanges 33, .33 extending laterally outwardly from the upper ends of the .side walls .29, 20 and forming a stiffening means therefor. The trough section II is shown as having similar flanges 34, 35 extending laterally outwardly from the tops of the side walls 2|, 2| thereof. The end of the plate I6, adjacent the plate I1, is shown .as being concave ly formed by an arc struck from the center .of the pin I2 and as having slidable engagement with a coaxial con frame member 39 extending along the "bottom of the plate 3'! and secured to the trough section It between the side walls thereof. In a like I manner the bottom pan I9 has a plate 4|) secured to the bottom thereof, in which is mounted a pivot pin 4|, depending therefrom and pivotally articulation with respect to each other. mounted in a cross-frame member 42, secured to the trough section I0 between the side walls thereof. The means for closing the gaps between the adjacent ends of the side walls 20 and 2| of the trough sections III and I are herein shown as comprising two upright closure plates 43, 43 interleaving the inner ends of the side walls 20, 20 and pivotally-connected thereto on vertical pivot pins 44, 44. The closure plates 183, 43 are shown as being of a generally arcuate :form, extending across the gaps between the adjacent ends of the side walls 20 and 2| and *slidably engaging the outer sides of the side walls 2|, 2|. The closure plates 43, 43 extend within retainer bars '45, 45 extending vertically along the side walls 2|, :21. The retainer bars 45, 45 are secured at their upper ends to the bottoms of the flanges til, :34 and at their lower ends to ears 4-6, 55 extending outwardly from the sidewalls 2|, 21 adjacent the bottom plate I! thereof. The gaps beneath the bottom plate I l are shown as being closed by closure plates 41, 41, interleaving the portions of the side walls 20, 20 "beneath the plate I1 and slidably engaging the side walls 2|, 2|. Each closure plate 41 is retained in engagement with an associated side wall 2| by a retainer bar 49. Each retainer bar '49 is shown as being mounted between the bottom :of :an ear 45 extending outwardly from the side wall 2|. in substantial alignment with the bottom plate I! and a downwardly spaced ear 59 extending laterally from the sidewall 2|, adjacent the bottom thereof. Spaced upwardly from and parallel to the bot-- tom plate I6 are inwardly extending hold-down flanges 5|, 5|. The hold down flanges 5|, 5] may be secured to the insides of the .side walls 20, 20, as by welding, and extend therealong parallel to the bottom pan It for substantially the length thereof. The under surfaces of the hold down flanges 5|, 5| may be slidably engaged by the upper surfaces of the flights Id, I4 to retain said flights to the bottom plate It during travel therealong, particularly when the trough sections H3 and I I are in various positions .of Each hold down 5| is shown as extending beyond the inner end .of the associated side wall .22 along, but spaced inwardly from the associated closure plate '43 and as having an advance outer edge 52 curved or inclined outwardly from-said closure plate and engaging said closure plate when the trough sections In and Il are in extreme positions of articulation. Hold down flanges 53, 53 similar to the hold down flanges 5|, 5|, but spaced 'there beneath, ' extend along the side walls 2|, 2| and slidably yond the end .of the side wall .2-.I and as having an outer edge .56 curving .inwardly toward the closure member 43 and engaging the same in the extreme positions of articulation of the trough sections. In Figures 5 and 6, the hold down flange 5I is shown as overlapping the hold down plate 55 in the region of articulation of the trough sections In and I I and as slidably moving therealong during articulation of the trough sections I and II. The hold down plate 55 is likewise shown as overlapping the hold down plate 53 and as being slidably movable therealong as the trough sections I0 and II are pivotally moved with respect to each other. The direction of travel of the conveyor chain I3 and flights I4, I4 along the trough sections In and I I is thus from the trough section I I to the trough section II) with the flights I4, I4 traveling from the hold down flanges 53, 53 to and beneath the hold down plates 55, 55 and hold down flanges 5I, 5|. Each hold down plate 55 is shown as having an elongated slot 5! formed therein and opening toward the adjacent end of the side wall 2|. Each side wall 2| is shown as having a slot 58 formed therein opening to the hold down plate 55 and receiving an inner end 59 thereof when the trough sections are in the position shown in Figure 5. The slot 51 is shown as being curved to conform generally to the path of travel of the side wall 2| at the inner end of the slot 58 in all positions of articulation of the trough sections In and II with respect to each other. The slot 51 is shown' as extending along oppostie sides of the side wall from the inner end of the slot 58, as one trough section is pivoted with respect to the other about the axis of the pivot pin I2. The edges of the slot 5'! in the hold down plate 55, engaging opposite sides of the side wall 2i from within the slot 58, thus serve to maintain the hold down plate 55 in position to close the gap between the hold down flanges 5I and 53 in all positions of articulation of the conveyor. The slot 51 in cooperation with the side wall 2| also pivotally moves the closure plate 43 and maintains said closure plate in engagement with the side wall H, as the trough sections I0 and II are pivoted with respect to each other about the axis of the pivot pin I2. It will be understood that modifications and variations of the present invention may be effected without departing from the spirit and scope of the novel concepts thereof. I claim as my invention: 1. An articulated conveyor unit comprising two pivoted trough sections, each having a bottom plate and upright side walls adapted to have an endless chain and flights movable therealong, and means closing the gaps between said side walls in all positions of articulation of said trough sections with respect to each other comprising closure plates pivoted to the ends of the side walls of the rear of said trough sections and extending along the outsides of the side walls of the forward of said trough sections, hold down guides extending inwardly of said closure plates in parallel relation with respect to said bottom plates and spaced there above and having slots extending therealong opening toward said side walls for engagement with opposite sides thereof to retain said closure plates to the outsides of the side walls of the forward of said trough sections in all positions of articulation of said trough sections with respect to each other. 2. An articulated conveyor unit comprising two aligned pivoted trough sections. each having bottom pans and generally parallel upright side walls adapted to have a center strand laterally flexible endless chain and flights movable therealong between said side walls, means closing the gaps bee tween said side walls comprising closure plates pivoted to the ends of the side Walls of the rear of said trough sections and extending along the outsides of the side walls of the forward of said trough sections, hold down members extending inwardly of said side walls in parallel relation with respect to said bottom pans, and means holding down said flights in the region of articulation of said trough sections comprising hold down plates secured to said closure plates and underlapping one of said hold down members and overlapping the other and having elongated arcuate slots extending therealong opening toward the side Walls of the forward of said trough sections, said last mentioned side walls also having slots opening toward said hold down plates and receiving the same and accommodating said slots in said hold down plates to slidably engage opposite sides of said side walls and move said closure plates to follow the forward of said trough sections in all positions of articulation of the conveyor. 3. In an articulated conveyor structure, two trough sections, each having a bottom pan and upright side walls extending along each side thereof, a vertical pivot pin pivotally connecting adjacent ends of said bottom pans together, a laterally flexible center strand chain having flights projecting laterally therefrom guided for movement between said side walls on the ends of said flights, hold down flanges extending inwardly of said side walls for engagement with said flights, means closing the gaps between said side walls comprising closure plates vertically pivoted to opposite side walls of the rear of said trough sections and extending along the outsides of the side walls of the forward of said trough sections, means retaining said closure plates to move along the side walls of the forward of said trough sections and closing the gaps between said hold down flanges in all positions of articulation of said trough sections with respect to each other comprising hold down plates secured to said closure plates and extending inwardly therefrom in overlapping relation with respect to the forward of said hold down flanges and in underlapping relation with respect to the rear of said hold down flanges, said hold down plates each having an elongated slot therein having slidable engagement with opposite sides of the walls to the next adjacent trough section and pivoted thereby upon pivotal movement of said trough sections with respect to each other for maintaining a continuous hold down flange in the region of articulation of the conveyor and moving said closure plates to maintain an uninterrupted flow of material along said trough'sections. 4. In an articulated conveyor structure, two trough sections, each having a bottom pan and upright side walls extending along each side thereof, a vertical pivot pin pivotally connecting adjacent ends of said bottom pans together, a center strand chain and flights guided for movement between said side walls, hold down flanges for the ends of said flights extending inwardly of said side walls, means closing the gaps between said side walls comprising closure plates vertically pivoted to the ends of opposite side walls of the rear of said trough sections and extending along the outsides of the side walls of the forward of said trough sections, means retaining said closure plates to move along the forward of said trough swims fa'ndwlosimg the {gains between mijacent ends 'of said mold (rownfianges comprising :hol-d d'ownzplates extending mwardlyfrom said' closure pl'ai tes in ov'erlapping reIationwith-resp'et to the rorwa'rd of said-hold'down fiamge's an'd in underlapping "relation with respect 1:0 the rear of said lml'd down flanges, said 110101 down pl-ates each having an elongated sI ot'merin, and' th'ez forward 0 1 said side walls eac ih having a, slot extend mg thereaslong in registry with the associated hold down,aeeommodafimgthe wail-1s 'af saidsrcst in the associated hd ld down to r'fga'ge oppdsit'e sides of an associated side wall ma *be' piyotall-y mm ed thereby :upon pivotal mdvement of "said trough sections with respect to each other. References Cited in the file of this patent UNITED STATES PA'IEN'TS N qm'ber 4 Name 1 Date I 2,366,407 :Jeffrey in--- Jan. 2, '1'945 10 2385108 part lidge q..- Aug "1, 1945 2,512,610 Carthage June 27, 1950
Feylost From The Wild Beyond the Witchlight, page 9. Overview * Skill Proficiencies: Deception, Survival * Tool Proficiencies: One type of musical instrument * Languages: One of your choice of Elvish, Gnomish, Goblin, or Sylvan Fey Mark Feywild Visitor Feature: Feywild Connection Suggested Characteristics * See Additional Background Features.
Wikipedia:Non-free content Wikipedia's goal is to be a free content encyclopedia, with free content defined as content that does not bear copyright restrictions on the right to redistribute, study, modify and improve, or otherwise use works for any purpose in any medium, even commercially. Any content not satisfying these criteria is said to be non-free. This includes all content (including images) that is fully copyrighted, or which is made available subject to restrictions such as "non-commercial use only" or "for use on Wikipedia only". (Many images that are generally available free of charge may thus still be "non-free" for Wikipedia's purposes.) The Wikimedia Foundation uses the definition of "free" described here. The licensing policy of the Wikimedia Foundation expects all content hosted on Wikimedia projects to be free content; however, there are exceptions. The policy allows projects (with the exception of Wikimedia Commons) to adopt an exemption doctrine policy allowing the use of non-free content. Their use should be minimal and confined (with limited exceptions) to illustrating historically significant events, to include identifying protected works such as logos, or to complement (within narrow limits) articles about copyrighted contemporary works. Non-free content should not be used when a freely licensed file that serves the same purpose can reasonably be expected to be uploaded, as is the case for almost all portraits of living people. Non-free content should be replaced by free content should such emerge. The non-free content criteria policy currently serves as the exemption doctrine policy of the English Wikipedia, while this document serves to provide guidance associated with this policy. Non-free content can be used on Wikipedia in certain cases (for example, in some situations where acquiring a freely licensed image for a particular subject is not possible), but only within the United States legal doctrine of fair use, and in accordance with Wikipedia's own non-free content criteria as set out below. The use of non-free content on Wikipedia is therefore subject to purposely stricter standards than those laid down in U.S. copyright law. Non-free content criteria Transcluded from Non-free content criteria; this is the part of the current page that is official policy Implementation The implementation of the non-free content criteria is done by having two specific elements on the non-free media's description page: * A copyright license template * The copyright license template identifies the type of copyright that the original work is under license. This is necessary to demonstrate that the image complies with United States fair use laws. * You can find a list of these copyright license templates at File copyright tags/Non-free. * A non-free rationale * The rationale to use the non-free content is necessary to show that the non-free content criteria have been met. The rationale should clearly address and satisfy all ten points of WP:NFCC. A generic template, Non-free use rationale, may be used for any non-free media. * Several other boilerplate rationale templates exist, such as for albums, film posters, and logos. More can be found at Category:Non-free use rationale templates, but editors are cautioned that these are generally tenuous in terms of supporting WP:NFCC, and are encouraged to improve upon rationales if they can do so. You are not required to use the template forms, but whatever form you chose needs to clearly address all ten criteria in WP:NFCCP. * It is important to remember that a non-free rationale is needed for each use of the image on Wikipedia. If the image is used in two separate articles, two separate rationales are needed, unique for both articles. If the image is used more than once in the same article, a separate and specific rationale is still needed for each use. Both the license and the rationale need to be included on the non-free media description page. The standard upload tool for Wikipedia will ask you enough questions during the process to fill in both the license and rationale for you, thus simplifying the process. If they are added manually, with or without the help of a template, it is recommended to put the rationale and license under separate sections "Rationale" and "Licensing" respectively. Failure to include a licensing template, or a rationale that clearly identifies each article the media file is used in, will lead to the media file being deleted within 7 days after being tagged with warning messages. Meeting the no free equivalent criterion Non-free content cannot be used in cases where a free content equivalent, with an acceptable quality sufficient to serve the encyclopedic purpose, is available or could be created. As a quick test, before adding non-free content, ask yourself: "Can this non-free content be replaced by a free version that has the same effect?" and "Could the subject be adequately conveyed by properly sourced text without using the non-free content at all?" If the answer to either is yes, the non-free content probably does not meet this criterion. Note that the replacement image does not need to exist; it is sufficient that it could be created – for example, in most cases, a photograph of a living person can be taken and released under a free licence. Another consideration for "no free equivalent" are "freer" versions of non-free media, typically which include derivative works. For example, a photograph of a copyrighted 3D work of art will also carry the copyright of the photographer in addition to the copyright of the artist that created the work. We would use a photograph where the photographer has licensed their photograph under a free license, retaining the copyright of the derivative work, instead of a photograph that has non-free licenses for both the photograph and work of art. Multiple restrictions For a vector image (i.e. SVG) of a non-free logo or other design, US law is not clear as to whether the vectorisation of the logo has its own copyright which exists in addition to any copyright on the actual logo. To avoid this uncertainty, editors who upload vector images of non-free logos should use a vector image that was produced by the copyright holder of the logo and should not use a vector image from a site such as seeklogo.com or Brands of the World where the vectorisation of a logo may have been done without authorization from the logo's copyright holder. If an editor bases a vectorisation they did by themself from a free image, they should indicate the source image so that freeness can be confirmed, and release their contribution (the labour of converting to vectors) under a free license to help with the aforementioned ambiguity. Meeting the previous publication criterion Very often, such as for most non-free content emanating from the news and entertainment industries, meeting this criterion is not in question. In rare cases however, non-free content may have been originally "leaked" and never subsequently published with the copyright holder's permission—such content must not be included in Wikipedia. Usually, an accompanying copyright notice is considered sufficient evidence that a publication in the media has been made with appropriate permission. If, in this regard, an item of non-free content is questioned or is likely to be questioned, then details of an instance of prior publication with permission must be determined and recorded at the non-free media's description page. Meeting the contextual significance criterion Two of the most common circumstances in which an item of non-free content can meet the contextual significance criterion are: * where the item is itself the subject of sourced commentary in the article, or * where only by including such non-free content, can the reader identify an object, style, or behavior, that is a subject of discussion in the article. In all cases, meeting the criterion depends on the significance of the understanding afforded by the non-free content, which can be determined according to the principles of due weight and balance. To identify a subject of discussion, depiction of a prominent aspect of the subject generally suffices, thus only a single item of non-free content meets the criterion. For example, to allow identification of music albums, books, etc., only an image of the front cover art of the object is normally used; for identification of specific coins and currency, images of the front and back are normally used. Sourcing While there is no specific requirement in the non-free content policy to identify the source from which a non-free file was obtained, editors are strongly encouraged to make note of the source on the media's description page; many of the non-free rationale templates already include a field for this information. This can aid in the cases of disputed media files, or evaluating the non-free or free nature of the image. Lacking a source is not grounds for media removal, but if the nature of the media file is disputed, the lack of a source may prevent the file from being retained. Non-free media must be from a published source; the unpublished non-free media is forbidden. Identification of the source will aid in validating the previous publication of the material. The source information should be sufficiently complete to allow any editor to validate that material. While completeness is not required, editors are encouraged to provide as much source information as they can. Some ways to source media files include: * Scanned images: Identifying the published work, page numbers, and the copyright owner. * Screenshots and video clips: Identifying the movie, television show, or other video source, its copyright owner, and the approximate timestamp where the shot or clip was taken. * Images from the Internet: Identifying the URL of the image itself or web page hosting the image, and the image's copyright owner (not necessarily the same as the website's). * Music samples: Identifying the album, artist, track number, and approximate time stamp of the sample. Number of items Articles are structured and worded to minimize the total number of items of non-free content that are included within the encyclopedia, where it is reasonable to do so. For example, an excerpt of a significant artistic work is usually included only in the article about the work, which is then referenced in the articles about its performer and its publisher. A single item of non-free content that conveys multiple points of significant understanding within a topic is preferred to multiple non-free items which each convey fewer such points. This is independent of whether the topic is covered by a single article, or is split across several. For example, an article about an ensemble may warrant the inclusion of a non-free image identifying the ensemble. This is preferable to including separate non-free images for each member of the ensemble, even if the article has been split with each member having their own sub-section of the article. Image resolution There is no firm guideline on allowable resolutions for non-free content; images should be rescaled as small as possible to still be useful as identified by their rationale, and no larger. This metric is very qualitative, and thus difficult to enforce. Some legal proceedings have discussed the issue, but are inconclusive here. At the low pixel count end of the range, most common pictorial needs can be met with an image containing no more than about 100,000 pixels (0.1 megapixels), obtained by multiplying the horizontal and vertical pixel dimensions of an image. This allows, for example, images with a 4:3 aspect ratio to be shown at 320 × 240 pixels (common for screenshots from TV, films, and video games), while allowing common cover art to be shown at 250 × 400 pixels. To scale an image down to a specific number of pixels, use this formula: * $$\text{new width} = \left\lfloor\sqrt{\tfrac{\text{target pixel count} \times \text{original width}}{\text{original height}}}\right\rfloor$$ At the extreme high end of the range, non-free images where one dimension exceeds 1,000 pixels, or where the pixel count approaches 1 megapixel, will very likely require a close review to verify that the image needs that level of resolution. Editors should ensure that the image rationale fully explains the need for such a level of detail. You also may wish to add the non-free no reduce template to the image rationale page to indicate that your image resolution purposely exceeds the 0.1 megapixels guideline, though this still requires you to include a valid rationale that explains this reasoning; large images using this template without a rationale to explain the large size may nonetheless be reduced. An original, high resolution image (that can be reasonably scaled down to maintain overall artistic and critical details) may lose some text detail. In such cases, that text should be duplicated on the image description page. Care should be given to the recreation of copyrighted text: for example, while duplication is appropriate for credits from a movie poster that contains factual data, such duplication would not be appropriate for an original poem embedded within an image. If a small area of a large image needs high resolution to see details that are discussed in the article text, it may be better to crop the section to show the critical portion at a higher resolution, than to try to reduce the full image. If cropping is performed, editors should indicate the original source of the image and what modifications were made. If you believe an image is oversized, either re-upload a new version at the same file location, or tag the image file page with a Non-free reduce template, which will place it in a maintenance category to be reduced by volunteers or a bot like DatBot. Note that these guidelines apply to the resolution as stored on the image file page; the reuse of these images in mainspace should follow the Manual of Style for image use, such as deferring to default thumbnail size to allow the end-user control of the image display. Video clips Non-free video clips may be used, but besides meeting all other criteria, there must be some significant justification for using a video clip over a single image. A single image is always considered a more minimal use under NFCC#3 than a video clip if it is an equivalent replacement, so the rationale for the video must describe why a single image is not sufficient, such as capturing a specific type of motion discussed in depth by sources. One such acceptable use is the video clip used on Dennō Senshi Porygon, which is necessary to show a strobing effect discussed at depth that is impossible to capture with still images. When non-free video clips are used, they should treated similarly to non-free audio clips: limited to less than 10% of the length of the original work or 30 seconds, whichever is shorter, encoded in a lossy format and resized to meet all other NFCC#3 factors. Guideline examples Non-free content that meets all of the policy criteria above but does not fall under one of the designated categories below may or may not be allowable, depending on what the material is and how it is used. These examples are not meant to be exhaustive, and depending on the situation there are exceptions. When in doubt as to whether non-free content may be included, please make a judgement based on the spirit of the policy, not necessarily the exact wording. If you want help in assessing whether a use is acceptable, please ask at Media copyright questions. It may also be useful to ask at Wikipedia talk:Copyrights, Wikipedia talk:Copyright problems, and Wikipedia talk:Non-free content; these are places where those who understand copyright law and Wikipedia policy are likely to be watching. * See also: Wikipedia Signpost/2008-09-22/Dispatches, a guide to evaluating the acceptability of non-free images. Acceptable use The following cases are a non-exhaustive list of established examples of acceptable use of non-free media on Wikipedia. Note that the use of such media must still comply with the Non-free content criteria and provide rationales and licensing information. Text Brief quotations of copyrighted text may be used to illustrate a point, establish context, or attribute a point of view or idea. In all cases, an inline citation following the quote or the sentence where it is used is required. Copyrighted text that is used verbatim must be attributed with quotation marks or other standard notation, such as block quotes. Any alterations must be clearly marked, i.e., [brackets] for added text, an ellipsis (e.g. ) for removed text, and emphasis noted after the quotation as "(emphasis added)" or "(emphasis in the original)". Extensive quotation of copyrighted text is prohibited. Please see both WP:QUOTE for use and formatting issues in using quotations, and WP:MOSQUOTE for style guidelines related to quoting. Audio clips All non-free audio files must meet each of the non-free content criteria; failure to meet those overrides any acceptable allowance here. Advice for preparing non-free audio files for Wikipedia can be found at Music samples. The following list is non-inclusive but contains the most common cases where non-free audio samples may be used. * 1) Music clips may be used to identify a musical style, group, or iconic piece of music when accompanied by appropriate sourced commentary and attributed to the copyright holder. Samples should generally not be longer than 30 seconds or 10% of the length of the original song, whichever is shorter (see Music samples). * 2) Spoken word clips of historical events, such as speeches by public figures, may be used when accompanied by appropriate sourced commentary and attributed to the speaker/author. Images Some non-free images may be used on Wikipedia, providing they meet both the legal criteria for fair use, and Wikipedia's own guidelines for non-free content. Non-free images that reasonably could be replaced by free content images are not suitable for Wikipedia. All non-free images must meet each non-free content criterion; failure to meet those overrides any acceptable allowance here. The following list is not exhaustive but contains the most common cases where non-free images may be used and is subject to the restrictions listed below at unacceptable use of images, notably #8, which forbids the use of press agency or photo agency (e.g., AP or Getty Images) images when the image itself is not the subject of commentary. * 1) Cover art: Cover art from various items, for visual identification only in the context of critical commentary of that item (not for identification without critical commentary). * 2) Team and corporate logos: For identification. See Logos. * 3) Stamps and currency: For identification of the stamp or currency, not the subjects depicted on it. * 4) Other promotional material: Posters, programs, billboards, ads. For critical commentary. * 5) Video screenshots: For critical commentary and discussion of the work in question (i.e., films, television programs, and music videos). * 6) Screenshots from software products: For critical commentary. See Software screenshots. * 7) Paintings and other works of visual art: For critical commentary, including images illustrative of a particular technique or school. * 8) Images with iconic status or historical importance: * 9) *Iconic or historical images that are themselves the subject of sourced commentary in the article are generally appropriate. * 10) *Iconic and historical images which are not subject of commentary themselves but significantly aid in illustrating historical events may be used if they meet all aspects of the non-free content criteria, particularly no free alternatives, respect for commercial opportunity, and contextual significance. However, if the image is from a press or photo agency (e.g., AP or Getty Images) and is not itself the subject of critical commentary, it is assumed automatically to fail the "respect for commercial opportunity" test. * 11) Images that are themselves subject of commentary. * 12) Pictures of deceased persons, in articles about that person, provided that ever obtaining a free close substitute is not reasonably likely. Note that if the image is from a press or photo agency (e.g., AP or Getty Images) and is not itself the subject of critical commentary it is assumed automatically to fail "respect for commercial opportunity". Unacceptable use The following is a non-exhaustive list of examples where non-free content may not be used outside of the noted exceptions. Text * 1) Unattributed pieces of text from a copyrighted source. * 2) Excessively long copyrighted excerpts. * 3) All copyrighted text poses legal problems when making spoken word audio files from Wikipedia articles, and should be avoided in such files, because the resulting audio file cannot be licensed under the GFDL. * 4) A complete or partial recreation of "Top 100" or similar lists where the list has been selected in a creative manner. (See Copyright in lists for further details.) Articles on individual elements from such lists can discuss their inclusion in these lists. Complete lists based on factual data, such as List of highest-grossing films, are appropriate to include. Lists that have acceptable free licensing (as with AFI 100 Years... series) may be reproduced in their entirety as long as proper citations and sourcing are included. Multimedia * 1) An excessive number of short audio clips in a single article. A small number may be appropriate if each is accompanied by commentary in the accompanying text. * 2) A long audio excerpt, to illustrate a stylistic feature of a contemporary band; see above for acceptable limits. * 3) A short video excerpt from a contemporary film, without sourced commentary in the accompanying text. The use of non-free media (whether images, audio or video clips) in galleries, discographies, and navigational and user-interface elements generally fails the test for significance (criterion #8). Images * 1) Pictures of people still alive, groups still active, and buildings still standing; provided that taking a new free picture as a replacement (which is almost always considered possible) would serve the same encyclopedic purpose as the non-free image. This includes non-free promotional images. * For some retired or disbanded groups, or retired individuals whose notability rests in large part on their earlier visual appearance, a new picture may not serve the same purpose as an image taken during their career, in which case the use would be acceptable. * In considering the ability to take a free photograph, it is expected that the photographer respect all local property and privacy laws and restrictions. For example, we would not accept a free photograph of a structure on inaccessible private property that is not visible from public locations. * 1) An album cover as part of a discography, as per the above. * 2) A rose, cropped from a record album, to illustrate an article on roses. * 3) A map, scanned or traced from an atlas, to illustrate the region depicted. Use may be appropriate if the map itself is a proper subject for commentary in the article: for example, a controversial map of a disputed territory, if the controversy is discussed in the article. * 4) An image whose subject happens to be a war, to illustrate an article on the war. Use may be appropriate if the image itself is a proper subject for commentary in the article: for example, an iconic image that has received attention in its own right, if the image is discussed in the article. * 5) An image to illustrate an article passage about the image, if the image has its own article (in which case the image may be described and a link provided to the article about the image) * 6) A Barry Bonds baseball card, to illustrate the article on Barry Bonds. The use may be appropriate to illustrate a passage on the card itself; see the Billy Ripken article. * 7) A magazine or book cover, to illustrate the article on the person whose photograph is on the cover. However, if the cover itself is the subject of sourced discussion in the article, it may be appropriate if placed inline next to the commentary. Similarly, a photo of a copyrighted statue (assuming there is no freedom of panorama in the country where the statue was when the photo was taken) can only be used to discuss the statue itself, not the subject of it. * 8) An image with an unknown or unverifiable origin. This does not apply to historical images, where sometimes only secondary sources are known, as the ultimate source of some historical images may never be known with certainty. * 9) A chart or graph. These can almost always be recreated from the original data. * 10) A commercial photograph reproduced in high enough resolution to potentially undermine the ability of the copyright holder to profit from the work. * 11) Board or card game artwork and photos where the game itself is shown more than de minimis; such images can nearly always be replaced by a free de minimis photograph of the game's layout while it is being played. Exceptions are made for parts of a board or card games that have received critical commentary. * 12) An image of a newspaper article or other publication that contains long legible sections of copyrighted text. If the text is important as a source or quotation, it should be worked into the Wikipedia article in textual form, with a citation to the newspaper article. * 13) A publicity image of a commercial product released by its manufacturer, if the product has already been sold or displayed to the public in such a way that free photographs of it could be taken. * 1) An image of a newspaper article or other publication that contains long legible sections of copyrighted text. If the text is important as a source or quotation, it should be worked into the Wikipedia article in textual form, with a citation to the newspaper article. * 2) A publicity image of a commercial product released by its manufacturer, if the product has already been sold or displayed to the public in such a way that free photographs of it could be taken. Non-free image use in list articles In articles and sections of articles that consist of several small sections of information for a series of elements common to a topic, such as a list of characters in a fictional work, non-free images should be used judiciously to present the key visual aspects of the topic. It is inadvisable to provide a non-free image for each entry in such an article or section. The following considerations should be made to reduce the number of new non-free images associated with such lists: * 1) Images that show multiple elements of the list at the same time, such as a cast shot or montage for a television show, are strongly preferred over individual images. Such an image should be provided by the copyright holder or scanned/captured directly from the copyrighted work, instead of being created from multiple non-free images by the user directly (as the "extent" of use is determined by the number and resolution of non-free images, and not the number of files.) * 2) Images which are discussed in detail in the context of the article body, such as a discussion of the art style, or a contentious element of the work, are preferable to those that simply provide visual identification of the elements. * 3) An image that provides a representative visual reference for other elements in the article, such as what an alien race may look like on a science-fiction television show, is preferred over providing a picture of each element discussed. * 4) If another non-free image of an element of an article is used elsewhere within Wikipedia, referring to its other use is preferred over repeating its use on the list and/or including a new, separate, non-free image. If duplicating the use of a non-free image, please be aware that a separate non-free use rationale must be supplied for the image for the new use. * 5) For media that involves live actors, do not supply an image of the actor in their role if an appropriate free image of the actor exists on their page (as per WP:BLP and above), if there is little difference in appearance between actor and role. However, if there is a significant difference due to age or makeup and costuming, then, when needed, it may be appropriate to include a non-free image to demonstrate the role of the actor in that media. * 6) Barring the above, images that are used only to visually identify elements in the article should be used as sparingly as possible. Consider restricting such uses to major characters and elements or those that cannot be described easily in text, as agreed to by editor consensus. Non-free image use in galleries or tables The use of non-free images arranged in a gallery or tabular format is usually unacceptable, but should be considered on a case-by-case basis. Exceptions should be very well-justified and alternate forms of presentation (including with fewer images) strongly considered. In categories that include non-free content, MediaWiki's code should be used to disable the display of the content while still listing it. User-created montages containing non-free images should be avoided for similar reasons. Within the scope of NFCC#3a, such montages are considered as multiple non-free images based on each non-free image that contributes towards the montage. If a montage is determined to be appropriate, each contributing non-free item should have its source described (such as File:Versions of the Doctor.jpg). A montage created by the copyright holder of the images used to create the montage is considered a single non-free item and not separate items. Exemptions Certain non-article pages are exempt from the non-free content policy. These uses are necessary for creating or managing the encyclopedia, such as special pages and categories that are used to review questionable non-free content uses. Categories that are exempt are listed in Category:Wikipedia non-free content criteria exemptions. Due to software limitations, TimedText pages for non-free video files will automatically include the video file, and as such, pages in the TimedText namespace are presumed to be exempted from NFCC#9. Fair use rationales are not required for such pages. Article images may appear in article preview popups. Background "Free" content is defined as that which meets the "Definition of Free Cultural Works". Material that is not free is permitted only if it meets the restrictions of this policy. This has been explicitly declared since May 2005. The stated mission of the Wikimedia Foundation, which supports Wikipedia servers and software, is "to empower and engage people around the world to collect and develop educational content under a free license or in the public domain, and to disseminate it effectively and globally." These concerns are embodied in the above requirements that all non-free content must meet, and our policy of deleting non-compliant content. Being generous to the world sometimes means being hard on ourselves. Please understand that these rules are not arbitrary; they are central to our mission. Wikipedia distributes content throughout the world with no restrictions on how people use it. Legally, we could use any copyrighted material for ourselves that is either licensed to us by the owner, or that fits the definition of "fair use" under US copyright law. However, we favor content that everyone can use, not just Wikipedia. We want them to be free to use, redistribute, or modify the content, for any purpose, without significant legal restrictions, particularly those of copyright. To honor its mission, Wikipedia accepts incoming copyright licenses only if they meet Wikipedia's definition of "free" use. This is a higher standard than we would need just for our own use. But our ability to use a work does not guarantee that others may use it. We reject licenses that limit use exclusively to Wikipedia or for non-commercial purposes. Commercial use is a complex issue that goes well beyond a company's for-profit status, another reason to be careful. In fact, we reject any licenses with significant limitations. That is not free enough. Similarly, Wikipedia imposes higher fair-use standards on itself than U.S. copyright law. There are some works, such as important photographs, significant modern artworks, that we cannot realistically expect to be released under a free content license, but that are hard to discuss in an educational context without including examples from the media itself. In other cases such as cover art/product packaging, a non-free work is needed to discuss a related subject. This policy allows such material to be used if it meets U.S. legal tests for fair use, but we impose additional limitations. Just because something is "fair use" on a Wikipedia article in the U.S. does not mean it is fair use in another context. A downstream user's commercial use of content in a commercial setting may be illegal even if our noncommercial use is legal. Use in another country with different fair use and fair dealing laws may be illegal as well. That would fail our mission. We therefore limit the media content we offer, to make sure what we do offer has the widest possible legal distribution. We do not want downstream re-users to rely solely on our assurances. They are liable for their own actions, no matter what we tell them. We therefore show them and let them make their own decision. To that end we require a copyright tag describing the nature of a copyrighted work, sourcing material saying exactly where any non-free content comes from, and a detailed non-free media rationale for every use of copyrighted content in every article, justifying why use in that article is permitted. A further goal of minimizing licensed and fair-use material is to encourage creation of original new content, rather than relying on borrowed content that comes with restrictions. In general Under United States copyright law, creative works published in the United States prior to are in the public domain. Some creative works published in the United States between and 1963 are still copyrighted. It is illegal (among other things) to reproduce or make derivative works of copyrighted works without legal justification. Unless a thorough search is conducted to determine that a copyright has expired or not been renewed, it should be regarded as copyrighted. Certain works have no copyright at all. Most material published in the United States before, work published before 1978 without a copyright notice, with an expired copyright term, or produced by the U.S. federal government is public domain, i.e., has no copyright. Some such as photos and scans of 2-dimensional objects and other "slavish reproductions", short text phrases, typographic logos, and product designs, do not have a sufficient degree of creativity apart from their functional aspects to have a copyright. Copyright law only governs creative expressions that are "fixed in a tangible medium of expression," not the ideas or information behind the works. It is legal to reformulate ideas based on written texts, or create images or recordings inspired by others, as long as there is no copying (see plagiarism for how much reformulation is necessary). If material does have a copyright, it may only be copied or distributed under a license (permission) from the copyright holder, or under the doctrine of fair use. If there is a valid license, the user must stay within the scope of the license (which may include limitations on amount of use, geographic or business territory, time period, nature of use, etc.). Fair use, by contrast, is a limited right to use copyrighted works without permission, highly dependent on the specific circumstances of the work and the use in question. It is a doctrine incorporated as a clause in United States copyright code, arising out of a concern that strict application of copyright law would limit criticism, commentary, scholarship, and other important free speech rights. A comparable concept of fair dealing exists in some other countries, where standards may vary. Anything published or later in other countries and still copyrighted there, is typically also copyrighted in the United States. See Non-U.S. copyrights. Applied to Wikipedia Never use materials that infringe the copyrights of others. This could create legal liabilities and seriously hurt the project. Uploading an image, audio or video file, or text quotation into Wikipedia, and adding that file to a project page, both raise copyright concerns. Editors who do either must make sure their contributions are legal. If there is any doubt as to legality, ask others for help, try to find a free equivalent, or use your own words to make the same point. Also, consider asking the copyright holder to release the work under an appropriate Creative Commons Attribution-Sharealike 3.0 Unported License (CC BY-SA) or a CC BY-SA-compatible license (dual-licensing under a GFDL license is also possible). See Boilerplate request for permission for a sample form letter. If a work has no copyright or is licensed to Wikipedia under an acceptable "free" license, it is a free work and may be used on Wikipedia without copyright concerns. See public domain, copyright, and Cornell University's guide to copyright terms for discussion of works that are not covered by copyright. Also see free license regarding free licenses and Image copyright tags/Free licenses for a list of copyright tags for these works. Restricted licenses to these works offer some legal rights, but Wikipedia ignores them because they are not free enough for its purposes. Instead, works covered by inadequate licenses are treated the same on Wikipedia as works with no licenses at all. If a work is not free, Wikipedia requires that it comply with Wikipedia's non-free use policy. As explained above, this policy is more restrictive than US law requires. Logically, material that satisfies the policy should also satisfy legal requirements as well. However, to be more certain of avoiding legal liability, and to understand the meaning of Wikipedia policy, editors should consider the legal rules as well. See fair use for further information, and the Stanford University summary of relevant cases, on the subject of fair use. Non-free material is used only if, in addition to other restrictions, we firmly believe that the use would be deemed fair use if we were taken to court. The Wikimedia Foundation reserves the right to remove unfree copyrighted content at any time. Note that citation sources and external links raise other copyright concerns that are addressed in other policies. Handling inappropriate use of non-free content Possibly inappropriate uses of non-free content can be either tagged with {{subst:proposed deletion}}, if deletion is uncontroversial per Proposed deletion, or reported and discussed at Files for discussion. Other Wikimedia projects This policy is specific to the English-language Wikipedia. Other Wikimedia projects, including Wikipedias in other languages, may have different policies on non-free content. A list of some of the projects and their policies on fair use can be read at Wikimedia Meta-Wiki.
FIG. 34 shows that after 21 weeks post-transplant, or in-life, the serum human C-peptide levels after fasting (first of three bars from the left for each animal), thirty-minutes (30′; second bar) and one hour (60′; third bar) following glucose administration or challenge. At 21 weeks post-implant, all but two animals (Animal No. 4317 and 4323; one from each cohort) produced robust levels of insulin as observed by mostly greater than 932 to 2691 pM of human sera C-peptide. However, when the stimulation indices are compared between the two groups (the difference between the 30′ or 60′ C-peptide serum level to that over the fasting serum C-peptide level) there was no significant difference in the in vivo function between the re-aggregated cell culture and Control cohorts. So, whether the transplanted cultures or grafts consisted of about 98% endocr ine/ CHGA+ re-aggregated cell culture or about 87% endocr ine/CHGA+ (Control) both appear to have in vivo function. Thus, Applicants have demonstrated for the first time that an in vitro endocr ine (CHGA+) cell population (a properly specified endocr ine cell) can give rise to physiologically functional insulin secreting cells when implanted. Thus, for the first time, Applicants have demonstrated that an in vitro population of properly specified endocr ine cells, which when transplanted develop and mature to functional pancreatic islet cells that secrete insulin in response to blood glucose. Further, the in vivo function is similar to that previously described by Applicant for PEC and reported in Schulz et al. (2012), supra. Further, it appears that so long as the properly specified endocr ine cells were made by the methods described herein via stages 1-7, additional enrichment of endocr ine cells (or depletion of non-endocr ine cells) via disassociation and re-aggregation is not necessary or required for in vivo function. These aspects of the invention have never been demonstrated up until now. Up until now, Applicants and others had only demonstrated in vivo function using a pancreatic progenitor population cell population and not an endocr ine cell population. Indeed, as described in Kelly et al. (2011) supra shows that the endocr ine (CHGA+) sub-populations of PEC or pancreatic progenitor preparations did not give rise to in vivo function. Example 19 Complex Base Media for Culturing Endocr ine Cells Complex base media such as CMRL and RMPI are widely used for culturing islet cells and preserving islet cell mass. CMRL media is typically more complex (more components or ingredients) than RPMI. And both CMRL and RMPI are more complex than DMEM, a commonly used and less complex medium for cell culture. Thus, Applicants explored whether such complex base media would be beneficial to maintain endocr ine cells during the later stages of differentiation, for example, stages 6 and 7. Islet transplantation requires culturing and engraftment of islets with adequate islet cell mass and minimal toxicity. Short-term islet culture has been linked to rapid degradation, and loss of viability and glucose control. See, S. Matsumoto, et al. (2003), A comparative evaluation of culture conditions for short-term maintenance (<24 hr.) of human islets isolated using the Edmonton protocol, Cell Tissue Banking, 4, p. 85.; and N.J.M. London, et al. (1998), Isolation, Culture and Functional Evaluation of Islets of Langerhans, Diabetes & Metabolism, 23, 200-207. In 1978, a study by Andersson et al. (1978) compared the effectiveness of TCM, RPMI, CMRL, DMEM and Hams F10 for culturing mouse islets. See Andersson A. (1978) Isolated mouse pancreatic islets in culture: Effects of serum and different culture media on the insulin production of the islets, Diabetelogia, 14, 397-404. Andersson’s results showed that Ham’s F10 provided for islet cultures with the most insulin content, but that RMPI provided for cultures with the best insulin biosynthesis rate; and that this was possibly due to the high glucose and nicotinamide in the media. Still other investigators such as Davalli et al. (1992) showed that TCM with adenosine phosphate and xanthine were better for culturing porcine islets than CMRL and RMPI. See Davalli, A.M. et al. (1992), In vitro function of adult pig islets: Effect of culture in different media, Transplantation, 60, 854-860. Davalii et al. further suggested that the high glucose found in Ham’s F10 and RMPI were possibly toxic to the porcine islets. Hence, islets from different species may have different culture requirements and there is no one formula suitable for all endocr ine and islet cells. See Andersson (1978), p.4 supra. Table 20 below describes only those components found in CMRL, RPMI and DMEM media which vary between each of the media types, i.e. Table 20 does not list those components shared in all 3 types of base media, but only listing those components which are contained in CMRL but not contained in RPMI and/or DMEM. Table 20 shows that CMRL is the most complex (greatest number of components) and DMEM is the least complex (the least number of components). Y, indicates presence of that component in that base media; and N, indicates the component is not present in the media. In brief, DMEM is lacking about 7 components (N, shaded boxes) which are present in CMRL and RPMI. TABLE 20 Comparison of CMRL, RPMI and DMEM cell culture media CMRL (cb) RPMI DMEM (db) Amino Acids Hyd roxy L-proline Y N L-Alanine N N L-Aspartic acid Y N L-Cysteine hyd rochloride-H20 N N L-Glutamic Acid Y N L-Proline Y N Vitamins Ascorbic Acid N N Biotin Y N Cholesterol N N Nicotinic Acid (Niacin) N N Para-Aminobenzoic Acid Y N Pyridoxal hyd rochloride N N Inorganic Salts Sodium Phosphaye mono basic (NaH2PO4-H2O) N Y Other Components 2′ Deoxyadenosine N Y 2′ Deoxycytidine N N 2′ Deoxyguanosine N N 5-Methyl-deoxycytidine N N Co-carboxylase N N Coenzyme A N N Diphosphopyridine nucleotide (NAD) N N FAD (flavin adenine dinucleotide) N N Glutathione (reduced) Y N Sodium acetate-3H2O N N Sodium glucuronate-H2O N N Thymidine N N Triphosphopyridine Nucleotide (NADP) N N Tween 80® N N Uridine 5′- triphosphate N N Methods for making endocr ine cell cultures were substantially similar to that described above except that during stage 6 (d15), or mid stage 7 (d23), or mid stage 7 (d26), or stages 6 and 7 (d15 to d25), DMEM or CMRL base media was used. In addition, some cultures received insulin like growth factor (IGF), nicotinamide (NC), glucose (Glc), and/or a rho-kinase inhibitor (Y-27632). In one experiment, stage 7 cells, starting at about d23, were cultured in either CMRL or DMEM base media with B27 supplement, Matrigel, BMP, IGF and Y-27632. Nanostring analysis of d26 (stage 7) samples showed that, in general, stage 7 cells cultured in CMRL had increased nonspecific gene expression of both hormone and pancreatic endocr ine markers as compared to the same stage cells when cultured in DMEM. See FIGS. 35A-C. Thus, CMRL may be important for endocr ine cell maintenance. Corresponding studies were also performed using RPMI as compared to DMEM and CMRL-based media (data not shown). CMRL and RMPI medias had similar effects on the cell cultures and both were improved over that of DMEM, thus the component(s) which are not present in DMEM and present in both CMRL and RPMI may be important for endocr ine cell maintenance during stage 6 and/or 7 including reduced glutathione, amino acids, hyd roxyl-L-Proline, L-Proline, L-Aspartic Acid, L-Glutamic Acid and vitamins, biotin and para-aminobenzoic acid. Example 20 Cryopreservation Does Not Reduce Endocr ine Cell Numbers A commercially viable cell therapy will require scale-up manufacturing of the cell product and a means to store the product long term for on-demand patient needs. Hence, any cell product, be it the cells alone or encapsulated cells, such as the ENCAPTRA® drug delivery device, may need to be cryopreserved or have other means for long-term storage without detrimental or toxic effects to the cells; and without affecting the cells’ therapeutic effect or in this case in vivo production of insulin in response to blood glucose. Thus, Applicant’s explored whether cryopreserved stage 7 endocr ine cells maintain their ability to develop and function in vivo similar to that observed for fresh (non-cryopreserved) and cryopreserved PEC as described in detail in Applicant’s U.S. Pat. Nos. 8,278,106 and 8,425,928, both titled, ENCAPSULATION OF PANCREATIC CELLS DERIVED FROM HUMAN PLURIPOTENT STEM CELLS, issued Oct. 2, 2012 and Apr. 23, 2013, respectively; and U.S. Application 61/775,480, titled CRYOPRESERVATION, HIBERNATION AND ROOM TERMPERATURE STORAGE OF ENCAPSULATED PANCREATIC ENDODERM CELL AGGREGATES, filed Mar. 7, 2013, and Kroon et al (2008) supra. Methods for making endocr ine cell cultures were substantially similar to that described above except cell cultures were all in CMRL based media during stage 7 and split into the following samples: 1) Not dissociated or re-aggregated but cryopreserved at about day 23 for several hours and then thawed and cultured for several days in CMRL base media with B27 supplement, BMP, TTNPB, and a rho-kinase inhibitor (No re-agg / Cryopreserved); 2) Not dissociated or re-aggregated and not cryopreserved (No-rea gg / No cryopreserved); and 3) Dissociated and re-aggregated and not cryopreserved (Re-agg / No cryopreserved). All 3 samples above were cultured in the same stage 7 media conditions. Samples were analyzed by flow cytometry on day 27. See Table 21 below. TABLE 21 Flow cytometry analysis of cryopreserved, fresh and re-aggregated stage 7 endocr ine cells Stage 7 endocr ine/non-Endocr ine flow cytometry Total Endocr ine (ChroA+) ChroA+ Nkx6.1-INS- ChroA+ NKx6.1-INS+ ChroA+ NKx6.1+ INS- ChroA+ Nkx6.1+ INS+ Total ChroA- ChroA-Nkx6.1+ INS- ChroA-Nkx6.1+ INS+ ChroA-Nkx6.1-INS+ ChroA-Nkx6.1-INS- 85.2 17.5 9.02 20.9 37.8 14.8 8.65 0.29 0.25 5.6 To lal Nkx6.1+ ChroA- Nkx6.1+INS ChroA- Nkx6.1+IN ChroA- Nkx6.1+IN ChroA- Nkx6.1+IN To lal Nkx6.1+ ChroA- Nkx6.1+IN ChroA- Nkx6.1+IN ChroA- Nkx6.1+IN ChroA- Nkx6.1+IN 67.6 8.65 0.29 20.9 37.8 32.3 17.5 0.25 5.6 9.02 Total INS+ ChroA- Nkx6.1+INS ChroA- Nkx6.1+IN ChroA- Nkx6.1+IN ChroA- Nkx6.1+IN Total INS+ ChroA- Nkx6.1+IN ChroA- Nkx6.1+IN ChroA- Nkx6.1+IN ChroA- Nkx6.1+IN 47.3 0.25 0.29 9.02 37.8 52.6 17.5 20.9 8.65 5.6 Stage 4 endocr ine/non-Endocr ine flow cytometry Total Endocr ine (ChroA+) ChroA+ Nkx6.1-INS- ChroA+ NKx6.1-INS+ ChroA+ NKx6.1+ INS- ChroA+ Nkx6.1+ INS+ Total ChroA- ChroA-Nkx6.1+ INS- ChroA-Nkx6.1+ INS+ ChroA-Nkx6.1-INS+ ChroA-Nkx6.1-INS- 39.4 15.0 7.74 8.78 7.85 60.1 0.57 46.1 0.23 13.2 To lal Nkx6.1+ ChroA- Nkx6.1+INS ChroA- Nkx6.1+IN ChroA- Nkx6.1+IN ChroA- Nkx6.1+IN To lal Nkx6.1+ ChroA- Nkx6.1+IN ChroA- Nkx6.1+IN ChroA- Nkx6.1+IN ChroA- Nkx6.1+IN 63.3 0.57 46.1 8.78 7.85 36.2 15.0 0.23 13.2 7.74 Total INS+ ChroA- Nkx6.1+INS ChroA- Nkx6.1+IN ChroA- Nkx6.1+IN ChroA- Nkx6.1+IN Total INS+ ChroA- Nkx6.1+IN ChroA- Nkx6.1+IN ChroA- Nkx6.1+IN ChroA- Nkx6.1+IN 16.4 0.23 0.57 7.74 7.85 83.1 15.0 8.78 46.1 13.2 Sample (d27) Total Endocr ine (CHGA+) ChroA+ Nkx6.1+ PDX1-/+ ChroA-NKx6.1-PDX1+ 1.Cryopreserved, No re-agg. 89.97 5.61 3.51 2. Fresh, No re-agg. (Control) 3. Fresh, 88.57 5.71 4.39 Re-agg. 92.85 3.69 2.05 Based on the flow cytometry analysis in Table 21, for those cultures which were not dissociated or re-aggregated (samples 1 and 2), cryopreserving the cells (sample 2) does not reduce the total number of endocr ine (CHGA+) cells as compared to the fresh cells (sample 1) (88.57 vs. 89.97). As expected, the total endocr ine (CHGA+) population is higher in cultures that had been dissociated and re-aggregated, and not cryopreserved (92.85). Thus, dissociating and re-aggregating affects endocr ine (CHGA+) cell numbers as described above in Example 17, but cryopreserving the stage 7 cultures does not appear to alter the cell composition. Further, samples were analyzed using cell immunohistochemistry with NKX6.1, C-peptide, INS, and GCG staining. See FIGS. 36A-36B. The endocr ine (CHGA+) cell aggregates from stage 7 (No re-agg / No cryopreserved; sample 2) showed NKX6.1 (nuclear) and C-peptide (cytoplasmic) co-expression or co-staining (FIGS. 36A-36C). Co-staining was not observed in similar analysis of PEC from a stage 4 differentiation protocol. Cell cultures from stage 7 were also primarily singly hormonal, as can be seen in FIGS. 36A-36C, which shows separate staining of INS (36A; cytoplasmic) and GCG (36B; cytoplasmic) cells, i.e. the majority of individual cells did not co-express INS and GCG as seen with endocr ine (CHGA+) sub-populations of PEC from stage 4. FIG. 37 shows photomicrographs with similar staining patterns but from cell cultures which had been dissociated and re-aggregated at the start of stage 7. FIG. 37 compares graft function from cell cultures 1 and 3 of Table 20 above. Analysis of the in vivo function of the transplanted stage 7 grafts of both cohorts of animals at 12 weeks showed that cultures which had been cryopreserved but not re-aggregated (Sample 1) had similar serum C-peptide levels at 30 and 60 minutes post glucose challenge as compared to those cultures which had not been cryopreserved but were re-aggregated (Sample 3). These C-peptide levels are comparable to those observed for weeks 10 and 15 in FIGS. 20 and 21 (e.g. at week 15, the average C-peptide level was 995 pM at sixty minutes post-glucose administration). So, it is expected and anticipated that a later analysis will show that C-peptide levels will increase to above 1000 pM at 30 and/or 60 minutes, in most of the animals, post glucose stimulation similar to that described in Example 18 and observed in FIG. 34 . In fact, animal no. 4490, a Sample 1 graft, had serum C-peptide levels at 12 weeks that were comparable to animal no. 4320 in Example 18 (FIG. 34 ) after 21 weeks post implant. Thus, cryopreserving stage 7 endocr ine cells does not appear to affect their in vivo function. Example 21 Increasing Glucose Concentrations Affects Hormone Expression Glucose levels in cell culture formulations range from 1 g/L (5.5 mM) to 10 g/L (55 mM), with some media having about 5.5 mM glucose which approximates normal blood sugar levels in vivo. Glucose levels approaching 10 mM are pre-diabetic levels and those above 10 mM are analogous to a diabetic condition. Stated in another way, above 10 mM glucose mimics hyperglycemic conditions in vivo. High glucose can cause, for example, post-translational secondary modifications including glycation, gly oxidation and carbonyl stress. Since the reports of the effects of high-glucose on different cell types in vitro vary, Applicant’s sought to study the effects of high glucose (i.e. 4.5 g/L or 24.98 mM glucose) on stage 6 and/or 7 endocr ine cells. Glucose is a soluble hexose sugar added to all cell culture media including Ames’ Medium; Basal Medium Eagle (BME); BG Jb Medium Fitton-Jackson Modification; Click’s Medium; CMRL-1066 Medium; Dulbecco’s Modified Eagle’s Medium (DMEM); DMEM/Ham’s Nutrient Mixture F-12 (50:50); F-12 Coon’s Modification; Fischer’s Medium; H-Y Medium (Hy bri-Max®); Iscove’s Modified Dulbecco’s Medium (IMDM); McCoy’s 5A Modified Medium; MCDB Media; Medium 199; Minimum Essential Medium Eagle (EMEM); NCTC Medium; Nutrient Mixture, Ham’s F-10; Nutrient Mixture, Ham’s F-12; Nutrient Mixture Ham’s F-12 Kaighn’s Modification (F12K); RPMI-1640; Serum-Free/Protein Free Hy bridoma Medium; Waymouth Medium MB; Williams Medium E and various proprietary media. L-15 Medium contains galactose in place of glucose. See Sigma-Aldrich Media Expert, available on the world wide web at sigma aldrich.com/life-science/cell-culture/learning-center/media-expert/glucose.html. Media containing greater than 10 mM levels of glucose supplementation include at least DMEM/Ham’s Nutrient Mixture F-12 (50:50) contains 17.5 mM of glucose; DMEM (Hi Glucose), GMEM and IMDM all contain 25 mM levels of glucose; and H-Y Medium (Hy bri-Max®) and Serum-Free/Protein Free Hy bridoma Medium contain 22.6 and 28.9 mM glucose, respectively. See Sigma-Aldrich Media Expert, available on the world wide web at sigma aldrich.com/life-science/cell-culture/learning-center/media-expert/glucose.html. Because different cell culture media contain widely different levels of glucose, and the effects of glucose in general vary from one cell culture system to the next, the effects of high glucose on endocr ine cells were studied. Again, methods for making endocr ine cell cultures were substantially similar to that described above except during stages 6 and/or 7, cell cultures were treated with CMRL base medium with and without exogenous high-glucose (total glucose concentration 4.5 g/L, or 25 mM). So, this is 3.5 g/L in addition to the 1000 mg/L or 1 g/L (5.5 mM) of glucose already found in CMRL media. Both conditions with and without exogenous glucose were treated with the same growth factors including Nicotinamide, BMP, RA (or TTNPB) and alternatively a rho-kinase inhibitor and Matrigel. Nanostring analysis showed that when high-glucose was added at the start of stage 6, by the end of stage 6 there was increased expression of INS, GCG and SST and decreased expression of GHRL (FIG. 39A and FIG. 39B). Thus, increased exogenous glucose during stages 6 and 7, with the exception of Grehlin, increases hormone expression. Example 22 Production of Immature Beta-Cells Examples 8-21 describe various iterative methods for production of endocr ine cells in vitro, including use of high Activin alone or combined with Wnt and/or Heregulin at stage 3; low Activin at stage 4 alone or combined with Heregulin; Noggin and a gamma secretase inhibitor with or without KGF, EGF and a rho-kinase inhibitor at stage 5; and one or more of the following nicotinamide, retinoic acid, BMP and Matrigel at stages 6-7. Endocr ine cell populations produced from such methods are not only singly-hormonal (e.g. INS only, GCG only or SST only; see FIG. 39 ) but also co-express other immature endocr ine cell markers including NKX6.1 and PDX1. Flow cytometry analysis was performed on two different stage 7 cultures using INS, SST and GCG hormone staining (data set A and data set B). The one set of data (A) was performed after many iterations of stages 1-7 were examined and the other set of data (B) was performed earlier in time e.g., in the absence of nicotinamide, BMP, CMRL and the like. See Table 22 below. Table 22 shows total percentages of INS, SST and GCG positive staining (see 5 left columns), total INS, SST and GCG negative staining (see 5 right columns), and single hormone staining (see middle column). The hormone staining from these two data sets (A and B) is compared to that obtained from stage 4 PEC cultures that had been maintained in culture until about day 26. TABLE 22 Comparison of Hormone Expressing Cells from Stage 7 and Stage 4 (PEC) Stage 7 (d27) Hormone flow cytometry (A) Total INS+ INS+ SST-GCG- INS+ SST-GCG+ INS+ SST+ GCG- INS+ SST+ GCG+ INS Only Total INS- INS-SST+ GCG- INS-SST+ GCG- INS-SST-GCG+ INS-SST-GCG- 49.3 22.7 11.1 11.6 3.95 22.7 50.6 5.65 1.21 7.27 36.5 Total SST+ INS-SST+ GCG- INS-SST+ GCG+ INS+ SST+ GCG+ INS+ SST+ GCG- SST Only Total SST- INS-SST-GCG+ INS-SST-GCG- INS+ SST-GCG- INS+ SST-GCG+ 22.4 5.65 1.21 3.95 11.6 5.65 77.6 7.27 36.5 22.7 11.1 Total GCG+ INS-SST-GCG+ INS+ SST-GCG+ INS+ SST+ GCG+ INS-SST+ GCG+ GCG Only Total GCG- INS-SST+ GCG- INS+ SST-GCG- INS+ SST+ GCG- INS-SST-GCG- 23.5 7.27 11.1 3.95 1.21 7.27 76.5 5.65 22.7 11.6 36.5 Stage 7 (d29) Horemone floweytomentry (B) Total INS+ INS+ SST-GCG- INS+ SST-GCG+ INS+ SST+ GCG- INS+ SST+ GCG+ INS Only Total INS- INS-SST+ GCG- INS-SST+ GCG+ INS-SST-GCG+ INS-SST-GCG- 22.7 12.9 4.3 4.3 1.08 12.9 77.3 1.68 0.46 2.57 72.6 Total SST+ INS-SST+ GCG- INS-SST+ GCG+ INS+ SST+ GCG+ INS+ SST+ GCG- SST Only Total SST- INS-SST-GCG+ INS-SST-GCG- INS+ SST-GCG- INS+ SST-GCG+ 7.5 1.68 0.46 1.08 4.3 1.68 92.4 2.57 72.6 12.9 4.3 Total GCG+ INS-SST-GCG+ INS+ SST-GCG+ INS+ SST+ GCG+ INS-SST+ GCG+ GCG Only Total GCG- INS-SST+ GCG- INS+ SST-GCG- INS+ SST+ GCG- INS-SST-GCG- 8.4 2.57 4.3 1.08 0.46 2.57 91.5 1.68 12.9 4.3 72.6 Stage 4 (d36) Hormone flow cytometry (PEC) Total INS+ INS+ SST-GCG- INS+ SST-GCG+ INS+ SST+ GCG- INS+ SST+ GCG+ INS Only Total INS- INS-SST+ GCG- INS-SST+ GCG+ INS-SST-GCG+ INS-SST-GCG- 15.8 3.69 5.14 1.93 4.99 3.7 84.3 2.26 3.52 8.72 69.8 Total SST+ INS-SST+ GCG- INS-SST+ GCG+ INS+ SST+ GCG+ INS+ SST+ GCG- SST Only Total SST- INS-SST-GCG+ INS-SST-GCG- INS+ SST-GCG- INS+ SST-GCG+ 12.7 2.26 3.52 4.99 1.93 2.26 87.4 8.72 69.8 3.7 5.1 Total GCG+ INS- SST- GCG+ INS+ SST- GCG+ INS+ SST+ GCG+ INS-SST+ GCG+ GCG only GCG- INS- SST+ GCG- INS+ SST- GCG- INS+ SST+ GCG- INS- SST- GCG- 22.4 8.72 5.1 4.99 3.52 8.72 77.7 2.26 3.7 1.9 69.8 In general, the total INS positive sub-population (the sum of singly INS-positive plus INS/SST co-positive plus INS/GCG co-positive plus INS/SST/GCG co-positive) are as much as 3-fold greater in stage 7 endocr ine cultures than in stage 4 PEC cultures (49.3 vs. 22.7 vs. 15.8). Accordingly, the total INS only (no co-expression with SST or GCG) sub-population is also higher for stage 7 endocr ine cultures than for stage 4 PEC cultures (22.7 vs. 12.9 vs. 3.7). Importantly, stage 7 cultures have more singly hormone expressing INS cells as a percentage of the total INS population (22.7/49.3= 46% in A and 12.7/22.7 = -57% in B) as compared to the stage 4 PEC cultures (3.7/15.8=23%). Hence, it appears that stage 7 endocr ine cell cultures consist of more singly hormone expressing cells, particularly, INS only expressing cells, than stage 4 PEC cultures. And at least Examples 18 and 20 demonstrate that stage 7 endocr ine cells, although still immature, do develop and mature to physiologically functional pancreatic islets capable of secreting insulin in response to blood glucose when transplanted; whereas the endocr ine (CHGA+) sub-populations of stage 4 PEC cultures are not capable of the same in vivo. See FIG. 44 and FIG. 45 and Examples above for more detail. Example 23 The Endocr ine (CHGA+) Cells From Stage 7 and Stage 4 (PEC) Are Distinguished from Each Other Although analysis of stage 7 (properly specified endocr ine) and stage 4 (PEC) cultures use the same antibodies (e.g. CHGA, INS, NKX6.1) for cell staining and analysis, properly specified endocr ine (CHGA+) cell sub-populations from stages 7 and 4 (PEC) are not the same. Earlier, Applicants reported that the origin of the functional beta cells observed after transplantation of stage 4 PEC cultures was due to pancreatic progenitors or non-endocr ine (CHGA-) cells and not from the primary endocr ine (CHGA+) cells. Kelly et al. (2011), supra, demonstrated that when PEC cultures were enriched (purified) for endocr ine (CHGA+) sub-populations and transplanted, they did not give rise to functional pancreatic islets or beta cells as compared to the un-enriched PEC cultures. See Kelly et al (2011), pp.3-4, FIG. 4 and Supplementary Table 1; Schulz et al. (2012) supra; and U.S. Pat. 7,534,608; 7,695,965; 7993920; and 8,278,106. Examples 17 and 18 and Table 19 described a nearly pure (98.2%) endocr ine (CHGA+) cell culture from stage 7 that when transplanted resulted in a robust functional graft in vivo (FIG. 34 ). In view of Examples 17 and 18 and other results described herein, the functional population of the stage 7 cultures is the endocr ine (CHGA+) cells from stage 7 and not the significantly smaller percentage of non-endocr ine (CHGA-) cells (2.08%). This is in contrast to the stage 4 PEC cultures, which Applicant’s had reported in detail in prior disclosures, where the in vivo function is attributed to the non-endocr ine (CHGA-) sub-populations. Therefore, in general, the endocr ine sub-populations from stages 7 and 4 are not comparable or the same. Table 23 shows flow cytometry data comparing total endocr ine (CHGA+) and non-endocr ine (CHGA-) sub-populations from stage 7 (properly specified endocr ine) and stage 4 (PEC) and demonstrates that the sub-populations are not equivalent and are distinguished from each other. For example, the total percentage of CHGA+ sub-population was significantly higher for stage 7 as compared to the stage 4 cultures (85.2 vs. 39.4). Accordingly, the total percentage of CHGA- sub-population was significantly lower for stage 7 as compared to stage 4 cultures (14.8 vs. 60.1). Further, true endocr ine cells, e.g. beta cells, not only express CHGA+ but also co-express at least INS and NKX6.1. Thus, a true endocr ine cell expresses CHGA+/INS+/NKX6.1+ (triple positive) and is able to function in vivo. Table 23 shows that stage 7 cultures had almost 5-fold more CHGA+/INS+/NKX6.1+ (triple positive) cells than stage 4 cultures (37.8 vs. 7.8); and even though the stage 4 CHGA+/INS+/NKX6.1+ sub-populations may appear properly specified, as discussed above, these endocr ine sub-populations from stage 4 do not function in vivo when transplanted, while the non-endocr ine (CHGA-) sub-populations from stage 4 do mature and function in vivo. Thus, Applicant’s refer to stage 7 endocr ine cells as “immature endocr ine cells” or “immature beta cells” and not endocr ine progenitor/precursor cells since these cells are developmentally committed to becoming mature beta cells in vivo. TABLE 23 Comparison of endocr ine (CHGA+) and non-endocr ine (CHGA-) cells from stage 7 and stage 4 (PEC) Stage 7 endocr ine/non-Endocr ine flow cytometry Total Endocr ine (ChroA+) ChroA+ Nkx6.1-INS- ChroA+ NKx6.1-INS+ ChroA+ NKx6.1+ INS- ChroA+ Nkx6.1+ INS+ Total ChroA- ChroA-Nkx6.1+ INS- ChroA-Nkx6.1+ INS+ ChroA-Nkx6.1-INS+ ChroA-Nkx6.1-INS- 85.2 17.5 9.02 20.9 37.8 14.8 8.65 0.29 0.25 5.6 Total Nkx6.1+ ChroA-Nkx6.1+ INS- ChroA-Nkx6.1+ INS+ ChroA+ NKx6.1+ INS- ChroA+ Nkx6.1+ INS+ Total Nkx6.1- ChroA+ Nkx6.1-INS- ChroA-Nkx6.1-INS+ ChroA-Nkx6.1-INS- ChroA+ NKx6.1-INS+ 67.6 8.65 0.29 20.9 37.8 32.3 17.5 0.25 5.6 9.02 Total INS+ ChroA-Nkx6.1-INS+ ChroA-Nkx6.1+ INS+ ChroA+ NKx6.1-INS+ ChroA+ Nkx6.1+ INS+ Total INS- ChroA+ Nkx6.1-INS- ChroA+ NKx6.1+ INS- ChroA-Nkx6.1+ INS- ChroA-Nkx6.1-INS- 47.3 0.25 0.29 9.02 37.8 52.6 17.5 20.9 8.65 5.6 Stage 4 endocr ine/non-end ocr ine flow cytometry Total Endocr ine (ChroA+) ChroA+ Nkx6.1-INS- ChroA+ NKx6.1-INS+ ChroA+ NKx6.1+ INS- ChroA+ Nkx6.1+ INS+ Total ChroA- ChroA-Nkx6.1+ INS- ChroA-Nkx6.1+ INS+ ChroA-Nkx6.1-INS+ ChroA-Nkx6.1-INS- 39.4 15.0 7.74 8.78 7.85 60.1 0.57 46.1 0.23 13.2 Total Nkx6.1+ ChroA-Nkx6.1+ INS+ ChroA-Nkx6.1+ INS- ChroA+ NKx6.1+ INS- ChroA+ Nkx6.1+ INS+ Total Nkx6.1- ChroA+ Nkx6.1-INS- ChroA-Nkx6.1-INS+ ChroA-Nkx6.1-INS- ChroA+ NKx6.1-INS+ 63.3 0.57 46.1 8.78 7.85 36.2 15.0 0.23 13.2 7.74 Total INS+ ChroA-Nkx6.1-INS+ ChroA-Nkx6.1+ INS+ ChroA+ NKx6.1-INS+ ChroA+ Nkx6.1+ INS+ Total INS- ChroA+ Nkx6.1-INS- ChroA+ NKx6.1+ INS- ChroA-Nkx6.1+ INS- ChroA-Nkx6.1-INS- 16.4 0.23 0.57 7.74 7.85 83.1 15.0 8.78 46.1 13.2 Example 24 Purification of Endocr ine Cells From Stage 7 Cell Populations Using A Zinc Sensor Beta cell secretory vesicles contain high concentrations of Zinc. Zinc is accumulated in the vesicles by at least the action of the Zinc transporter SLC30A8 (or ZnT8). Zinc-binding fluorescent probes have been used to visualize Zinc in cells by absorbing and emitting more light at specific wavelengths when bound with Zinc than without Zinc. Yet, reports show that most of the probes may not localize properly to detect Zinc within beta cell vesicles. Interestingly, PyDPy1 (or Py1; Chemical Communications, 2011, 47:7107-9) can get into vesicles, but has not been used with beta cells to date. Py1 increases fluorescence intensity 50-80x when bound to Zn2+. For the first time, Applicants have demonstrated production of an endocr ine cell, or immature beta cell, population capable of in vivo function. It is therefore advantageous to have a means for further enriching this population for purposes of further in vitro analysis, for use as a screening tool, or to provide an enriched immature beta cell population for transplant. Stage 7 endocr ine cells, or immature beta cells, were treated with Py1, and the cells were sorted via fluorescence for those containing high Zinc content. A Py1 zinc sensor (custom synthesized by ChemoGenics Biopharma, Research Triangle Park, NC) was resuspended at 10 mM in DMSO and diluted in DPBS(-/-)/ 0.25% BSA (Buffer A) to 5 micromolar final concentration (staining solution). Differentiated stage 7 cell aggregates were washed twice with DPBS(-/-) and dissociated with Accu max. The Accu max was quenched with addition of base medium containing B-27. The resulting cell suspension was filtered through a 40 micron mesh, centrifuged, washed in Buffer A and resuspended in staining solution. Staining was continued for 15 minutes, cells were centrifuged, washed in Buffer A and resuspended in Buffer A for sorting. Cells were sorted by flow cytometry into CMRL/50% FBS at 4° C. Following the sort, cells were centrifuged and resuspended in RNA isolation buffer. In the results shown in FIG. 40 , the cells encompassed or gated by the polygon are live cells that have increased fluorescence due to the presence of Py1 sensor bound to Zinc. This cell population was further divided into two approximately equal gates and sorted into two tubes, designated Bright and Dim according to their fluorescence intensity. RNA was prepared from the cells and subjected to Nanostring analysis. As seen in FIGS. 42A-C, the dim population is enriched for markers of the beta cell lineage such as INS, IAPP, PDX1, NKX6.1, PAX4, PCSK1, G6PC2, GCK and SLC30A8. This indicates that beta cells or immature beta cells can be purified from a stage 7 cell population using a Zinc sensor. A previous experiment determined that 630,000 INS mRNA units by Nanostring correspond to about 49% INS-positive cells by flow cytometry (data not shown). Thus a purified Dim population having an INS Nanostring value of 1,200,000, would correspond to about 93% INS-positive cells (1,200,000/630,000 x 49%). Conversely, the bright sorted population was enriched for markers of the alpha cell lineage such as GCG, ARX and SLC30A8 (FIGS. 41, 42, 43 ). And similar to beta cells, glucagon cells are also known to take up Zinc, So, Py1 sensor can be used to bind Zinc in both beta and glucagon cells, but separated or sorted by their different levels of fluorescence. Summary of Methods for Making PEC (Stages 1-4) Endocr ine Cell (Stages 1-7) Cultures In summary, the inventions described herein are directed to at least PEC and immature endocr ine cells and methods for making such cells comprising at least stages 1-4 for production of PEC, and stages 1-7 for production of endocr ine cells. FIGS. 43, 44 and 45 are diagrams summarizing certain aspects of Applicant’s cell compositions and methods of production described herein. Applicants have previously reported that endocr ine (CHGA+) sub-populations following a stage 4 differentiation protocol do not develop into mature and functioning pancreatic islet cells when transplanted in vivo. Refer to the “EN” cell type after stage 4 in FIGS. 43 and 44 . These endocr ine (CHGA+) sub-populations had early or pre-mature NGN3 expression (NGN3 expression before PDX1 and NKX6.1 co-expression). See also Rukstalis et al. (2009), supra. In contrast, the non-endocr ine (CHGA-) sub-populations of PEC that did not express NGN3 develop and mature into endocr ine cells in vivo. This delayed NGN3 expression until after in vivo maturation of non-endocr ine (CHGA-) sub-populations of PEC was shown in Example 10 (FIGS. 15-16 ) where the combination of Activin, Wnt and Heregulin effectively repressed NGN3 as compared to the control, and transplantation of this PEC as compared to the control gave rise to improve in vivo function. See FIGS. 15-16 . Applicant’s then endeavored to obtain a properly specified endocr ine cell culture that was capable of developing and maturing to pancreatic islets in vivo and make insulin in response to blood glucose levels similar to that observed in all the instances with PEC, specifically the non-endocr ine (CHGA-) sub-population of PEC. To this end, Applicant’s performed many iterative experiments to suppress or inhibit NGN3 expression at stages 3 and 4. Examples 8-11 describe in detail Applicant’s use of Activin alone or in combination with other agents such as Wnt and Heregulin, and at various concentrations, to affect NGN3 expression or suppression. FIG. 44 also summarizes the effect of activin at stages 3 and 4. Once it was demonstrated that NGN3 expression could be delayed and such did not affect the in vivo function, Applicant’s explored methods for inducing expression of endocr ine markers in stages after PEC (stage 4) formation. Examples 11-14 and 16-22 describe the many iterations and methods employed to optimize culture conditions to produce properly specified endocr ine populations that could give rise to in vivo function. Specifically, Applicant used gamma secretase inhibitor at stage 5 to induce NGN3 expression and endocr ine differentiation. Applicant used reagents such as CMRL at stages 6 and 7 to increase endocr ine marker expression; BMP to increase INS and PDX1. FIGS. 44 and 45 summarize and describe these efforts. It will be appreciated that initially the use of multiple methodologies are required to characterize and identify cells (e.g. Q-PCR, ICC, flow cytometry analysis, C-peptide assays and the like). After having fully characterized and identified such cells under certain cell differentiation culture conditions, Q-PCR and Nanostring multiplex RNA were often used as sole methods to analyze whether such a cell type was obtained. The methods, compositions, and devices described herein are presently representative of preferred embodiments and are exemplary and are not intended as limitations on the scope of the invention. Changes therein and other uses will occur to those skilled in the art which are encompassed within the spirit of the invention and are defined by the scope of the disclosure. Accordingly, it will be apparent to one skilled in the art that varying substitutions and modifications may be made to the invention disclosed herein without departing from the scope and spirit of the invention. As used in the claims below and throughout this disclosure, by the phrase “consisting essentially of” is meant including any elements listed after the phrase, and limited to other elements that do not interfere with or contribute to the activity or action specified in the disclosure for the listed elements. Thus, the phrase “consisting essentially of” indicates that the listed elements are required or mandatory, but that other elements are optional and may or may not be present depending upon whether or not they affect the activity or action of the listed elements. Also, it will be appreciated that in embodiments where numerical values, such as amounts, concentrations, percentages, proportions or ranges, are recited the value that is referred to can be “at least about” the numerical value, “about” the numerical value or “at least” the numerical value. 1. A method of producing mature beta cells in a subject, comprising transplanting into the subject an in vitro cell culture comprising human cells, wherein at least 10% of the human cells are unipotent immature beta cells that co-express INS and NKX6.1 and which do not substantially express NGN3, thereby producing mature beta cells in the subject. 2. The method of claim 1, wherein at least 30% of the human cells are unipotent immature beta cells that co-express INS and NKX6.1 and which do not substantially express NGN3. 3. The method of claim 1, wherein the cell culture is in a semipermeable encapsulation device. 4. The method of claim 1, wherein the cell culture is in a perforated encapsulation device. 5. The method of claim 1, wherein the unipotent immature beta cells are in cell aggregates. 6. The method of claim 5, wherein the cell aggregates are 50 to 600 microns in size. 7. The method of claim 5, wherein the cell aggregates are 50 to 250 microns in size. 8. The method of claim 1, wherein the unipotent immature beta cells further express pancreatic duodenal homeobox factor-1 (PDX-1). 9. The method of claim 1, wherein the unipotent immature beta cells further express MAFB. 10. The method of claim 1, wherein the unipotent immature beta cells are capable of maturing into mature beta cells. 11. The method of claim 1, wherein the cell culture further comprises Chromogranin A (CHGA)-negative cells. 12. The method of claim 1, wherein the subject has diabetes. 13. A method of producing insulin in a subject, comprising transplanting into the subject an in vitro cell culture comprising human cells, wherein at least 10% of the human cells are unipotent immature beta cells that co-express INS and NKX6.1 and which do not substantially express NGN3, thereby producing insulin in the subject. 14. The method of claim 13, wherein at least 30% of the human cells are unipotent immature beta cells that co-express INS and NKX6.1 and which do not substantially express NGN3. 15. The method of claim 13, wherein the cell culture is in a semipermeable encapsulation device. 16. The method of claim 13, wherein the cell culture is in a perforated encapsulation device. 17. The method of claim 13, wherein the unipotent immature beta cells are in cell aggregates. 18. The method of claim 17, wherein a) the cell aggregates are 50 to 600 microns in size or b) the cell aggregates are 50 to 250 microns in size. 19. The method of claim 13, wherein a) the unipotent immature beta cells further express pancreatic duodenal homeobox factor-1 (PDX-1) and/or b) the unipotent immature beta cells further express MAFB. 20. The method of claim 13, wherein the subject has diabetes.
import { Component, OnInit, ViewChild } from '@angular/core'; import { routerTransition } from '../../router.animations'; import { UserService, AlertService, FaceService } from '../../shared'; import { NgbModal } from '@ng-bootstrap/ng-bootstrap'; import { Router } from '@angular/router'; import { FeedbackService } from './feedback.service'; @Component({ selector: 'app-feedback', templateUrl: './feedback.component.html', styleUrls: ['./feedback.component.scss'], animations: [routerTransition()] }) export class FeedbackComponent implements OnInit { feedback: any = {}; loading = false; faceImagePath = null; faceImageSrc = null; rate = 0; @ViewChild("fileInput") fileInput; constructor(private modalService: NgbModal, private alertService: AlertService, private router : Router, private faceService: FaceService, private feedbackService: FeedbackService ) {} ngOnInit() { } update(content) { this.loading = true; if(this.faceImageSrc){ const formData = new FormData(); formData.append("file", this.faceImageSrc); formData.append("comment", this.feedback.comment); this.feedbackService.add(formData).subscribe( res => { this.rate = res.message; this.loading = false; this.feedback = {}; this.faceImagePath = null; this.faceImageSrc = null; this.modalService.open(content); }, error => { this.alertService.error(error.json().message); this.loading = false; }); }else{ this.alertService.error('Face required'); this.loading = false; } } readImage() { var that = this; let fileBrowser = this.fileInput.nativeElement; if (fileBrowser.files && fileBrowser.files[0]) { this.faceService.rotateImageO(fileBrowser.files[0]).subscribe( data => { this.faceService.resizeImageO(data, 500).subscribe( data => { var reader = new FileReader(); reader.onload = function(fileReader){ var reader = fileReader.target; that.faceImagePath = reader; } reader.readAsDataURL(data); that.faceImageSrc = data; }, error => { this.loading = false; this.alertService.error(error.message); }); }, error => { this.loading = false; this.alertService.error(error.message); }); } } }
Card Tips:Quill Pen of Gulldos From Yugipedia Jump to: navigation, search • Note, that returning WIND monsters from the Graveyard into the Deck is not a Cost, so you can return WIND Synchro and/or Xyz monsters to use again. • This card works well with "Reborn Tengu". If you use Quill Pen to target 2 copies of Reborn Tengu in your graveyard and one copy of Reborn Tengu on the field, one of the copies of Tengu returned to the deck will be special summoned by the copy that returned to your hand. • This card can replace "Contact with Gusto" in a Gusto deck. Very useful for dealing with troublesome monsters such as "Wind-Up Zenmaines" or "Stardust Dragon". • This card can work well in a "Ritual Beast" Deck. • Use this card mainly to recover any lost "Ritual Beast Ulti" Fusion Monster. • Since the Limitation on "Ritual Beast Ulti-Cannahawk", any play sequences with "Spiritual Beast Rampengu" becomes even riskier than before, thus also means less chances of any "Ritual Beast Ulti" monster hitting your Graveyard. • This card works best as backup plan when combined with "Constellar Ptolemy M7", which normally always Summoned using 2 Level 6 "Ritual Beast Ulti" monsters. After "M7" spent its materials (one of which can be used to recover "Ulti-Cannahawk"), you can use this card to recover the spent materials and either return an opponent's card, or you can return "M7" to use it again at a later time.
Thread:Fructosejerv0412/@comment-22439-20150626070424 Hi, welcome to ! Thanks for your edit to the Sheep page. Please leave me a message if I can help with anything!
Hi, my name is Andrew Hall, and I'm a Program Manager Lead in the Visual Studio team working on .NET Azure Tooling. Hey there, I'm Paul Yiknovich. I'm another Program Manager Lead working on Azure Development. What's a little known fact about us, Paul? Yeah, what people really don't know is that we're actually twin brothers, and that makes us totally interchangeable. It is true. We were just separated at birth, so different last names, but yeah, one for one swap, except Paul knows twice as much as I do, but I'm better looking. Well, you can tell by the height we're brothers, right? That's exactly true. That's right. We're sitting down though, so maybe. Oh, that's true. So today, we're going to talk about five Azure services every .NET developer needs to know. The goal of what we want you to come away with from this session today is, if you're looking to get started with Azure, we want to make sure that you get a really crystal clear view on where you should look to start with Azure. I think the five services that we've picked, we think are going to be the things that probably 80 to 90 percent of apps when they first try to go to Azure are going to need. Yeah, and we looked at a few different ways. We looked at how do you modernize the infrastructure that your app is using? How do you modernize the actual application to make it work just totally optimized for the Cloud, and then even how do you modernize your DevOps? So you're going to see that theme throughout. It's a perfect transition. Okay. So our agenda for today. So we're going to actually start off a little bit with some Azure terminology. We're going to make sure that we're not throwing around terms accidentally that we haven't defined. Every technology has done specific terminology, and you know what or you don't. So we make sure that that's clearly defined. As you mentioned, we're going to talk about hosting applications in Azure. So you have your code that was running either locally somewhere or some other provider. How do you actually create something in Azure where that can run, and you can hit it and you have a public endpoint to do that. As you mentioned, then we want to talk about how do we modernize it to take advantage of more of the power of the Cloud. Then as you said, introduce how do we start to onboard to a modern DevOps process. So with that, let's flip through some terminology real quickly. So one of the fundamental things you always have to deal with in Azure is an Azure account. So basically this is the credentials that you sign into Azure with. So if you were to go to the Azure portal or when you're in Visual Studio trying to work with Azure, you're going to have to be logged in with your Azure account. So I could use like my school account, my work account, my MSA, Microsoft account. That is correct. Hotmail, that kind of thing. Then under the account that kind of fits into the hotel, you have what's called subscriptions. So subscriptions is the billing plan that Azure resources are created inside. So this is when you create Azure resources, basically where is the money coming from if it's a paid resource. There's a lot of free resources which are worth mentioning. For a quick call out, if you have a Visual Studio subscription, you get Azure credits every month. So if you pick your Visual Studio subscription, for example, when I say unit of billing, it'll just use your monthly credits which reset every month. If your credits run out, you're not going to go bankrupt, I promise, Paul. Azure will just turn everything off. At least not for this. Not for this. Okay. Azure will just turn everything off, and you won't ever get a bill. You just can't actually access the resource until your credits. Yeah, I get that question a lot too about the like, if I'm using a free trial and I hit the limit of free, what happens? Yes. Yes. Great question. So if you're using a free trial and you hit the limit of free, Azure will turn everything off, and unless you have explicitly opted in, and it's an opt-in, the default is everything turns off and you can't use the resources, you could go in and you could say, okay, let me go over my free credits, and here's a credit card that goes with that. So it's like a spending cap and you could lift the cap but that's up to you. That is correct. So it's safe by default. It is safe by default. You will never, without choosing to say, you can bill me more than my credits allow, ever end up in a position where you get a really large bill you didn't expect. So free is truly free? Free is truly free. Okay. And even if you're trading against credits, it's only credits, you're never dealing with real money. Okay. So next thing, we talked about creating resources inside accounts. So we should talk about resource, and a resource is maybe obviously, but any service that you would create inside Azure, anything that's going to actually take any CPU or memory or storage or anything like that, but I want to be able to take it in. Kind of like an instance of a service. Instance of a service is a great way to think about it. So resource groups are one of the most fundamental primitives in Azure that you're going to deal with and it groups all of your resources that you create, and all the resources that you create actually have to be placed into a resource group. The analogy I use for resource groups is like creating stuff on your local machine, that you always put it in a folder. A resource group is basically a folder. Folder or even like a namespace almost, right? Yeah. Namespace would be another good analogy there. Yeah. So anytime you- I'm in my OO kind of a mood today, so I guess I'll just- Perfect. Yeah, I'll take it. Anytime you create a thing, you always ask you and the general guidance for resource groups is you want to organize things that are going to have the same lifecycle together in the same resource group. It's almost like if I have a unit of deployment, and these are things that should deploy together, that's a good idea for resource group. That's correct. Okay. Or delete even. Exactly. Where it becomes really convenient is if I'm doing some testing, I want to get started, I think we'll look at that here in a little bit, and I say, oh, I want to try this application up in Azure, and I'm going to- It's going to have an ASP to an application that has a SQL database associated with it and has some storage account associated with it or anything else. If I put that all in the same resource group and I go, hey, yeah, I got it working, but I don't need this anymore, I just delete the resource group and it gets rid of all of those resources for me, just like deleting the folder, right? If you have some temporary stuff on your disk as opposed to having to go through and potentially clean up and delete four files. So for DevTest, you can just create each of these namespaces but they're called resource groups and then at the end you can purge them. Exactly. Okay. And then it stops trading at your credit. Yeah, I'll be showing it a bit later. Like once you use Azure for a while, you can really rack up a lot of resources. I think that's going to be handy. Absolutely. And the last term that you may hear us use sometimes is provisioning. So in Azure, provisioning is simply the act of creating Azure resources. Okay. But if you use any of the management APIs or anything like that, it'll be dealt with as provisioning. So I may use that term sometimes. It's extra fancy. It's extra fancy, that's right. So the next thing we want to talk about a little bit is the idea of hosting versus services. And this is really a distinction that you and I came up with. It's captured in the Azure docs but they won't necessarily talk about it exactly this way. And they're all services, right? There's like you're making a distinction between hosting services and other ones. Exactly, really hosting and software as a service. Okay. So hosting and the definition that we're going to use today is where you provide your code. And then what Azure's going to do is Azure's going to execute the code that you wrote according to the application model of that code. An example of that would be an ASP.NET application, right? It gets deployed or published into an Azure environment and Azure is going to hook up and it understands how to call into and route requests to the ASP.NET application. At the end of the day, all of the interesting stuff that's happening is a custom code that I provide it, right? If I have a web API that says some and I accidentally subtract instead of some, it's going to return that wrong number because Azure's not doing the math for me. It's whatever I wrote. But maybe another way to think about it. So if I just, if I have a web app, a web API, today I'm used to, you know, I work with my, let's say at my IT team, I get some servers with IES and I just push into that. So now in the cloud this sounds like, yep, exactly. A similar thing. It's kind of like IES is a service, right? It's exactly correct. Yeah. So then service or really software as a service from definition that we're going to use here is where you provide the data or information and Azure's implementation actually takes the action on what it's provided. So if we go back to my analogy of adding two numbers, if Azure had a software as a service, a service that was addition, I couldn't make a mistake in saying A minus B instead of A plus B because I would just say, hey Azure, here's two numbers A and B and it would do, use its own implementation and return the result to me. And a great example of this for our purposes is blob storage, right? So I basically give it streams of data. It stores it for me. I can ask it to do things. I can say list stuff for me, you know, delete, write, update. But at the end of the day, Azure has an implementation that's actually dealing with how that gets copied into a persistent storage mechanism for me and I just have an API service. So hosting, I put my app or my APIs in a hosting service. The service you have, I'm using Microsoft's APIs. That's correct. Directly and they're already in the cloud and that's high scale, I'm guessing, availability, all the good stuff we want for them. All the good stuff that we want. That is absolutely correct. That's a good way to think about it. And so with that, we've been talking for a little while. Why don't we roll over and do our, look at our first two services from a demo perspective. So to introduce what we're gonna do, I'll start here in Visual Studio. And I have a ASP.NET application that I've written that's gonna behave like a photo gallery. So it has, what it does right now is it has a SQL, local SQL DB that I can log in as a user. And so let's go ahead and do that. Anybody needs to talk to me? There you go. So we'll log in and perfect. Nope, I don't want to remember that password. And so now I'm a logged in user. So that used my SQL, local DB account that was running on my local machine. I can upload images. So if I want to pick another image, I can say, okay, so I'll upload that. But you're not putting those images in the database, right? I'm not putting the images in the database. I'm just storing them on, on disk right now. So they're going to a, so now I return to my gallery and I see the other image that I loaded. If you look over here in Visual Studio, there's a user images folder. And so that image appeared in that folder. Cool. So at this point, what I'd like to do is before I do anything else, I need to get this application off of my local machine or off of on-premise. And I need to get that running up in Azure. Okay. And so I have come two requirements right now. You don't want to put in a request for a machine and wait? I really don't. Oh, okay. I've heard that. I'm your guy though. Okay, I've heard the cloud's the future, Paul. All right. And I don't want to put in a request. I don't want to run it on-premise. In fact, my boss has told me that he'd like to stop running the server in his office. It's making it really hot in the summer. Yes. So I need two requirements here. I need to host the ASP.NET part of the application. So going back to where we're kind of talking about hosting. And then I'm going to need a database because SQL LocalDB obviously isn't going to work up in Azure. Okay, we can do that. So if we go ahead and cut over to Paul's machine, can you talk me through how I would do that in Azure? Yeah, sure. Okay, so we're going to switch to my machine. Great, we're already over. All right, so the first service that we're going to use, this is service number one of the five, is we're going to use app service web apps. Now, and we're looking at it right here. So I said create a resource in the Azure portal and we'll create a web app. I really think of a web app kind of like it's IES as a service, as a managed service. Microsoft runs the operating system. It runs IES. It handles the hardware and the infrastructure. So this is really good for web apps. It's good for web APIs. It's good for web jobs. And this is really where I start for most workloads. Now, there's plenty of good reasons when you're going to do something more intensive that goes beyond what a web app does. And we have other services for that, but this is a great place to start for that. So let's just create a web app. My initials, pie, not to be confused with Python. Be really explicit. Andrew talked before about a subscription. I have any number of subscriptions here. That's quite a lot. And we talked about this idea of a resource group. So this will be handy and I'll show you in a second. I'm going to create it with a few strings I can find like Andrew. I'm even going to just say dev test for myself. Just makes it a little easier to find it later. We'll start with Windows, but we're using .NET Core, right? So we could do Windows or Linux. We have the choice there. And there's this thing called a plan. I don't know if we, did we talk about that in the concepts? We haven't talked about plan yet. Okay, cool. So that's a nice one for me. So I think about a plan as the, like kind of like the hardware that you're going to be running. So it's literally, it's literally like a virtual machine that has a size. It has a certain amount of RAM and CPU. And this gives you the flexibility to choose a plan. Right, so I could go ahead and create one. One thing I want to be careful about is I like to create it in a data center that makes total sense. So we're in the Northwest. I'm going to pick West US two. And then you have lots of options for pricing tiers. I'm going to start with the initial one, but you could see if my demands were higher, I just pick a higher plan with better hardware. Okay, and that's really all we have to do. I like to turn on application insights that will help us down the road. And maybe a sneak peek of what one of the services could be. We'll just go ahead and create it. Now, if you like working with a UI or a GUI, the portal's a great place. It's kind of a ubiquitous place to create everything in the cloud. I think I just want to talk about, one of the tools in our tool belt is the Azure CLI, or command line interface. And really everything you can see in the management interface of the portal, you can also do at the command line. So what that would enable me to do is it sounds like that would enable me to check in scripts that would let me have repeatable deployments. For example, if I wanted automate it to stand up my environment for testing purposes or whatever on a just a scriptable, repeatable way, I could assume I could write scripts that use the Azure CLI. But I could check in with my source code. Right, like if you're, whatever scripting language you're comfortable with, whether it's PowerShell or Batch or even Bash, you can go ahead and script this. And it's totally useful just like you said for automation, for repeatable actions. It'll just take the guesswork and the variability out of things. Another thing I find is, once you know what you want to do, the CLI is very definitive and it's very fast. And you can reduce a lot of clicks in a user experience using the CLI with a one-liner. So here you can see, we did a few clicks, right? But to replicate the same thing, AZ, web app, create some of the parameters for the resource group, the name and the plan. It's gonna happen. Make it so. Great. Couple of questions for you real quick, Paul. Sure. So I'll just mention, you showed the portal and I think it'd be awesome if you could tour us through kind of the various capabilities you mentioned that app service is a great place to start. App service web apps is the place to start for hosting web applications. Yeah, that's good. So what makes it such a great place to start? Obviously, it was quick to create. One of the things we'll show later is you showed the portal, you can create and everything from Visual Studio with app service, so web apps, which is a great thing. We wouldn't have to go through the portal just to get a quick web test environment and I'll show that in a little bit. But let's say I actually want to use it for production workloads. Does it scale? What other capabilities does it give me? Or is it really just a place for me to dev test and then I have to find somewhere else to host production? Well, let me answer that right away. Yes, it's absolutely a place to scale. We even run a lot of our mission critical apps on app service. So we need a front end or an API, it's great for that. So let's take a look at some of the things you can do. First thing is, if I can type, I mentioned that having these resource groups make it really easy to just search. So I love this kind of text box in the middle of the portal. You just start typing in and you can find things. So right at, because we were careful about the resource group, we're gonna find this PY Andrew that we made. What I'm looking at now, this is the app service plan and it lives inside of this resource group, right? So I can see now the resource group has my app service. It has my plan or basically the host server hardware that I'm running on and even has app insights. So it's keeping it all organized. And again, one of the things about getting to prod is you probably have a few different environments, whether it's like Dev's test, integration testing, maybe a staging prod. So you can organize the different resources and you can actually even make sure that there's different access control or RBAC for each one. You'll really wanna think about that. Like prod, probably don't want the testers, testing against prod. RBAC, resource-based access control, I assume. Mm-hmm, yes, or roles-based, either one. Okay, so let's go back to my app service and just take a, we'll take kind of a lap around it. So we saw all the same information before, kind of a subscription. There's a really easy way to diagnose and solve problems or to get insights over the app. But some of the things I wanna get to are kind of the cool things that make this more Pazzy, right, so. Quick question right before you click back, sorry. If you go back to your overview page, I noticed that there was a URL. So I just wanna make sure I understand, because I think what we just happened is in less than five minutes, and really it was probably less than one minute, you instead of, as you kind of talked about submitting a ticket and getting something created on a local network on my own sort of company, right, or going and bothering my boss or whoever controls the hardware, you stood up a full hosting environment that has a public endpoint in about a minute to two minutes. Right, that's a really good catch. So it's a fully networked, secure hosting environment with the URL on the internet. Pretty cool, like you got that right away and then you can even redeploy over and over the content to that hosting environment and that networking will be durable. Like that host name, the networking to it will just kind of stick around. So think of your app service as this configurable host in the cloud that can kind of stick around and you can keep dynamically pushing new content to it. Very awesome. Yeah. Okay, sorry, I didn't mean to interrupt. Yeah, okay, perfect. So let's talk about where this goes beyond, let's just say ordinary IS on a VM, right? So the first thing is there's a first class idea of application settings in an app service. I really like this because it kind of lets you configure things on the fly. You can do it dynamically, change it anytime. So let's say the framework version that I'm expecting, the Bitness 32 or 64 version of HTTP. And then all the way down to things like, I think of them as environment variables. The term here is app settings, but I can dynamically set configuration values, key value pairs and they will live in the environment. And so you talk about getting to prod, a pattern I would do is I would say, here's my database connection string. So let's go ahead and add it. And we won't pick a real one right now, but kind of like this, right? And I am going to write my app in a durable way against DB connection, but I know in each environment I can actually override the value, kind of like change that environment variable so that the right database connection exists and is loaded in that environment. It's also very secure, right? Because only people who have access to this environment can even get at these settings. Right, so this is a much more secure way than trying to have something in my build server that tries to inject it into my web.config, right? It is, and sadly, when we think about security, we even have to think about all the internal people who have access to things. So it's just kind of out of sight, out of mind. It keeps things really isolated to where they belong and that's probably a good thing. Some other things, you mentioned prod, so I just want to mention, I'm not going to hook it up now, but we can set up SSL and TLS. So use basically HTTPS for your website. So that's really cool. Another thing I want to talk about is scale, right? Like one of the promises of the cloud is really just kind of near infinite scale. So to show that, again, if I wanted to provision 10 machines, because I know that we're coming into a holiday, it's going to be a really important time for our site, I'd have to request it. But just with Azure, it's as easy as sliding the slider bar. Should we make Azure work a little bit? I'm not paying for this. Let's go ahead. Looks like you also have autoscale, right? So you, instead of me being responsible for coming in and changing the slider, if things get popular in Europe while I'm asleep, autoscale, Azure will detect that until they just spin up automatically on my behalf. Right, it sounds like you know a lot about that. So what would you want to scale based on, do you think? I generally request. The request volume that's coming in, if I was seeing a major dip in response rate. Very cool. Yeah, so it's kind of like you can be metrics based, think about a threshold. So one would be, you know, the requests are going hot, like more requests than you think the machine can handle through testing. Another one might be like CPU or memory. You can really just use that as a trigger to scale out. And it's not just about scaling out, it's also about scaling back, which is really cool and that avoids the waste. So right now, by the way, servers are just being provisioned and that'll be pretty cool. Can you remind me to delete those after? I will do that. Okay, excellent. And I think for getting to cloud, for getting to prod, the main things I wanted to show you were, you know, you can do SSL, you can scale out, you can autoscale, and you can set environment variables. Makes sense. So next question, if we think about our app, so we figured out we're going to host it, that's going to be app service. Yep. The next thing I want to know is, so remember I have a SQL database that my application has to work. Yeah, totally. Using SQL local DB. Do I have to go create a VM and go install SQL server on it? You don't have to. I mean, and by the way, that's a perfectly fine option. Like if that's the first step you want to take, we have VMs that are preconfigured with SQL and you can pay like a little bit more than a normal VM to use SQL. But we want to go all in on the cloud here, right? So like, why do that? So we have this thing called Azure SQL DB and that would be the next thing we'd check out. So let's go to, that's right here. So we have the SQL database. And then the really cool thing here is I can create both a database and even a database server. If I wanted to prime it for more DevTest, we could throw in AdventureWorks or our own backup. And then a lot like app service plan, there are plans for databases. And here you kind of pay by the DTU. That's kind of like a mathematical scale unit. But there's some good guidance here to kind of understand how many DTUs you might need. So I would just check that out. When you say DTU, what does that mean? Well, let's click the what is DTU. I knew you'd ask. Yeah, so yeah, I want to say it's a transaction unit. Yeah, am I right? Yes, data transaction unit. But for me, it's more complicated to explain because it's an aggregate, right? It's an index unit over a number of other units, right? Like we're seeing things like IO, IOPS, CPU, uptime, retention and things like that. So basically you pay a bit more, you get to do more transactions. But the cool thing about that as well though is instead of paying a fixed fee to your point about wasting money with having resources up that we don't need and we don't need them, same thing from a database perspective, right? By using the software as a service Azure SQL DB, only actually paying for what I'm using. So if I'm trying to get started, you know, it's my own dev test thing or it's even a line of business application that's relatively small, I'm not gonna pay much because I'm not using, doing many transactions or storing a lot of data in that. So I don't have to pay $15 a month to reserve a dedicated database. You can even see this right here, a scalable database in the cloud that gets patched for you by Microsoft, $15 USD. Right, just to get started with that. So pretty cool. So that's what, so just to kind of wrap this all up. So we created, you know, you described a data-centric web app, right? So we've got the web app that can work in DevTest or Prod, we've got databases that can work in DevTest or Prod and basically we've just modernized our infrastructure and even the way we pay for things just for those two steps. Very cool. So one question and then we can go ahead and move on. So a question was you created resources in West US and what if you want to change the geographical area that something's located? Is that possible? Yes, it is possible. So we have commands and operations where you can move from one resource group to another. And you also talked about the plans here in SQL and app service, right? I pick a specific plan without blowing away my data and recreating it, do I have the ability to upgrade that plan? Yes, absolutely. So these sliders are basically re-entrant and maybe I'll give you a visual example. So let's go back to PY app for Andrew. Okay, we're gonna save our database right now. Yeah, we didn't go ahead and do that right now. So here, just to give you a quick example. So scale out the way I think about that. It's going from let's say one machine or one instance to many. Scale up is about improving the hardware and the skew and the plan that you're actually using. So I would go to scale up in this case and you can see that I can actually dynamically change the hardware that I'm using. And Azure is pretty smart about this. It'll like spin up the new hardware, it'll start draining the traffic from the current instance and kind of redirecting it to the new one. So I can start cheap if I need. And then without losing any data, without recreating or redeploying anything, as my site gets popular or perhaps I move it from a staging to production or preview to RTW status and I see traffic uptick. It's just a couple of clicks and a couple of clicks, scale it out, scale it out. Or you could do a command line parameter. In the end, you're going to get a lot more economy that way. But you can also scale up and just increase your hardware. But my first instinct would always be to scale out. Perfect. Makes sense. That would be a great transition for our next status. And so just a quick review. So the first thing we talked about is Azure App Service. And we think that is the place that you should look at to start when you're hosting applications in Azure. Right? As my requirements get more advanced, I may look at some of the other hosting options available in Azure. But this would be the place that we'd recommend starting until it doesn't meet your needs. So for example, if I need a hyperscale distributed app with microservices, I might look at service fabric. Yeah, exactly. That, right? And so I think the other thing that's really worth mentioning is we didn't talk about it specifically. It's a fully managed service, which means that Microsoft or Azure takes care of automatically applying all the framework and operating system updates. So I'm not going to be in a place where I'm vulnerable to the next ransomware attack that got patched six months ago. Too soon. But I'm sorry. Because they're going to be updating that on a regular basis. Yeah, absolutely. I go back to when it comes to the hardware, the operating system, and even your middleware like IES, Microsoft is on the hook for that. So really, as a developer, we're on the hook for applications when we use a service like this. Yeah, and so I think the one thing we didn't talk about, which we'll show here in a few minutes, is the concept of deployment slots. It's another really cool capability. We'll get to that in a little bit. But and then we'll also show the first class integration with Visual Studio. App service is just brain dead symbol to get started with from Visual Studio. It is. In fact, even a lot of the stuff I did, you're going to show me up with how the tooling makes it easy. So everything you just showed at the portal, I can actually do from Visual Studio without ever having to log into my web browser. But it's good. I figured the VS guy would do that. Any good infomercial, you can't open the milk container until I have the screw on straw, right? That's right. You have to struggle a little bit. You need to drop your laptop first and then. And then we talked about SQL database. So a lot of the same concepts that come in with Azure App Service, right? Most of my ASP and applications that I work on personally all have a SQL database that back them. And so Azure SQL database is a great software as a service place that just makes it really quick. It'll create the database again. Speaking of dwelling on the page, have you ever created a database cluster? Because you haven't lived until you've created a database farm or cluster. You know, I have to say I haven't. OK. We have to save that. Leaving that to Microsoft, I'm happy about that. For a more advanced scenario. So with that, let's go ahead because one of the things we talked about is a designer app for scale and modernization. So let's go ahead back to our application and talk about adding some more capabilities. My mouse aside, mouse, there we go. And so we go ahead to Visual Studio. So one of the things that we mentioned a little bit ago is that when this was running in an on-premises server, I was storing all of the images that users uploaded into local file storage. And that works when we publish it to Azure. So if I have my site, I can publish it, and it'll work. But the problem with that is that data is not the storage for local file storage isn't persistent, isn't guaranteed to be there. And so any images that are uploaded that I store on that local app service VM, if upgrades occur or whatever and I get a new VM, none of the stuff that I store that got dynamically created at runtime is going to guarantee to live. And so to really modernize this application and take advantage of the cloud, what I need to do is I'm going to move over to storing that stuff in Azure Storage. Because that's going to be the persistent storage model for storing large objects, so blob storage in Azure. Right, it's durable, right? So it'll always be around, I don't have to worry about getting wiped out when I redeploy or like delete a server even. And one of the things that I do a lot of times, even as you change files and things like that, is one of the options when you publish is to delete any files that are inconsistent on the server if you want to clean stuff up. And if you have that option set, like I do in a lot of my projects. I do too, and I debate this with people, but. Yeah, exactly. Well, that's right, you remove a CSS file, I don't need that up there anymore. And so the fastest way is that, but it'll blow away any of that data. If I put it in storage, which would be the correct model to do that, then I'm not going to run into any issues. So I'm going to get started on my local machine here with Azure Storage. And I don't even have to use Azure, no subscription required. The Azure Storage emulator installs by default with both the web and Azure workloads. Yeah, that might be the Explorer, which is it. That is a Explorer, you're right. But thank you by the way for pumping that. Yeah. I know the people who work on that. Storage emulator is what I meant to do. So I'm just going to go ahead and manually start that, and I'll see it'll fire up here and perfect. It started and it's running, which means that my application can connect to it locally. And so what I want to do is I want to update my application, and I already have the code written here under services. So I have an Azure Storage service. Oh, yep, it's running. Thank you for telling me that. And so what I'm going to do is my application starts up here, is it's going to actually store the information in Blob Storage. So there's lots of good tutorials online. I don't necessarily need to show how to write the code. This is creating all the connection information for storage, getting the, grabbing the connection string, which in my case, because I'm working with the storage emulator, it's just an app settings at JSON, right. No idea what that was. Yeah, there's no like big secret there, right? That for some reason, I don't know why that wants to open, but are you in run mode? No, I'm not in run mode, but we'll just open it as a, sure. So yeah, so you can see my connection string is as simple as use development storage. So that's like the built-in connection string for local. That's the built-in connection string for local. Okay, cool. And then we just have to take note and remember to go back to those app settings correct. Don't let me forget that. I won't, I promise. And the last thing I have to do here is I just have to, so in my startup code, because I use an interface, so both of my file provider and my storage provider implemented an iStorage service interface. I'm just gonna swap that to not use the storage connection rather than the local file connection. Okay. And so when we go ahead and start up this app. It's almost like you made a provider, like one for files, one for storage. That's correct. That's not the same, but you're re-implementing it. That's cool. That's correct. So now we've, before we did a little bit more of kind of like a lift and shift. You know, just changing our infrastructure. Now you're really modernizing the app at this point. I'm modernizing the app at this point, that's correct. Okay. So what I wanna do, log in. I should have had it remember me, but I won't make that mistake again. Now it can remember me. Minute to learn lifetime to master these things. That's right. All right, so what I'm gonna do is, now I'm gonna go ahead and say I'm gonna upload an image. And let's just pick a different image. This one will work. That's pretty sweet. Yeah. Did you take those? I did not. They come from our official Microsoft image samples that were allowed to use for things like this. Oh, that's good. So perfect, protect the innocent. So it uploaded and you mentioned storage explorer. So it's worth showing that at this point. Cause what I wanna do is I wanna make sure that this image actually uploaded correctly into storage. And so storage. It's almost like observability tools for storage. Correct. So from the first step in my development process is I think I wrote the code that put something in Blob Storage. Yep. Let's make sure that indeed it actually put it into Blob Storage. And so I will mention that Visual Studio has some basic tools built in for working with storage. But it's a storage explorer is a free download and it's just a lot more powerful, a lot more UI. Yep. And so. Works on Mac and Linux too even. Yep. It's not showing me my local one. Do you know why that is? So what I'm doing here, click this little thing. So we mentioned various subscriptions. And so I'm going to, it'll load my subscriptions and I can pick the subscriptions that I wanna see here while I'm in storage explorer. Apparently you have a lot of them. I do have a lot of them. It's the hazards of working on Azure here. Oh, even at mine. I do have yours. Yep. Okay. I have my Visual Studio Enterprise subscription. That's because of that RBAC. Right, so the local node should have it right there. That's your local and then it should be an operative word. All right, that's all right. So we'll go ahead to mention that Visual Studio has great tools built in. So we'll go to Cloud Explorer. So I hit Control Q, great tip in Visual Studio. Cloud Explorer. And so this is a tool built into Visual Studio. Again comes by default with Visual Studio. It's kind of a subset of stuff. Subset of stuff, but sure enough I have local here. And so I can browse in my various subscriptions. You know what? Okay. We have an issue with, I have a gut feeling. Yeah, there's an issue. Is there any weird subscription filter you have on? Not that it should affect local, but yeah. Anyways, it worked because it got uploaded. We can show it on my machine. Hazards of live demo, perfect. Yeah, your machine would be great. It did force me to sell an upgrade. So maybe that broke something with local. But it's going in, but one of the things you mentioned a little bit ago was you talked about designing the application for scale. And now that we're uploading images and now that these are gonna have a public endpoint on the cloud, I mentioned that what these were came from a pre-approved list of images in this particular case. But what I wanna make sure is I wanna make sure that nobody steals our images without us being able to trace back and say that we actually owned them. So I wanna add some logic to this to stamp a watermark into the image when it's uploaded before it gets displayed back to the user. So anybody coming and browsing the image gallery, if they go download it, it's gonna have a little watermark stamped into it. Cool. And when I think about modernizing my app and designing it for the cloud or for scalability and for the cloud, I don't actually want that to happen in that process. Because we talk about scaling up my application and somebody uploads images if this gets amazingly popular, like I expect it to do and we'll be retiring next year. Paul, with all the money we make off of our great image gallery site. Then, right, stamp, like image processing and stamping watermarks is significantly more CPU intensive operation than serving up pages. And so it'd be great to move that into a worker type pattern where that upload occurs, the watermarking or image processing occurs in a scalable way that happens independently outside my process. It also lowers the risk to the application because the only logic out of the application is reading and writing from storage. And if anything goes wrong with that process, the image won't show up, it's not gonna take my site down. It kind of, it's almost like a queue worker pattern. We've done this for decades, right? Like you queue up work and something else that could be a little beefier could actually go take care of that work, right? Exactly, okay. And so the way that I wanna do that, the perfect project type for this is Azure Function Project. And so Azure Functions are event-based serverless applications, which means that I don't have to take care of any of the provisioning or scaling or anything like that. I just write some code and Azure will take care of deciding how many instances I need. It will run it only when events occur. I'm only gonna pay for the CPU and memory resources I use per function execution. So it's just a great model for this. As I mentioned, it moves all of the kind of risky logic into a separate service. So the way I added as I'm in the new project dialogue, I clicked on the solutions at add and Azure Functions is the project type. We'll go ahead and actually switch over here in a minute. I have one in a different branch and then I can choose the various type of triggers that I want. In this case, blob trigger is exactly what I want because we're gonna push an image in and we wanna process it. I think we're just explaining triggers for a sec. Yeah, so what a trigger does is a trigger is whenever an event occurs, I mentioned they're event-driven, it's gonna execute code to process that event. And so in this case, a blob trigger is gonna say, whenever anything shows up in this blob container, call some code that can do some processing on it. And of course I define all of the code that does that processing on it. Okay, so kind of like your inbox is actually the blob container. That is correct. There's lots of other ones, Cosmos DB. I could do things when data got written into new tables or not into new tables, but got written into tables. I could, HTTP trigger gives me an HTTP endpoint if I need to call it FES. Is that almost like a webhook? It's almost like a webhook, exactly. And there's a lot of webhooks integration, there's timer triggers, there's service bus triggers, there's queue triggers for message passing. It seems like a lot of those things you call services, they have events and you can wire them up with a function. That's correct. Okay. So in this case, I mentioned I already have kind of a project created. So let's go ahead and cancel out of this and I'll just switch over to our next branch that has that. It won't let me switch branches while I have pending changes. So let's go ahead over. Probably a good thing. Yeah, probably is a good thing. All right, so let's go over here to our next branch. And so I can see that I have an Azure Functions project now available here to me. And so what I wanna do is I wanna go ahead and add that blob trigger. So we have the basic project infrastructure there, but I'm gonna right click, I'm gonna say add new Azure Function. Let's pick the name of this. This is gonna be, we're gonna process the image. So image uploaded, perfect. And so this is the list of triggers available that we just talked about it. This is gonna be a blob trigger in this particular case. And this is gonna be the name of the connection string. So Azure Functions always have this connection string called Azure WebJobStorage that points to a storage account that backs the function. I'm just gonna share that for the purposes here with my function app. So Azure WebJobStorage is the name of that connection string. It's already in my local about settings. And then the path to my thing is images. So this is the container that I wrote it into in my code. This is the blob, name of the blob container basically. Yep. So I'm gonna click okay. It's gonna add that function for me. And so now when we go into functions and we talk about one of the really cool things if you think back to this Azure Storage Service, whenever I was dealing with my storage account, I had all of this code to read and write and create the connection to blob storage and read information to and from my things. And then Azure Function, they do what's called a binding. And so I had this blob trigger attribute which was generated on my behalf. And it simply gives it the name to path to the connection string in that particular case. And then it's gonna automatically do all of that for me. It's gonna initialize the connection. It's gonna listen for changes. When something new gets written into it, it's gonna read that into a stream and it's gonna pass that to me as a user in a stream. So this my blob is whatever image has gotten put into that queue since I did something with it. And so now we're just gonna do a little bit of wire up to, so let me go ahead and just copy and paste this over a new version of the function. I'll explain what it does here in a second. So that's the input blob that we just talked about. And what I can do is I can add an arbitrary other number of binding as well. So in this case, I mentioned that I was gonna watermark the image and so I need to write back out the modified image. And so I'm gonna have an out parameter in this particular case that's gonna be an out blob stream. And so I just put the new bits or new bytes for the modified watermarked image into this output blob. And it's gonna automatically just get written into my whatever queue that I want for that. And so my application website's now gonna read them from the watermark. So on one hand, I mean, it sounds like it's doing a lot of process and machinery for you, which it is, but at the end of the day, it lets you think about more like functional programming, like there's an input, there's an output in the code that you write is what decides what gets output. That's correct. And the runtime handles all of the, I'd say, heavy lifting and connection goo on my behalf to the Azure resource. And it keeps a quality of service between you and the storage. Exactly. So now that we have this function written, let's go ahead and I wanna start both my website and my function app. So let's go down here to my properties and my solution. And we'll go ahead and set it to have multiple startup projects. We'll can launch the regular application without debugging. And we'll go ahead and launch our function app with debugging just because I wanna show the breakpoint get hit when we push something into that storage queue. So let's go ahead and hit a five. One of the really cool things about Azure Functions from a serverless platform compared to a lot of the other options available out there is the full integration with Visual Studio, full local development and debugging experience available to me as a user. And there should be something in that queue already. So I actually expect from our last upload it never got processed. So I'll expect this function to actually get hit right away before I even upload an image as a user. Because there's a pending image to process in there. Yeah, that gives you a lot of flexibility. Using an emulator, it saves cost and time basically. Exactly. It's a lot less to administer. So you can see that in this particular case my breakpoint got hit. The name of the image that I got uploaded, I'm assigning a grid to them to make sure they never get duplicated. But that's the image that I uploaded. And so you can see, so F5, right. Perfect, processed it. Let's go back here. Let's upload another image. It feels like the event driven program we've done forever, right? It's the event driven program we've done forever in the cloud. Boom, upload another image, gets hit. Perfect. But the other thing that I think we talked about a little bit before that was really, really important here was the fact that it's now running a sub-process. So I added functionality to my existing application that we wanted to move into Azure in an event driven way with very minimal risk by using a storage or queue or some other pattern as the communication mechanism in between. And now if I come back here to my gallery. Perfect, I can see all those images and I can see some watermarks that are getting stamped in here. So I have it set to stamp in the lower right hand corner and so you can see it showing up there. Sweet. So you're really, now it's really, it's more than a web app now. It's like really a cloud compute intensive app, right? Exactly. That's pretty cool. Even though it's a simple example. Exactly. Another thing just to re-emphasize there is as you mentioned, all this was on my local machine working with emulators. Yep, right. Yeah, that's cool. But you could also use the cloud resources if that's the way you use it. I could, and I'll show because we talked, I promised that I would show, put you to shame from the portal from Visual Studio. So now that we have our application working. Andrew, Andrew. I want to go ahead and let's just say I want to publish all of this up into Azure. I'd have to do each project individually. We'll pick the web app first. So I'm going to pick app service, right? I'm going to create a new app service. And then so it's going to load my subscriptions. We talked about that. If I had just one, that would have been pretty much instant. I have a lot that at Paul Yuck Corp. I can put this on your bill, right? Yep. So this is the app name that you were showing earlier. This is going to be the public URL. Yeah. I can choose my plan. So we talked about right, all that information that's available to me, my sizes. That's handy. It even summarizes what the hardware is. That's cool. Exactly. And I can even start with a free plan if I want. Yep. Which again is 100% free. I picked the region. You showed the ability to add SQL. I can create a SQL database that'll automatically be associated with my application and it'll write the connection string in my behalf into those settings that you talked about. So not only is it creating it, it's setting up the patterns in your app for DevOps. Exactly. I can create a storage account because our application now uses storage. I just want to hit a question that was on there too from Vlad. He's saying, you know, what's behind SQL DB service? Is it an instance of a SQL server? I mean, that is a good way to think about it. So there's two resources. There's the database resource and it's backed by a server resource. Now the service is much more special than an ordinary server on a VM because it's replicated and it's meant to run at scale all over the world. In fact, it's even backed by service fabric and it's stateful cache in a ring. So it's an extra fancy SQL server. But yes, at the end of the day, you would literally see a SQL server in the portal that you manage. Yeah, and the other thing that's worth mentioning because the SQL team likes to talk about this as well is that they actually, it's built from the same source code that the on-premises SQL server that you would run or install is. I mean, there are a few differences and we have good docs that talk about it. And even a lot of the differences are important to run at a cloud scale. But also keep in mind, there's something called SQL managed instance. So if you need absolute perfect fidelity with a SQL server on-prem that you want it run by Microsoft, go check that out. Yeah. Yes, just to finish summarizing, what I did here is I have myself set up to publish a new app service that will turn application insights on on my behalf that will create a SQL Azure database and will create a storage account and write those connection strings in on my behalf. And so when I click publish the create button, it would take probably about two to three minutes creating all of that stuff. And then I'd have a site that's up and running with a SQL server connection string written, storage account connection string written and application insights on it. I've been told, you know, I listen to Hanselman and these folks really don't write the connection string to code. So how do you- This isn't writing the connection string code. This is writing into the app service settings that you showed. Beautiful. So just the way that you showed going- So it's really that environment variable pattern. Yeah, exactly. It's like the 12 factor app. It's just the tools doing it on my behalf. That's awesome. There's nothing stored in code. Okay. You are putting me out of business, sir. All right, with that, let's go just to quickly resummarize. We talked about Azure storage. So they have multiple different storage types. What we showed here specifically was blob storage, but there's also tables, queues and files depending on your needs. It's automatically replicated, backed up. I can do optional geo-replication if I am trying to scale across different geographies and it just has a lot of capabilities available for me as mentioned. Once I put stuff in there, it's never coming out of there unless I choose to explicitly delete it. Okay. Azure Functions, we took up their serverless, which is really great because I never have to create or manage any virtual machines or clusters. Even the auto scaling thing that you showed the slider for App Service, Azure Functions will auto scale on my behalf. Right. No, nothing to have to do. It's event-driven. One other thing to even mention, there's like, I think you mentioned there's consumption-based billing. You just pay for the consumption, but you could also use one of those App Service plans if you want a more kind of always running host dedicated to your functions. I could. So you have two options. Yep, that's correct. Or if I wanted to just share my App Service by resources like my website wasn't that popular. Or you just have some resources left over. Why not? You're already paying for it. Exactly. Yeah. Okay. Perfect. And once again, we have great integration with Visual Studio for all of this stuff as we demo it. So the last thing we've talked about it a couple of times, we've hinted at it, that I'm curious if you can teach me about a little bit, Paul, but how do I modernize my DevOps process as the first time for everything? So we can do it. So you want to Paul's machine. Okay. So let's go to my machine and... Well, Paul's getting set up. I'll answer one question that came in and said, can you opt out of patches or roll them back if there's an issue on App Service? And the answer to that is no. No, yeah. You can patch your own app, but not the OS. Like you really, with this kind of a service, you really want us patching it. Now that said, I remember I said before, if something breaks with the operating system, the hardware or even IES, Microsoft's on the hook. And if we see that a patch is harming apps, Microsoft will take care of that. So like leave the rollback to Microsoft in that case. If you want more control, I would suggest looking at containers because containers you can explicitly choose the exact patch level. Perfect. Okay. Cool. So I'm back in my resource group and I'm kind of like a cooking show. I'm gonna date myself. I grew up with Julia Childs, but so here we have another app service and this one I took it just a little bit further along with some more setup. But basically what I did is I set up application insights. And this is by the way, service number five. We've been hinting at it. We're not very good at keeping secrets, but this is application insights. And it's a key part of application monitoring that is app centric. Okay. So we've already set it up. It's really easy just through this flow. If it's not set up the first time, we'll go ahead and create it. Also you showed Visual Studio, makes it quite easy. So you've got lots of ways. You click on app insights and we're gonna start showing you a lot of telemetry that I think is a developer we care about. Let's actually make sure this is the one I think I wanna look at 15, 28. Is that the right one? Yep. There we go. You probably need to increase your time range on the previous one. We've been on the stage for an hour, so we haven't been sending in any traffic. Totally, okay. So the first thing I wanna just call out is you can even use application insights to basically discover what's going on. Like what are the services? What is the service health? What are your dependencies? So here we can see that that web service, sorry that web application as a service that you deployed, that's represented as this node and then it's making calls off to storage. And I can just visualize that from a glance, which is super handy. So there's basically like this instrumentation engine that's keeping an eye on dependencies and figuring this stuff out. Another thing that's super important to watch, I think from the developer and the DevOps side is the failure rate. So I can see at a glance that I've got a bit of a failure rate going on and I can inspect those failures and just go ahead and investigate those right away. So in terms of just like you wanna understand the server failures or the exception failures in your app, this is an incredibly handy tool. And without writing any code, you're just using standard.net framework calls, right? And so those exceptions were captured. I think the other interesting thing I was pointing out is you're running the same copy that I just showed running on my machine, which we tested and saw it working. So it seems like most of the requests are succeeding but there's something that for some reason up on our production site, it's failing on occasion. Yes. So from here, is there any way I can drill in to actually figure out? Yeah, totally. So like one thing we'd be used to just looking at our server, the server logs would show 500, which is helpful. It kind of points us in the right direction. We know what's failing, but we don't know why, right? And so just to kind of figure out why, I can actually see there's exceptions underneath object not set to an up. Is this the one we wanted? Yes, we're getting an all reference exception sometimes but not all times. And it has to do with my Azure storage service. So I have a hypothesis that, remember how you promised you were gonna remind me on the connection string? Yeah, so actually this is an all reference exception. So you scroll over to the call stack, you have one frame higher, I just wrote this code so I remember it. And so- So it's even a good way to keep an eye on churn from your developers. Yeah, so actually I can see from the parameters, it looks like what's happening. It turns out the issue is I shouldn't be letting people upload images when they're not logged in. And so actually what you're catching, if we go look at that lit place in the code, you'll see that I'm trying to use the username which when you're not logged in is null. Got it. So the cool thing is kind of from a monitoring and observability standpoint, like I'm immediately alerted to there were failures. I understood the failure rate and we could zoom into any particular failure, get a stack and we're already in the right place. But this is cool because I forgot to test that scenario and obviously escaped into production because everything worked on my machine because I clicked that, remember me. And so every time I launched my site and tested it, I was just logging in like I would expect. So- That's awesome. You helped me find a failure in production that I didn't understand the repro steps for locally. Perfect. And just a couple more things I'll call out. Like if all telemetry, you can search over from the search view. Even we have a capability, I love this, my favorite new feature. You can watch a live metric stream and a live log stream. It's kind of like tailing with superpowers. Cool. So that was service number five. Very cool. And we'll back- One other question that came up before we do that. Can I set up alerts? Like so do I have to log into the portal? Absolutely. And that's the best practice. You can set up alerts programmatically or through the portal. There's another alerts blade on the portal that I didn't go through. And basically you pick the criteria and off you go. And by the way, application insights and absolutely any resource in Azure through what's called Azure monitoring has the capability to do alerts. So it's just built in. You don't have to buy like a pager duty or anything else for that. Very cool. So I can actually get emails when things start failing. I don't have to- Emails, pages, texts, you know, you can choose. Social, if you want to have it go into Slack. Great. Excellent. So go ahead and come back to my machine. So to wrap up, because we want to make sure that you get out on time. So the last service that we, service number five, application insights is part of the broader Azure monitor service. And so obviously we didn't have time to tour everything. I think each of these services that we just showed you could do- They may have a few hours, yeah. But hopefully we kind of wet your whistle and helped you understand more places to go explore to get started with Azure. But as we just saw that, having application insights on or monitoring can save me a lot of headache for remote debugging or when something fails, if I turned it on at the beginning, I just have the data. And if we show, I get call stacks and variable information. Hopefully the days of like remote desktop or FTPing or SSHing in or over just use this tool, it's so much easier. So a couple of honorable mentions. Obviously we didn't have time to talk about everything. A couple of other hosting options that are really good options depending on what you need. We have Azure Kubernetes Service. If you're looking to work with containers, it's a great place to work with containers, awesome orchestration system. Azure Service Epic Mesh is really good for distributed microservices. They have a very interesting programming model that just helps- Yeah, let's say you want to have a stateful collection instead of a cache, you could just write to the collection. And then a couple of the services that'll really help you. A few of the questions that unfortunately we didn't have time to get to, we talked about securing things. Yeah, and so Key Vault's a great service to go explore for storing certificates and secrets and things like that that are actually even more secure than storing MS environment variables. Perfect. So with that, let's kind of just go back to a couple of resources that'll help you keep getting going. We generally announce and talk about a lot of these things on the web developer blog. So that's aka.ms.webdevblog.net.azuredev, so aka.ms.netazuredev. We'll take you to our.netazured developer center where we have links to Quick Starts. Every single thing you and I talked about, there's a Quick Start that will help you get started in five minutes doing each of those services. Right, even the same sample app, all of it, right. And then lastly, we have a migrate to cloud page. If you're looking to take an existing application and migrate that from on-premises to Azure, the migrate to cloud aka.ms.migratetocloud will point you to page, so I'll give you resources to get started with that. We saw Cesar was here earlier. Watch his talk too, because that's a great resource for that. Yeah, that, thank you very much. Thank you.
Wiki should be updated to provide guidance on Bolts compatibility There is now a Bolts sample available at: https://github.com/Microsoft/WinObjC-samples. The Supported 3rd-Party Libraries wiki should be updated to provide information on how and when Bolts can be used. https://github.com/Microsoft/WinObjC/wiki/Supported-Third-Party-Libraries This particular page is referenced by our external submission site, so it would be useful to let folks know how to link in Bolts.
Portion of a watch FIG. 1 is a front view of a portion of a watch showing my new design; FIG. 2 is a perspective view thereof; FIG. 3 is a side view thereof; FIG. 4 is a side view thereof taken opposite FIG. 3; FIG. 5 is a bottom view thereof; and, FIG. 6 is a top view thereof. The broken line showing of the top and bottom of the band and the side buttons is for illustrative purposes only and forms no part of the claimed design. The un shown back of the watch forms no part of the claimed design. The ornamental design for a portion of a watch, as shown and described.
Page:Rousseau - Profession of Faith of a Savoyard Vicar, 1889.djvu/83 without temerity, form a modest conjecture or two on this subject. I reflect that, if the mind of man had remained perfectly free arid pure, what merit could he have pretended to in admiring and pursuing that order which he saw already established, and which he would lie under no temptation to disturb? It is true he would have been happy, but he could not have attained that most sublime degree of felicity—the glory of virtue and the testimony of a good conscience. We should in such a case have been no better than the angels, and without doubt a virtuous man will be one day much superior. Being united on earth to a mortal body by ties not less powerful than incomprehensible, the preservation of that body becomes the great concern of the soul, and makes its present apparent interests contrary to the general order of things, which it is nevertheless capable of seeing and admiring. It is in this situation that by making a good use of his liberty, it becomes at once his merit and his reward; and that he prepares for himself eternal happiness in combating his earthly, passions, and preserving the primitive purity of his will. But even supposing that in our present state of depravity our primitive propensities were such as they ought to be, yet if all our vices are derived from ourselves, why do we complain that we are subjected by them? Why do we impute to the Creator those evils which we bring on ourselves, and those enemies we arm against our own happiness? Ah! let us not spoil the man of nature, and he will always be virtuous without constraint, and happy without remorse. The criminals who pretend they are compelled to sin,
#! /bin/bash if [[ -d "pcl/" ]]; then echo "Removing old pcl directory" `rm -rf ./pcl` `mkdir ./pcl` fi printf "Copying includes." `cp include/pcl/* ./pcl/` for f in * ; do if [[ -d "$f/include/pcl/" && ! -L "$f/include/pcl/" ]]; then `cp -R $f/include/pcl/* ./pcl/` printf "." fi done printf "Complete\n"
Restrictions of beta server Thanks to great feature of @psh ( #857 ), now we are able to use beta cluster. Thus I wonder restrictions of using commons beta, not to vandalise. What are our restrictions by using our new feature? Can we Upload same photos several times? Upload with bad title/description/categories? Anything is OK on beta server except copyright violations.
jQuery Uncaught Typerror: Property '$' of Object is not a Function I recieve an error (Uncaught TypeError: Property '$' of object [object Object] is not a function) when trying to run the following jQuery code: $(document).ready(function(){ $('.login').click(function(){ $.getJSON('https://www.cloudflare.com/api_json.html', $('form').serialize()) }); }); I do not receive this error on the test page I set up, I only receive it on my actual site. I've tried replacing my first line of code with jQuery(document).ready(function ($) { because other posts suggested my CMS was loading jQuery in no-conflict mode, and it just returned a different error (Uncaught ReferenceError: jquery is not defined). Does anyone have any helpful suggestions? I have the following code in my header: <script type='text/javascript' src='https://ajax.googleapis.com/ajax/libs/jquery/1.7.2/jquery.min.js'></script>. That should load jQuery right? How should I verify that it is loading for sure? There are also several other scripts on the site that are using jQuery and seem to be working fine. it should be jQuery not jquery. jQuery variables are case sensitive. You have used jquery(document).ready(function($){ small case q, it should be jQuery(document).ready(function($){ You can shorten it to jQuery(function($){ $('.submit-login-info').click(function(){ $.getJSON('https://www.cloudflare.com/api_json.html', $('.logininfo').serialize()) }); }); Thanks, that worked, but now I get a different, unrelated error: XMLHttpRequest cannot load https://www.cloudflare.com/api_json.html?tkn=&email=&z=&a=rec_load_all. Origin http://www.tech-bytes.org is not allowed by Access-Control-Allow-Origin. Should this posted under a different post since it is unrelated to the original issue? Also, should answers that were intended to be helpful but didn't actually solve the issue be downvoted, or are only answers not intended to help supposed to be downvoted? This is a very common error faced with newbies in Ajax, you can search for Same Origin Policy with Ajax. There are lot of questions within SO itself regarding this The same origin policy prevents the browsers from making ajax calls to different domain than from which the page was loaded. One of the solution is to use jsonp I tried using JSONP, but cannot figure out how to do it. I tried adding &callback=mycallback because that's what [the CloudFlare API] (http://www.cloudflare.com/docs/client-api.html) specified (I think it wants somethign specified in place of mycallback, but I am not sure what) and also &callback=? because that's what I kept reading on other posts. Do I need to specify that I am trying to use JSONP and not JSON somehow, or am I doing something else wrong? I have the code stored here. I would recommend you ask a separate question about jsonp with cloudflare, and post a link here so that I also can have a look Alright, am I doing everything correctly; would what I've done generally work as JSONP, and it is just something with CloudFlare? In generally a jsonp will look like $.getJSON('https://www.cloudflare.com/api_json.html?callback=?', $('.logininfo').serialize()), I haven't gone through cloudflare api yet I posted a new question about this here. JS is case-sensitive: jquery(document).ready(function($){ ^__ This should be capitalized First of all make sure jQuery js is loaded. Also check for other javascript library that are conflicting with jQuery $. And use jQuery.noConflict(); Official Document : http://api.jquery.com/jQuery.noConflict/ In your source code, jQuery is all lowercase, "jquery". That seems to be your issue. Change it to "jQuery" and see if it works. <script type="text/javascript"> $(document).ready(function() { $('a[href=#top]').click(function(){ $('html, body').animate({scrollTop:0}, 'slow'); return false; }); }); </script> Then you can use $ withing the scope of that function, without conflicting it with other library outside the scope
Possible to post to Apollo server with protobuf? I am working on a project that currently uses protobuf to communicate from the client to the api. I would like to do a proof of concept showing communication via graphql, and am wondering if its possible to send protobuf to apollo server instead of json. also asked in spectrum for apollo server: https://spectrum.chat/apollo/apollo-server/is-there-support-in-the-road-map-at-all-for-sending-protobuf-from-the-client~dbdbc639-d7c7-4d0f-9caa-b58bb3744a90 I see a few protobuf related packages, https://www.npmjs.com/package/apollo-engine-reporting-protobuf https://www.npmjs.com/package/@apollo/protobufjs and am wondering if there are any plans for this or if it's possible. asking here because i've received no response on the community site. If you can use apollo-server-express, you can achieve that as long as you can convert the protobuf to json again on the server, which I guess it is possible with protobuf.js. apollo-server or apollo-server-express expects three key-pair values in the request body operationName variables query This is how request body looks like (logged using morgan-body) If you are using apollo-server-express, you can use your custom express middleware in which you can deserialize the protobuf to json and add the json to the req.body for Apollo Server. Changing from apollo-server to apollo-server-express just takes a few steps and no change in your apollo server config // import it from apollo-server-express instead of apollo-server const { ApolloServer } = require('apollo-server-express'); const express = require('express'); const app = express(); app.use((req, res, next) => { // Intercept the request before it reaches the Apollo Server // Use protobuf.js to deserialize the request body into json // Add the json to req.body // Call the next() to go to next middleware next(); }); const server = new ApolloServer({/* Your apollo server config */}) server.applyMiddleware({ app }); app.listen({ port: 4000 }, () => console.log(` Server ready at http://localhost:4000${server.graphqlPath}`));
In Expl3, if sequences are declared globally, why is there a \l_tmpa_seq I guess everything is in the title... But I'll rephrase here. The documentation says: \seq_new:N -> Creates a new ⟨seq var⟩ or raises an error if the name is already taken. The declaration is global. The ⟨seq var⟩ initially contains no items. But it also defines: \l_tmpa_seq -> Scratch sequences for local assignment. These are never used by the kernel code, and so are safe for use with any Latex-defined function. However, they may be overwritten by other non-kernel code and so should only be used for short-term storage. So what's the point in having a local but global variable? This is the also case for other types. if you do \seq_set:Nn the value is set locally. if you do \seq_gset:Nn it is global. the \l_ or \g_ tells you which is appropriate for this particular variable. so the existence of the variable can be global, even though it is set or changed locally. @cfr: when you write that \l_or \g_l tells you wich is appropriate... do you mean that it's just a convention? I could use \seq_gset:Nn \l_tmpa_seq {A,B,C}? most of the syntax in expl3 is just a convention. there are some checks and there may be more in the future, but, for the most part it would simply be too expensive to enforce the conventions as the underlying TeX does not make (m)any of these distinctions. so you could (probably) say e.g. \seq_gset_from_clist:Nn \l_foo_bar_tl {A,B,C} and not get an error. the pressure to conform to the conventions laid down by the LaTeX Project is largely social :-). @cfr social and \debug_on:n { all }. @Skillmon I think the second would be ineffective in the absence of the first even in the case of people who've actually heard of that. (which I certainly have not except very vaguely and extremely recently. certainly I've never used it.) \documentclass{article} \pagestyle{empty} \begin{document} \ExplSyntaxOn \seq_new:N \l_pinkfloyd_local_seq \seq_new:N \g_pinkfloyd_global_seq \cs_new_protected:Nn \pinkfloyd_demo: { local : ~ \seq_use:Nn \l_pinkfloyd_local_seq {,} \par global : ~ \seq_use:Nn \g_pinkfloyd_global_seq {,} \par } \seq_set_from_clist:Nn \l_pinkfloyd_local_seq {aardvarks,bonobos,civets} \seq_gset_from_clist:Nn \g_pinkfloyd_global_seq {aardvarks,bonobos,civets} \pinkfloyd_demo: \group_begin: \seq_set_from_clist:Nn \l_pinkfloyd_local_seq {A,B,C,D,E} \seq_gset_from_clist:Nn \g_pinkfloyd_global_seq {W,V,X,Y,Z} \pinkfloyd_demo: \group_end: \pinkfloyd_demo: \ExplSyntaxOff \end{document} Your example is great, even the variable names are well chosen! actually, the example should set the variables inside the group. Variables in expl3 are all declared globally, but may be set either locally or globally. Thus when you do \seq_new:N \l_my_foo_seq the name is reserved globally and set at global scope to an empty sequence. However, this sequence should only be set locally.
Cameron Cameron Staton Muti-Genre Project Concrete Poem 182 word count Essay 485 word count Interview471 word count Moonshine Jug 57 word count Prohibition Definitions79 word count Prayer311 word count Obituary 136 word count Citation
Remote control for welders and method therefor ABSTRACT A method and system for remotely controlling operational parameters of welders by communicating over a welding cable thereof with a remote device electrically coupleable between a workpiece and an electrode holder of the welder. The operational parameters include remotely enabling and disabling the welder, remotely choking the engine thereof, remotely controlling coarse and fine current adjustment, and remotely controlling other parameters otherwise controllable from a control panel of the welder. The remote device having one or more operational parameter signal circuits electrically coupleable between the workpiece and the electrode holder for producing unique operational signals on the welding cable, a sensor circuit that detects the operational signals and produces corresponding operational sense signal in response thereto, and an operational control circuit that actuates operation controls corresponding to the operational parameters on the control panel of the welder in response to the operational sense signals. BACKGROUND OF THE INVENTION The invention relates generally to welders, and more particularly to systems and methods for remotely controlling welders using a remotedevice interposable between a workpiece and an electrode holder and communicating along the welding cable to control various operational parameters otherwise controllable locally from a control panel thereof. Welding machines, or welders, are known generally and used widely for a variety of welding applications. Engine-driven welders include a combustion engine-driven generator that provides an electric welding current, and static welders obtain welding current from an electrical converter connected to an electrical power supply line. The welding current and other operational parameters of these and other welders are generally selectable and controllable locally from a control panel thereof. In the Bobcat™ 225G Plus engine-driven welder, available from Miller Electric Mfg. Co., an Illinois Tool Works Company, Appleton, Wis., forexample, the control panel includes AC/DC selects, polarity reversing selects, and coarse and fine current adjustment controls. The controlpanel of the Bobcat™ 225G Plus engine-driven welder also includes engine choke, start, run/off and low idle controls. Other welders may control some or all of these operational parameters. Static welding machines forexample do not require engine choke and idle controls. Other operational parameters include wire feed and auxiliary power controls. In many applications, especially engine-driven welder applications,welding is performed at a site remote from the welder and the controlpanel thereof. For convenience, it is known to control a limited numberof welding parameters remotely from the control panel. Namely, coarse and fine current adjustment. U.S. Pat. No. 4,216,367 entitled "Wireless Remote Control for Electric Welder" issued Aug. 5, 1980 to Risberg and assigned commonly herewith, for example, discloses a device for remotely controlling weld current in an engine-driven welder to precise percentages of a current value selected previously at the control panel thereof. U.S. Pat. No. 4,216,367 discloses, more particularly, an adjustable rheostat temporarily disposable between a weld stick andworkpiece to generate a calibrating current, which flows through thewelding cable. The calibrating current is sensed by a current transformer and compared to a previously set and stored welding current reference, which is adjusted automatically after a slight delay to a new desired weld current value. The present invention is drawn toward advancements in the art of welders, and more particularly to systems and methods for remotely controlling various operational parameters of welders including engine-driven and static welders. The inventors of the present invention recognize the desirability of remotely starting and stopping welders, particularly engine-driven welders, in addition to or besides controlling weld current. More particularly, remotely starting and stopping welders without any additional cables between the remote site and the control panel thereof.Remote welder control eliminates the necessity of walking back and forth between the control panel and the work-site, which may be located a significant distance therefrom. In some applications, moreover, a plurality of welders are arranged in large banks of welders, which are operated by many welding operators at remote locations. The operators may not know, and it may be very difficult to ascertain, which welder is coupled to a particular welding cable. The advantage of remotely enabling and disabling the welder under these circumstances is self evident. Remote enabling and disabling engine-driven welders is desirable also for reducing engine and generator operation time, thereby extending the useable life thereof as well as other components of thewelder. The known prior art however does not teach remote enabling and disabling of welders. It is therefore an object of the invention to provide novel methods and systems for remotely controlling welders, and more particularly for remotely controlling operational parameters thereof by communicating over a welding cable with a remote device electrically coupleablebetween a workpiece and an electrode holder. It is also an object of the invention to provide novel methods and systems for remotely controlling at least one, and preferably more than one, operational parameter of welders otherwise controllable locally from a control panel thereof, including remotely enabling and disabling the welder, remotely choking engine-driven welders, remotely controlling coarse and fine current adjustment, remotely reversing current polarity,among other operational parameters thereof. It is a more particular object of the invention to provide novel methods and systems for remotely controlling welders having generally a weldingcable and an electrode holder for supplying weld current to a workpiece,and one or more operation controls for locally controlling corresponding operational parameters from a control panel thereof. The novel methods and systems include a remote device having one or more operational parameter signal circuits electrically coupleable between the workpieceand the electrode holder for producing corresponding unique operational signals on a welding cable, a sensor circuit for detecting the operational signals and producing corresponding operational sense signals in response thereto, and an operational control circuit foractuating corresponding operation controls of the welder in response tothe operational sense signals, thereby remotely controlling operational parameters of the welder corresponding to the unique operational signals produced by the remote device on the welding cable. It is another more particular object of the invention to provide novel methods and systems for remotely controlling welders with a remotedevice having one or more constant current source circuits electricallycoupleable between the workpiece and the electrode holder for selectively producing one or more corresponding unique operational signals on the welding cable. It is a related object of the invention toprovide a battery powered remote device having one or more constant current source circuits coupled to a timer circuit that produces unique intermittent operational signals on the welding cable, whereby the intermittent operational signals reduce battery power consumption. It is still another more particular object of the invention to provide novel methods and systems for remotely controlling welders having a current transformer sensor circuit for detecting operational signals produced by an alternating carrier wave on the welding cable. It is a further object of the invention to provide novel methods and systems for remotely controlling welders with a remote device electrically coupleable between a workpiece and an electrode holder,wherein the remote device includes a first electrode coupleable to theworkpiece and one or more second electrodes coupleable to the weld stick, or electrode holder, for generating unique operational signals onthe welding cable, whereby the second electrodes are readily andaccessibly contactable by the electrode holder to selectively remotely control corresponding operational parameters of the welder. It is a related object of the invention to magnetically and electrically couple the first electrode to ferromagnetic workpieces. These and other objects, aspects, features and advantages of the present invention will become more fully apparent upon careful consideration ofthe following Detailed Description of the Invention and the accompanying Drawings, which may be disproportionate for ease of understanding,wherein like structure and steps are referenced generally by corresponding numerals and indicators. BRIEF DESCRIPTION OF THE DRAWINGS FIG. 1a is a block diagram of a remotely controllable engine-driven welder system according to an exemplary embodiment of the present invention. FIG. 1b is a block diagram of a remotely controllable static welder system according to an exemplary embodiment of the present invention. FIG. 2a is a top plan view of a pendant for housing a remote devicedisposeable between a weld stick and workpiece for controlling operational parameters of a welder. FIG. 2b is a partial sectional view along lines b--b of FIG. 2a. FIG. 3 is an exemplary schematic block diagram of a remote device according to an exemplary embodiment of the present invention. FIGS. 4a and 4b are detailed electrical circuit diagram of a remotedevice according to an exemplary embodiment of the present invention. FIGS. 5a and 5b are detailed electrical instrumentation circuit according to an exemplary embodiment of the present invention. FIGS. 6a and 6b are partial detailed microprocessor based controller according to an exemplary embodiment of the present invention. FIGS. 7a and 7b are partial detailed microprocessor based controller according to an exemplary embodiment of the present invention. DETAILED DESCRIPTION OF THE INVENTION FIG. 1a is a system 10 for remotely controlling one or more operational parameters of an engine-driven welder having a welding cable and an electrode holder for supplying welding current to a workpiece W, and one or more operation controls for controlling corresponding operational parameters thereof. The system 10 remotely controls one or more operational parameters otherwise controllable locally from a controlpanel of the welder, including remotely enabling and disabling thewelder, remotely choking the engine and controlling the idle thereof,remotely controlling coarse and fine current adjustment, and remotely reversing current polarity, among other operational parameters of thewelder. In FIG. 1a, the engine-driven welder portion of the system 10 comprises generally a combustion engine 20 rotatably coupled to a generator 22having a rectifier 24 at its output for supplying welding current through a first cable portion 32 coupled to the workpiece W, and a second cable portion 34 coupled to an electrode holder 36, which includes an electrode, or weld stick, or stinger 37. The engine-driven welder includes a control panel 40 with operation controls for controlling operational parameters thereof locally as is known generally. The operation controls include, for example, a start switch,a run switch, a choke switch, a variable coarse current increase and decrease control 25, and a variable fine current increase and decrease control 27. The operation controls may be electrical contacts or solid state switching devices or other controls. Most ignition systems, forexample, include generally a start switch and a run switch, wherein thestart and run switches may be electrical contacts or solid state switches. The coarse current control is often a multi-tapped transformer, and the fine current control is usually a variablerheostat. FIG. 1b is a static welder system 11, which is similar in most respects to the engine-driven welder system 10 of FIG. 1a. except that the combustion engine 20 and generator 22 are replaced by an electrical converter 26 that converts an alternating current supply line input toan appropriate welding current supplied through a first cable portion 32coupled to the workpiece W, and a second cable portion 34 coupled to an electrode holder 36 having an electrode 37. The static welder also includes a control panel 40 with operation controls for controlling operational parameters thereof locally similar to the controls on theengine-driven welder with the exception of the choke and possibly someother operation controls. The static welder system 11 also includes generally coarse and fine current controls, among other controls. The welder systems of FIGS. 1a and 1b further comprise generally a remote device 50 having one or more operational parameter signal circuits electrically coupleable between the workpiece W and theelectrode holder 36 for selectively producing unique operational signal son the welding cable, a signal sense circuit 60 for detecting the operational signals produced on the welding cable and for producing correspondingly unique operational sense signals in response thereto,and an operational control circuit 70, which may control drivers or relays 72 coupled to and actuating corresponding operation controls ofthe welder in response to the unique operational sense signals, thereby remotely controlling the operational parameters of the welder. FIGS. 2a and 2b illustrate a hand held battery powered pendant 80 for housing the one or more operational parameter signal circuits of the remote device 50 for selectively producing the one or more corresponding operational signals on the welding cable when electrically coupled between the workpiece W and the electrode holder 36. The pendent 80 hasexposedly mounted thereon a first electrode 51 of the remote device 50coupleable to the workpiece W and one or more second electrodes 52 ofthe remote device, only one of which is shown in FIGS. 2a and 2b,corresponding to the one or more operational parameter signal circuits,which are coupleable to the electrode holder 36 for generating corresponding operational signals on the welding cable. The pendent 80includes a battery storage bay 81 for mounting and electrically coupling one or more batteries B to the operational parameter signal circuits ofthe remote device 50 also housed therein. The batteries B are required to power the remote device 50 to produce operational signals on the weld cable during some phases of operation as discussed further below. In one embodiment, the first electrode 51 is preferably disposed on an end 82 of the pendent 80 where it is more readily electricallyengageable with the workpiece W by an operator wielding the pendant 80.The one or more second electrodes 52 are disposed preferably on an upper surface 84 of the pendant 80 where they are readily and accessiblycontactable by the electrode holder 36, including the electrode 37thereof, to remotely control corresponding operational parameters of thewelder as discussed further below. The first and second electrodes 51and 52 of the remote device 50 include preferably pointed tips to facilitate electrically contacting the workpiece W and the electrodeholder 36, despite the tendency for the formation of oxides on the electrodes 51 and 52, and on the electrode 37 of the electrode holder36, and on the workpiece W. In another embodiment, the first electrode 51 is or includes a magnetic portion 53 disposed on a lower surface 86 of the pendant 80, wherein the magnetic portion 53 is magnetically and electrically coupleable toferromagnetic workpieces. The pendent 80 may thus be adheredmagnetically to the workpiece about a surface thereof by the operator during use. According to this aspect of the invention, the pendant 80may be repositioned and remain adhered and electrically coupled to theworkpiece when not in use, thereby reducing the possibility ofmisplacement thereof and eliminating the necessity of having to pocket or otherwise secure the pendant 80 between uses, which is a particularly convenient feature of the invention. FIG. 3 illustrates schematically a remote device 50 having the first electrode 51 and a plurality of second electrodes 52, 54, 55, 56, 57,and 58, corresponding to a plurality of remotely controllable operational parameters of the welder, coupled to operational parameter signal circuits 90 that produce corresponding operational signals on thewelding cable when electrically coupled between the workpiece W and theelectrode holder 36. The operational parameter signal circuits 90include, for example, among others a start signal circuit, a choke signal circuit, coarse current increase and decrease circuits, and fine current increase and decrease circuits for producing on the weldingcable unique start, choke, coarse increase and decrease, and fine increase and decrease signals, respectively, depending on the particular type of welder. FIGS. 4a and 4b illustrate an exemplary operational parameter signalcircuit of the remote device 50 for producing an operational signal onthe welding cable when electrically coupled between the workpiece W andthe electrode holder 36. The circuit includes a constant current source formed generally by op-amp OA1, transistor Q1, resistors R6 and R7, and capacitor C3. The output of the constant current source, and more particularly the collector of transistor Q1, is coupled to a corresponding one of the second electrodes 52 and 53-58, which is selectively coupleable to the electrode holder 36 to produce an operational signal on the welding cable. In the exemplary embodiment,diode D1 is disposed between the collector of Q1 and the electrode. The amplitude of the operational signal produced by the exemplary constant current source circuit is dependent on the values of resistorR1, R2, R3, R7. R8, R9 and R10. In embodiments that control more than one operational parameter of the welder, there is a corresponding numberof constant current source circuits electrically coupled in parallel inthe remote device 50, wherein the collector of the transistor of each constant current source circuit is coupled to a corresponding first electrode, which is selectively contactable by the electrode holder 36in the hands of a welding operator to produce the desired operational signal on the welding cable. The current amplitude corresponding to the operational signals produced on the welding cable is unique for each remotely controlled operational parameter of the welder, as discussed further below, by appropriately selecting different resistive values R1,R2, R7 and R8 for each of the constant current source circuits. FIGS. 4a and 4b illustrate the operational parameter signal circuit powered by a battery BAT1 coupled between the input pin 5 of the op-ampOA1 and the first electrode 51. The battery BAT1 also supplies power tothe collector of the transistor Q1. The battery BAT1 is required topower the remote device 50 before starting the welder. After the welder is powered, however a DC voltage supplied by the welder is available onthe welding cable between the workpiece W and the electrode holder 36,which may be as high as 85 volts or more on some welders. The DC voltage from the welder is thus available to power the operational parameter signal instead of the battery BAT1 when the welder is powered and the remote device 50 is electrically coupled between the workpiece W and theelectrode holder 36. The remote device 50 of the exemplary embodiment requires at least approximately 4.8 volts to operate, although other embodiments may be configured to operate on less voltage. The remote device 50 also includes a switch SW1 for reversing the powersupply polarity to the operational parameter signal circuit, and more particularly to the constant current source thereof. The power supply polarity reversal is required when the polarity of the welder is reversed at the control panel thereof as is desired for some welding applications. FIGS. 4a and 4b illustrate a timer circuit coupled between the op-ampOA1 and the first electrode 51 including a 555 timer U1 having its output pin 3 coupled to the input pin 5 of the op-amp OA1 by transistorQ3 to intermittently drive the constant current source, thereby producing an intermittent operational signal on the welding cable. Inone embodiment, the timer circuit is configured to produce an ON/OFF pulse with a 20 percent duty cycle, which in one configuration is ON for100 milliseconds and is OFF for 400 milliseconds. The duty cycle of the555 timer U1 is configurable by appropriate selection of resistorsR13-R15, R18, R19 and C10. Other duty cycles may be used alternatively.The timer circuit is powered by a regulated voltage V+REG, as known generally, which is powered by either the battery or the DC voltage available on the welding cable. The intermittent operational signals generated by the constant current source reduce battery power consumption, which is particularly desirable when the remote device 50is used to remotely start the welder and to remotely control other operational parameters thereof before the welder is running. In embodiments having a plurality of constant current source circuits electrically coupled in parallel for controlling a corresponding plurality of operational parameters of the welder, as discussed above, asingle timer circuit may modulate one or more of the constant current source circuits through a common driver transistor Q3. Thus, the remotedevice 50 includes generally a plurality of constant current source circuits arranged in parallel, and coupled to a corresponding second electrode for producing a corresponding plurality of unique operational signals on the welding cable when electrically coupled between theworkpiece W and the electrode holder 36. The constant current source circuits may be driven by an intermittent drive signal from a timercircuit, which is desirable for reducing power consumption when remotely controlling operational parameters of the welder under battery power. In the exemplary embodiment generally, the start signal circuit is a constant current source circuit, and the start signal produced on thewelding cable is preferably a relatively low amplitude intermittent operational signal of approximately 0.5 amps and having a 20 percent duty cycle thereby minimizing battery power consumption since the remotedevice 50 operates on battery power to produce the start signal. Other operational signals produced by the remote device on the welding cable prior to starting the welder, for example the choke signal, are also preferably low amplitude intermittent signals to reduce battery power consumption. The higher amplitude current signals are preferably reserved for remotely controlling operational parameters while thewelder is running, for example the coarse and fine current adjustments,since the energized welding cable has a low DC voltage thereon, which may be used by the remote device 50 for producing the higher amplitude signals. FIG. 4a also illustrates an indicator circuit with a visual display for indicating when the welder is providing a DC voltage on the weldingcable, and thus that the welder is powered. In the case of an engine-driven welder, the visual display is indicative that the engine is running. The indicator circuit includes a current limiting resisterR20, and a zener diode D6 in series with a lamp LED1, which is preferably mounted in a visually conspicuous location on the pendant 80. FIGS. 1a and 1b illustrate a signal sense circuit 60 for detecting the operational signals produced on the welding cable, and for producing corresponding operational sense signals in response thereto. The signal sense circuit 60 is disposed generally along the welding cable, wherein the welding cable includes the welding cable portions 32 and 34 and anyother conductor in the rectifier 24 and generator 22 or electrical converter 26 supplying current thereto, and through which the operational signals are communicated and detectable by the signal sense circuit 60. The signal sense circuit 60 may include a differential amplifier coupled across a low resistance resistor, for example a 5 ohm resistor, disposed in series along the welding cable, wherein the resistor is protected by two diodes arranged in parallel with the resistor, but with polarities reversed. In a preferred embodiment, the signal sense circuit 60 is a current sensor circuit, for example a commercially available current transformer. According to this embodiment, the remote circuit 50produces preferably an alternating carrier wave for the operational signals. The alternating carrier wave distinguishes the operational signals from noise on the welding cable thereby facilitating detection of the operational signals by the current transformer. FIG. 4aillustrates, more particularly, an oscillator circuit coupled to the input pin 5 of the constant current source op-amp OA1, whereby the timercircuit modulates the AC signal from the oscillator circuit at the input pin 5 of the constant current source. The timer circuit is not an essential feature, although it remarkably increases the longevity of the batteries. The oscillator circuit includes generally an op-amp U2, wherein the oscillation frequency is dependant upon the resistors R1-R3, R9, R10 andthe capacitor C6 coupled thereto. In the exemplary embodiment, the frequency is preferably not less than approximately 900 Hertz, but maybe more or less. Frequencies at this exemplary level, and higher,increase the signal strength from the current transformer sense circuit60, and facilitate improved filtering thereof by an instrumentation circuit thereby providing less noisy signals to the controller 70discussed further below. In embodiments having a plurality of constant current source circuits electrically coupled in parallel for controlling a corresponding plurality of operational parameters of the welder, as discussed above, a single oscillator circuit may provide the AC carrier signal to the plurality of constant current source circuits. FIGS. 5a and 5b illustrate an exemplary instrumentation circuit coupled to the output of a current transformer signal sense circuit 60. In FIG.5a, the instrumentation circuit includes a differential amplifier circuit formed by op-amps A11-A13. The gain of the amplifiers A11, A12and A13 is set by feedback resistors R33-R35, R43, R44, R45 and R47 and corresponding parallel capacitors C4, C6 and C7. The input to the instrumentation amplifier is clamped by zener diode D5, and a highwattage resistor R39 dissipates excessive power applied thereto. In embodiments that include the alternative resistive signal sense circuit discussed above, the instrumentation circuit portion of FIG. 5a may notbe required. FIG. 5a illustrates further a signal stabilization and wave shaping circuit to stabilize and shape the operational sense signal generated bythe sense signal circuit 60 in response to the operational signal produced by the remote device 50 on the welding cable. The signal stabilization and wave shaping circuit includes op-amps A14 and A23,rectifying diodes D21 and D31, resistors R31, R32 and R40-R42, and capacitor C5 for rectifying any negative component of the signal from op-amp A13. The signal stabilization and wave shaping circuit may alternatively be coupled to the output of the alternative resistivesignal sense circuit 60 discussed above. FIG. 5b also illustrates an operational signal sense disable circuit coupled generally to the output of the signal sense circuit 60, and inthe exemplary embodiment to the output of the signal stabilization and wave shaping circuit at capacitor C5. A transistor Q7 pulls the capacitor C5 to ground in response to an input signal thereto indicating that the welder is providing power either for welding or to some other load coupled to an accessory AC outlet at the control panel of thewelder. In engine-driven welders, the input signal driving transistor Q7 is supplied, for example, from an idle module or an excitation rectifier ofthe welder. The signal is rectified by diode D6 and filtered by capacitors C31, C32 and resistor R48, and is isolated optically from the transistor Q7 by switch U6. When transistor Q7 is turned ON, by the presence of the input signal from the idle module or the excitation rectifier of the welder, the operational sense signal at the output of capacitor C5 is grounded, thereby blanking any input to the controller,and more particularly the comparator circuits of FIG. 6a discussed further below, to disable the remote control of the welder. In static welders the input signal driving the transistor Q7 is supplied from aload sensor, like a current transformer, in the electrical converter 26. The operational sense signals generated by the signal sense circuit 60are provided generally to a bank of comparator circuits of FIG. 6a. Inthe exemplary embodiment, the operational sense signals are supplied tothe instrumentation circuit and the signal stabilization and wave shaping circuit, which in turn supply the operational sense signal tothe comparator circuits of FIG. 6a. There are eight comparator circuitsU1 and U2. Six of the comparator circuits correspond to remotely controllable operational parameters of the welder, which may for example be the operational parameters identified on the remote device in FIG. 3.There may be more or less, however, corresponding to the desired numberof operational parameters controlled remotely. The six comparator circuits produce an output signal when the amplitude of the operational sense signal produced by the signal sense circuit 60 input thereto is ata threshold level determined by the resistive divider network formed by resistors R4-R7 and R16-R19 and R21. In the exemplary embodiment, forexample, the threshold voltage for the comparator circuits differs by0.5 volts and ranges between 0.5 volts and 4 volts, and the six comparator circuits corresponding to the operational sense signals have threshold voltages between 0.5 and 3.0 volts. In the exemplary embodiment, the comparator having the lowest threshold of 0.5 volts and the comparator having the highest threshold of 4.0volts define upper and lower limits on the range of signals interpreted by the microcontroller U4, wherein signals outside these range limits are disregarded as noise, and may be indicative that the welder is supplying power to a load, either via the welding cable or accessory AC outlets thereon. FIG. 6b illustrates the comparator circuit outputs coupled to a microprocessor based controller U4. The microcontroller U4 is programmed to determine which operational signal was produced by the remote device50 on the welding cable based on the combination of comparator output signals produced in response to the operational sense signal input thereto. For example, a 1.5 volt signal generated by the sense signalcircuit 60 in response to an operational signal produced on the weldingcable will produce outputs from the three comparators circuits having thresholds of 0.5 volts, 1.0 volts and 1.5 volts. The microcontroller U4is programmed to interpret the outputs from the comparator circuits as corresponding to a particular operational signal, and responsivelyactuates or de-actuates one or more corresponding operation controls atthe control panel of the welder as discussed further below. Thus by producing unique operational signals on the welding cable with the remote device 50, which in the exemplary embodiment are defined by unique current amplitudes, the remote device 50 can transmit operational signals from the remote weld site along the welding cable back to the microcontroller U4, which interprets the operational signals and performs some control function at the control panel of the welder inresponse thereto as discussed further below. FIG. 7a illustrates outputs of the microcontroller U4 coupleable to several exemplary operation controls of an engine-driven welder controlpanel including more particularly an idle operation switch, a choke operation switch, a starter operation switch and a run operation switch.Generally, the outputs on lines 2, 3, 6 and 8 of the microcontroller drive corresponding FETs Q6, Q10, Q11 and Q3, which energizecorresponding relays CR3, CR2, CR1 and CR4, respectively, in response to outputs from the microcontroller U4. The relays CR3, CR2, CR1 and CR4are tied to the existing run, starter, choke, and idle operation switches in the control panel of the engine-drive welder. Engine-driven welders include generally a start switch and a run switch for starting and running the welder. The run switch is usually actuated"on" initially by a switch at the control panel of the welder. To start the engine-driven welder, the start signal circuit in the remote device50 is electrically coupled between the workpiece W and the electrodeholder 36 to produce a start signal on the welding cable. This is performed by touching the electrode holder 36 onto the run/stop electrode 54 illustrated in FIG. 3 and contacting the first electrode 51to the workpiece W. The sensor circuit 60 detects the start signal and produces a start sense signal in response thereto, which is transmitted to the controller 70, or microcontroller U4. If the welder is not running when the microcontroller U4 receives the start signal sense signal from the signal sensor 60, the microcontroller U4 energizes the starter relay to actuate the start switch of the welder. The microcontroller U4 determines whether the welder is running by monitoring the ignition coil frequency as discussed further below. Afterthe welder starts and is running, the Microcontroller U4 de-energizesthe starter relay, thereby de-actuating the start switch. If the engine-driven welder is already running when the microcontrollerU4 receives the start sense signal, the microcontroller will at least momentarily de-energize the run relay in response thereto, thereby stopping the engine-driven welder. The starting and stopping of thewelder may be controlled remotely by electrically coupling the start signal circuit between the workpiece W and the electrode holder 36. Thefirst start signal produced by the remote device on the welding cable starts the welder and a subsequent start signal stops the welder. FIG. 7a illustrates an ignition coil of the engine-driven welder coupled to the microcontroller, which monitors the frequency thereof to determine whether the engine-driven welder, and more particularly theengine 20 thereof, is running or is merely being cranked by the starter motor thereof. The ignition coil frequency signal is filtered by resistor R8 and capacitor C9, and is optically coupled by U7 to the trigger input of a 555 timer circuit U5, which provides an input to pin11 of the microcontroller U4. A relatively low frequency signal,approximately 8 Hertz in the exemplary embodiment, is indicative thatthe engine 20 is being cranked by the starter motor but is not yet running, whereas a higher frequency signal is indicative that theengine-driven welder is running. The ignition coil frequency signal is used by the microcontroller U4 to determine when the engine is running and when to disable the starter switch when the engine-driven welder is started remotely as discussed above. In the exemplary embodiment, the microcontroller U4 is programmed to disable the starter switch when the ignition coil frequency signal is approximately 16 Hertz. In engine-driven welder applications, the microcontroller U4 is preferably programmed to increase the idle of the engine-driven wel derby energizing relay CR4 on the output 8 thereof prior to starting theengine so that the engine starts at high idle. The engine-driven welder includes a load sensor circuit that increases or decreases the idle based on the electrical load drawn through the welding cable or by an auxiliary electrical outlet located on the front panel, as is known. The load sensor circuit will subsequently reduce the engine idle after starting in the absence of a load. Engine-driven welders may also include an audio or visual indicator thatis momentarily energized prior to starting the engine as a safety feature. FIG. 5c illustrates, more particularly, the microcontroller U4output 7 coupled to a FET Q9. The controller U4 is programmed to temporarily energize a horn or other indicator driven by the FET Q9prior to starting the welder in response to a start signal from the remote device. The microcontroller U4 may also be programmed to temporarily energizethe choke relay CR1 to choke the engine-driven welder prior to starting,or while cranking the engine if the engine does not start after a predetermined time period upon actuating the starter switch. In the exemplary embodiment of FIG. 3, the choke is also remotely controllable by the remote device 50, which may be desired prior to starting theengine-driven welder. To choke the engine-driven welder, a choke signalcircuit in the remote device 50 is electrically coupled between theworkpiece W and the electrode holder 36 to produce a choke signal on thewelding cable. The sensor circuit 60 detects the choke signal and produces a choke sense signal in response thereto, which is transmitted to the controller 70, or microcontroller U4 in FIG. 6b. FIG. 7a also illustrates an output line 9 of the microcontroller U4coupled to FET Q1 for driving lamp LED1, which may be programmed for diagnostic purposes, for example to indicate the presence of noisy signals from a current sensor circuit 60 which are outside the upper range limit of the comparator circuits of FIG. 6a during calibration ofthe system. The microcontroller U4 may more generally be coupled to alternative or additional operation controls as discussed hereinabove, including coarse and fine current controls, a polarity reversal switch, an AC/DC select switch, and a wire feed switch, among others, on the control panel ofthe welder. In the case of the coarse and fine current adjustment controls, the microcontroller U4 includes corresponding fine current increase and decrease outputs and coarse current increase and decrease outputs, which drive corresponding stepper motors or solenoids or other known control devices that increase and decrease the coarse and fine currents. Forexample, each time the electrode holder 36 is touched onto the electrode57 of the remote device 50 corresponding to the fine current increase,as shown in FIG. 3, a fine increase signal is produced on the weldingcable and is detected by the signal sensor 60, which in turn generates a corresponding current increase sense signal that is transmitted to the controller 70. The controller 70 then responsively produces a control signal that operates the fine current increase control, usually a stepper motor or solenoid controlled variable rheostat, to increase the fine current some predetermined incremental amount programmed into the controller. The other current controls operate similarly. The controller70 thus controls the operation controls on the control panel of thewelder in response to corresponding operation signals produced on thewelding cable by the remote device 50 thereby permitting remote control thereof. While the foregoing written description of the invention enables one of ordinary skill in the art to make and use what is at present considered to be the best mode of the invention, it will be appreciated and understood by those of ordinary skill the existence of variations,combinations, modifications and equivalents within the spirit and scope of the specific exemplary embodiments disclosed herein. The microcontroller may for example be replaced with analog and or other digital circuit y. The present invention is therefore to be limited not by the specific exemplary embodiments disclosed herein but by all embodiments within the scope of the appended claims. What is claimed is: 1. A remotely controllable welder having a weldingcable and an electrode holder for supplying weld current to a workpieceand a start switch for starting the welder, the system comprising:a remote device having a start signal circuit having a constant current source circuit electrically coupleable between the workpiece and theelectrode holder, the start signal circuit produces a start signal onthe welding cable when coupled between the workpiece and electrodeholder; a sensor circuit disposed along the welding cable, the sensorcircuit producing a start sense signal in response to detecting thestart signal on the welding cable; a start control circuit coupled between the sensor circuit and the start switch, the start control circuit at least temporarily actuating the start switch in response to detection of the start sense signal; whereby the welder is energizablewhen the start signal circuit is electrically coupled between theworkpiece and the electrode holder. 2. The system of claim 1, the sensorcircuit is a current sensor circuit, the start signal circuit further comprises an oscillator circuit coupled to the constant current source circuit, the oscillator circuit produces an alternating carrier wave forthe start signal. 3. The system of claim 1 further comprising a timercircuit coupled to the constant current source circuit, the timercircuit produces an intermittent start signal, a battery power supply coupled to the constant current source circuit and the timer circuit. 4.The system of claim 1 further comprising a pendant housing the remotedevice, first and second electrodes disposed on the pendant and coupled to the start signal circuit, the first electrode couples the start signal circuit to the workpiece and the second electrode couples thestart signal circuit to the electrode holder. 5. The system of claim 4,the first electrode is a magnetic electrode mounted on a surface of the pendant, the magnetic electrode is magnetically and electricallycoupleable to ferromagnetic workpieces. 6. The system of claim 1, thewelder having a coarse current increase control and a coarse current decrease control, the remote device further comprising:a coarse current increase circuit electrically coupleable between the workpiece and theelectrode holder, the course current increase circuit producing a coarse increase signal on the welding cable when coupled between the workpieceand electrode holder; the sensor circuit producing a coarse increase sense signal in response to detecting the coarse increase signal on thewelding cable; a coarse increase control circuit coupled between the sensor circuit and the coarse current increase control, the coarse increase control circuit actuating the coarse current increase control in response to detection of the coarse increase sense signal; a coarse current decrease circuit electrically coupleable between the workpieceand the electrode holder, the course current decrease circuit producing a coarse decrease signal on the welding cable when coupled between theworkpiece and electrode holder; the sensor circuit producing a coarse decrease sense signal in response to detection of the coarse decrease signal on the welding cable; a coarse decrease control circuit coupled between the sensor circuit and the coarse current decrease control, the coarse decrease control circuit actuating the coarse current decrease control in response to detection of the coarse decrease sense signal,whereby coarse current is increasable and decreasable in response to electrically coupling a corresponding one of the coarse current increase circuit and the coarse current decrease circuit between the workpieceand the electrode holder. 7. The system of claim 1, the welder having a fine current increase control and a fine current decrease control, the remote device further comprising:a fine current increase circuit electrically coupleable between the workpiece and the electrode holder,the fine current increase circuit producing a fine increase signal onthe welding cable when coupled between the workpiece and electrodeholder; the sensor circuit producing a fine increase sense signal inresponse to detecting the fine increase signal on the welding cable; a fine increase control circuit coupled between the sensor circuit and the fine current increase control, the fine increase control circuitactuating the fine current increase control in response to detection ofthe fine increase sense signal; a fine current decrease circuit electrically coupleable between the workpiece and the electrode holder,the fine current decrease circuit producing a fine decrease signal onthe welding cable when coupled between the workpiece and electrodeholder; the sensor circuit producing a fine decrease sense signal inresponse to detecting the fine decrease signal on the welding cable; a fine decrease control circuit coupled between the sensor circuit and the fine current decrease control, the fine decrease control circuitactuating the fine current decrease control in response to detection ofthe fine decrease sense signal, whereby fine current is increasable anddecreasable in response to electrically coupling a corresponding one ofthe fine current increase circuit and the fine current decrease circuit between the workpiece and the electrode holder. 8. The system of claim1, the welder is an engine-driven welder having a run switch, the system further comprising a run control circuit coupled between the sensorcircuit and the run switch, the run switch de-actuatable in response toa start sense signal produced on the welding cable when the start signalcircuit is electrically coupled between the workpiece and the electrodeholder after the engine-driven welder is energized, whereby theengine-driven welder is de-energized. 9. The system of claim 1, thewelder is engine-driven welder having an indicator switch for at least temporarily enabling an indicator, the system further comprising an indicator control circuit coupled between the sensor circuit and the indicator switch, the indicator control circuit actuating the indicator switch at least temporarily in response to the start sense signal before the start switch is closed. 10. The system of claim 1, the welder is an engine-driven welder having choke switch for at least temporarily enabling a choke, the remote device further comprising:a choke signalcircuit electrically coupleable between the workpiece and the electrodeholder, the choke signal circuit producing a choke signal on the weldingcable when coupled between the workpiece and electrode holder; the sensor circuit producing a choke sense signal in response to detecting the choke signal on the welding cable; a choke control circuit coupled between the sensor circuit and the choke switch, the choke control circuit actuating the choke switch in response to detection of the choke sense signal, whereby the choke is at least temporarily enabled inresponse to electrically coupling the choke signal circuit between theworkpiece and the electrode holder. 11. A remotely controllable welder having a welding cable and an electrode holder for supplying weld current to a workpiece, and at least one operation control for controlling an operational parameter of the welder, the system comprising:a remote device having a constant current source circuit electrically coupleable between the workpiece and the electrode holder,the constant current source circuit produces an operational signal onthe welding cable when coupled between the workpiece and electrodeholder; a sensor circuit disposed along the welding cable, the sensorcircuit producing an operational sense signal in response to detecting the operational signal on the welding cable; an operational control circuit coupled between the sensor circuit and the operation control,the operational control circuit actuating the operation control inresponse to detection of the operational sense signal, whereby the operational parameter of the welder is controllable when the constant current source circuit is electrically coupled between the workpiece andthe electrode holder. 12. The system of claim 11 further comprising a timer circuit coupled to the constant current source circuit, the timercircuit produces an intermittent operational signal, a battery powersupply coupled to the constant current source circuit and the timercircuit. 13. The system of claim 11 further comprising an oscillator circuit coupled to the constant current source circuit, the oscillator circuit produces an alternating carrier wave for the operational signal,the sensor circuit is a current sensor circuit. 14. The system of claim13, the current sensor is disposed about the welding cable. 15. Thesystem of claim 11, the welder having a plurality of at least two operation controls for controlling corresponding operational parameters of the welder, the system further comprising a plurality of at least two constant current source circuits arranged in parallel in the remotedevice, each of the plurality of constant current source circuits electrically coupleable between the workpiece and the electrode holder and producing a unique operational signal on the welding cable when coupled between the workpiece and electrode holder; the sensor circuit producing correspondingly unique operational sense signals in response to detecting the unique operational signals on the welding cable; the operational control circuit coupled between the sensor circuit and the plurality of operation controls, the operational control circuitactuating one of the operation controls of the welder in response to detection of a corresponding one of the unique operational sense signals, whereby the operational parameters of the welder are selectively controllable by electrically coupling one of the plurality of constant current source circuits between the workpiece and theelectrode holder. 16. The system of claim 15, the plurality of operational signals produced on the welding cable by the plurality of constant current source circuits having different amplitudes. 17. Thesystem of claim 11 further comprising a pendant housing the remotedevice, first and second electrodes disposed on the pendant and coupled to the plurality of constant current source circuits, the first electrode couples the constant current source circuit to the workpiece,and the second electrode couples the constant current source circuit tothe electrode holder. 18. The system of claim 17, the first electrode isa magnetic electrode disposed on a surface of the pendant, the magnetic electrode magnetically and electrically coupleable to ferromagneticworkpieces. 19. A method for a remotely controllable welder having a welding cable and an electrode holder for supplying weld current to aworkpiece, a start switch for starting the welder and an indicator switch for at least temporarily enabling an indicator, the method comprising:producing a start signal on the welding cable by electrically coupling a start signal circuit between the workpiece and the electrodeholder; producing a start sense signal in response to detecting thestart signal with a sensor circuit; at least temporarily actuating thestart switch in response to detection of the start sense signal;actuating the indicator switch at least temporarily in response to thestart sense signal before actuating the start switch, whereby the welder is energized in response to electrically coupling the start signalcircuit between the workpiece and the electrode holder. 20. The method of claim 19, the welder having a coarse current increase control and a coarse current decrease control, the method further comprising:producing a coarse increase signal on the welding cable by electrically coupling a coarse increase signal circuit between the workpiece and the electrodeholder; producing a coarse increase sense signal in response to detecting the coarse increase signal with a sensor circuit; actuatingthe coarse current increase control in response to the coarse increase sense signal; producing a coarse decrease signal on the welding cable by electrically coupling a coarse decrease signal circuit between theworkpiece and the electrode holder; producing a coarse decrease sense signal in response to detecting the coarse decrease signal with a sensorcircuit; actuating the coarse current decrease control in response tothe coarse decrease sense signal, whereby coarse current isincrementally increasable and decreasable in response to electrically coupling a corresponding one of the coarse current increase circuit andthe coarse current decrease circuit between the workpiece and theelectrode holder. 21. The method of claim 19, the welder having a fine current increase control and a fine current decrease control, the method further comprising:producing a fine increase signal on the welding cable by electrically coupling a fine increase signal circuit between theworkpiece and the electrode holder; producing a fine increase sense signal in response to detecting the fine increase signal with a sensorcircuit; actuating the fine current increase control in response to the fine increase sense signal; producing a fine decrease signal on thewelding cable by electrically coupling a fine decrease signal circuit between the workpiece and the electrode holder; producing a fine decrease sense signal in response to detecting the fine decrease signal with a sensor circuit; actuating the fine current decrease control inresponse to the fine decrease sense signal, whereby fine current isincrementally increasable and decreasable in response to electrically coupling a corresponding one of the fine current increase circuit andthe fine current decrease circuit between the workpiece and theelectrode holder. 22. The method of claim 19, further comprising coupling the start signal circuit to the workpiece with a first electrode on a remote device, and coupling the start signal circuit tothe electrode holder with a second electrode on the remote device. 23.The method of claim 22 further comprising magnetically and electrically coupling the workpiece to the start signal circuit with a magnetic electrode. 24. The method of claim 19, producing the start sense signalin response to detecting the start signal with a current sensor circuit disposed along the welding cable. 25. The method of claim 19 further comprising producing the start signal on the welding cable by electrically coupling a constant current source circuit between theworkpiece and the electrode holder. 26. The method of claim 19 further comprising producing the start signal with an alternating carrier wave,and detecting the start signal with a current sensor circuit. 27. The method of claim 19 further comprising modulating the start signal to produce an intermittent start signal, and powering the start signalcircuit with a battery. 28. The method of claim 19, the welder is an engine-driven welder having run switch, the method further comprising atleast temporarily de-actuating the run switch with a run control circuit in response to a start sense signal produced on the welding cable whenthe start signal circuit is electrically coupled between the workpieceand the electrode holder after the welder is energized, whereby thewelder is de-energized. 29. The method of claim 19, the welder is an engine-driven welder having a choke switch for enabling a choke, the method further comprising:producing a choke signal on the welding, cable by electrically coupling a choke signal circuit between the workpieceand the electrode holder; producing a choke sense signal in response to detecting the choke signal with a sensor circuit; at least temporarilyactuating the choke switch in response to the choke sense signal,whereby the choke is enabled at least temporarily in response to electrically coupling the choke signal circuit between the workpiece andthe electrode holder. 30. A remotely controllable engine-driven welder having a welding cable and an electrode holder for supplying weld current to a workpiece, a start switch for starting the welder and an indicator switch for at least temporarily enabling an indicator, thesystem comprising:a remote device having a start signal circuit electrically coupleable between the workpiece and the electrode holder,the start signal circuit produces a start signal on the welding cable when coupled between the workpiece and electrode holder; a sensorcircuit disposed along the welding cable, the sensor circuit producing a start sense signal in response to detecting the start signal on thewelding cable; a start control circuit coupled between the sensorcircuit and the start switch, the start control circuit at least temporarily actuating the start switch in response to detection of thestart sense signal; an indicator control circuit coupled between the sensor circuit and the indicator switch, whereby the indicator control circuit actuates the indicator switch at least temporarily before thewelder is started. 31. The system of claim 30, the start signal circuit comprises a constant current source circuit. 32. A remotely controllable welder having a welding cable, an electrode holder for supplying weld current to a workpiece, and at least one operation control for controlling an operational parameter of the welder, the system comprising:a pendant housing a remote device having an operational parameter circuit; a magnetic first electrode disposed on the pendant; a second electrode disposed on the pendant; the first and second electrodes coupled to the operational parameter circuit; a sensorcircuit disposed along the welding cable and coupled to the operation control; the first electrode is magnetically and electricallyconnectable to the workpiece, whereby the electrode holder is electrically connectable to the second electrode to electrically couple the operational parameter circuit between the workpiece and theelectrode holder. 33. A method for a remotely controllable welder having a welding cable and an electrode holder for supplying weld current to aworkpiece, and an operation control for controlling an operational parameter of the welder, the method comprising:producing an operational signal on the welding cable by electrically coupling a constant current source circuit between the workpiece and the electrode holder; producing an operational sense signal with a sensor circuit in response to detecting the operational signal on the welding cable; actuating the operation control with an operational control circuit in response to detecting the operational sense signal, whereby the operational parameter of the welder is controlled remotely. 34. A remotely controllable welder for welding a workpiece, comprising:a welding cable having a welding electrode coupled thereto; a remote device; a constant current source circuit disposed in the remote device and electricallyconnectable between the workpiece and the welding electrode; a sensorcircuit disposed along the welding cable; the constant current source circuit produces an operational signal on the welding cable when coupled between the workpiece and welding electrode; the sensor circuit produce san operational sense signal in response to detecting the operational signal on the welding cable. 35. The welder of claim 34 further comprising a timer circuit coupled to the constant current source circuit, a battery coupled to the constant current source circuit andthe timer circuit, the timer circuit operates the constant current source circuit intermittently, whereby the operational signal produced thereby is intermittent. 36. The welder of claim 34 further comprising an oscillator circuit coupled to the constant current source circuit,the oscillator circuit produces an alternating carrier wave for the operational signal, the sensor circuit is a current sensor circuit. 37.The welder of claim 34,a plurality of constant current source circuits electrically connected in parallel and disposed in the remote device,each of the plurality of constant current source circuits electricallyconnectable between the workpiece and the welding electrode, the plurality of constant current source circuits each produce a unique operational signal on the welding cable when coupled between theworkpiece and welding electrode, the sensor circuit produces a unique operational sense signal in response to detecting a corresponding one ofthe unique operational signals on the welding cable. 38. The welder of claim 37, the welder having a plurality of operation controls for controlling corresponding operational parameters thereof, the welder further comprising:an operational control circuit coupled between the sensor circuit and the plurality of operation controls, the operational control circuit actuating one of the operation controls of the welder inresponse to generation of a corresponding one of the unique operational sense signals. 39. The welder of claim 37, the unique operational signals produced by the plurality of constant current source circuits have different amplitudes. 40. The welder of claim 37, a pendant housing the remote device, a common electrode disposed on the pendant and coupled to the plurality of constant current source circuits, a plurality of discrete electrodes disposed on the pendant, each of the plurality of discrete electrodes coupled to a corresponding one of the plurality of constant current source circuits, whereby the common electrode is connectable to the workpiece and the welding electrode iscontactable with one of the plurality of discrete electrodes to produce one of the unique operational signals on the welding cable. 41. The welder of claim 34, a pendant housing the remote device, a first electrode disposed on the pendant and coupled to the constant current source circuit, a second electrode disposed on the pendant and coupled to the constant current source circuit, whereby the first electrode isconnectable with the workpiece and the welding electrode is contactablewith the second electrode to produce the operational signal on thewelding cable. 42. The welder of claim 41, the first electrode is magnetic.
#!/usr/bin/env python # # Copyright (C) 2014 The MoKee OpenSource Project # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. # from xml.dom.minidom import parse, Document, Node import os, sys reload(sys) sys.setdefaultencoding('utf-8') class bcolors: HEADER = '\033[95m' OKBLUE = '\033[94m' OKGREEN = '\033[92m' WARNING = '\033[93m' FAIL = '\033[91m' ENDC = '\033[0m' def usage(): print '' print 'Usage: translate [language] [file name] [1]' print '' print 'Compares target XML (strings) with main XML in res/values' print 'and points out missing translations. If parameter 1 is given,' print 'generates new file' print '' print 'Note: Must be in res folder, output file will be in home directory' try: if len(sys.argv) == 1: usage() exit(1) if len(sys.argv) == 4: if sys.argv[3] != "1": print bcolors.FAIL+'Invalid parameters!'+bcolors.ENDC usage() exit(1) else: GENFILE = True elif len(sys.argv) == 3: GENFILE = False elif sys.argv[1] == 'help': usage() exit(1) FILE1 = os.path.abspath(os.path.join('values/',sys.argv[2])) FILE2 = os.path.abspath(os.path.join('values-'+sys.argv[1],sys.argv[2])) except IndexError: print bcolors.FAIL+'Insufficient parameters!'+bcolors.ENDC usage() exit(1) if not os.path.isfile(FILE1): print bcolors.FAIL+FILE1+' does not exist!'+bcolors.ENDC exit(1) file1 = parse(FILE1) doc = Document() NODE_NIL = [] NODE_LIST1 = [] if os.path.isfile(FILE2): NOFILE=False file2 = parse(FILE2) NODE_LIST2 = [] else: NOFILE=True def compareXML(): for node in file1.getElementsByTagName('string'): if not str(node.getAttribute('translatable')) == 'false': NODE_LIST1.append(str(node.getAttribute('name'))) for node in file2.getElementsByTagName('string'): if not str(node.getAttribute('translatable')) == 'false': NODE_LIST2.append(str(node.getAttribute('name'))) for i in NODE_LIST1: if i not in NODE_LIST2: NODE_NIL.append(i) def noCompare(): for node in file1.getElementsByTagName('string'): if not str(node.getAttribute('translatable')) == 'false': NODE_LIST1.append(str(node.getAttribute('name'))) NODE_NIL.append(NODE_LIST1) def genNode(parent, target): for i in parent.childNodes: if i.nodeType == Node.TEXT_NODE: nodeText = doc.createTextNode(i.nodeValue) target.appendChild(nodeText) else: newNode = doc.createElement(i.nodeName) j=i.attributes count=0 while count < j.length: k=j.item(count) newNode.setAttribute(k.name, k.value) count+=1 genNode(i, newNode) target.appendChild(newNode) def genNew(): if len(NODE_NIL) == 0: return if not os.path.exists('values-'+sys.argv[1]): os.makedirs('values-'+sys.argv[1]) docPath = os.path.join('values-'+sys.argv[1], sys.argv[2]) commentText = '\n\ Copyright (C) 2014 The MoKee OpenSource Project\n\n\ Licensed under the Apache License, Version 2.0 (the \"License\");\n\ you may not use this file except in compliance with the License.\n\ You may obtain a copy of the License at\n\n\ http://www.apache.org/licenses/LICENSE-2.0\n\n\ Unless required by applicable law or agreed to in writing, software\n\ distributed under the License is distributed on an \"AS IS\" BASIS,\n\ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n\ See the License for the specific language governing permissions and\n\ limitations under the License.\n' comment = doc.createComment(commentText) doc.appendChild(comment) root = doc.createElement('resources') root.setAttribute('xmlns:xliff', 'urn:oasis:names:tc:xliff:document:1.2') doc.appendChild(root) if not NOFILE: values = file2.getElementsByTagName('string') for node in NODE_LIST1: tempChild = doc.createElement('string') tempChild.setAttribute('name', node) root.appendChild(tempChild) if NOFILE or (node in NODE_NIL): nodeText = doc.createTextNode('') tempChild.appendChild(nodeText) else: for i in values: if str(i.getAttribute('name')) == node: parent = values[values.index(i)] break genNode(parent, tempChild) doc.writexml(open(docPath, 'w'), addindent=' ', newl='\n', encoding="UTF-8") doc.unlink() print '' print bcolors.HEADER+'New file created at:'+bcolors.ENDC print bcolors.OKGREEN+docPath+bcolors.ENDC def diffPrint(): print '' if NOFILE: print bcolors.OKBLUE+'No translations done before!'+bcolors.ENDC elif NODE_NIL == []: print bcolors.OKBLUE+'All translations done!'+bcolors.ENDC else: print bcolors.OKBLUE+'Strings without translation:'+bcolors.ENDC for i in NODE_NIL: print bcolors.OKGREEN+i+bcolors.ENDC def main(): if NOFILE: noCompare() else: compareXML() diffPrint() if GENFILE: genNew() if __name__ == "__main__": main()
<?php /* * This file is part of the Symfony package. * * (c) Fabien Potencier <[email protected]> * * For the full copyright and license information, please view the LICENSE * file that was distributed with this source code. */ namespace Symfony\Component\Serializer\Tests\Fixtures; /** * @author Jérôme Desjardin <[email protected]> */ class DeepObjectPopulateParentDummy { /** * @var DeepObjectPopulateChildDummy|null */ private $child; public function setChild(?DeepObjectPopulateChildDummy $child) { $this->child = $child; } public function getChild(): ?DeepObjectPopulateChildDummy { return $this->child; } }
'use strict'; var angular = require('angular'); angular .module('mwl.calendar') .controller('MwlCalendarDayReversedCtrl', function($scope, moment, calendarHelper, calendarEventTitle, $window) { var vm = this; vm.calendarEventTitle = calendarEventTitle; vm.leftInterval = setInterval(function() { }, 1000); vm.rightInterval = setInterval(function() { }, 1000); function refreshView() { vm.dayViewSplit = vm.dayViewSplit || 30; vm.dayViewHeight = calendarHelper.getDayViewHeight( vm.dayViewStart, vm.dayViewEnd, vm.dayViewSplit ); var view = calendarHelper.getDayView( vm.events, vm.viewDate, vm.dayViewStart, vm.dayViewEnd, vm.dayViewSplit, vm.dayViewEventWidth ); vm.allDayEvents = view.allDayEvents; vm.nonAllDayEvents = view.events; vm.viewWidth = view.width + 62; vm.events.forEach(function(event) { event.actualEnd = ''; event.left = (moment(event.startsAt).hours() * 60 + moment(event.startsAt).minutes()) / 60; }); } $scope.$on('calendar.refreshView', refreshView); $scope.$watchGroup([ 'vm.dayViewStart', 'vm.dayViewEnd', 'vm.dayViewSplit' ], refreshView); vm.eventDragComplete = function(event, columnChunksMoved, minuteChunksMoved) { var minutesDiff = Math.floor(minuteChunksMoved / vm.dayViewSplit / 3) * vm.dayViewSplit; if (typeof vm.columns !== 'undefined') { if (typeof event.column === 'undefined') { event.column = 0; } var newColumn = event.column + Math.round(columnChunksMoved / 2); if (newColumn < 0) { newColumn = 0; } else if (newColumn > vm.columns.length) { newColumn = vm.columns.length - 1; } } if (event.actualEnd !== '') { event.endsAt = event.actualEnd; } var newStart = moment(event.startsAt).add(minutesDiff, 'minutes'); var newEnd = moment(event.endsAt).add(minutesDiff, 'minutes'); var tempNewEnd = moment(event.endsAt).add(minutesDiff, 'minutes'); tempNewEnd.set('hour', vm.dayViewEnd.split(':')[0]); tempNewEnd.set('minute', vm.dayViewEnd.split(':')[1]); if (newEnd > tempNewEnd) { event.actualEnd = newEnd.toDate(); newEnd = tempNewEnd; } else { event.actualEnd = ''; } delete event.tempStartsAt; vm.onEventTimesChanged({ calendarEvent: event, calendarNewEventStart: newStart.toDate(), calendarNewEventEnd: event.endsAt ? newEnd.toDate() : null, calendarNewColumn: newColumn }); clearInterval(vm.leftInterval); clearInterval(vm.rightInterval); event.left = (newStart.hours() * 60 + newStart.minutes()) * 3; }; vm.eventDragged = function(event, columnChunksMoved, minuteChunksMoved) { var minutesDiff = Math.floor(minuteChunksMoved / vm.dayViewSplit / 3) * vm.dayViewSplit; event.tempStartsAt = moment(event.startsAt).add(minutesDiff, 'minutes').toDate(); var document = typeof $window.document === 'undefined' ? '' : $window.document; var posx = 0; var e = $window.event; if (e.pageX || e.pageY) { posx = e.pageX; } else if (e.clientX || e.clientY) { posx = e.clientX + document.body.scrollLeft + document.documentElement.scrollLeft; } clearInterval(vm.leftInterval); clearInterval(vm.rightInterval); var docWidth = parseInt($window.getComputedStyle(document.getElementById('calendar')).width) - 100; if (posx >= docWidth) { vm.leftInterval = setInterval( function() { document.getElementById('calendar').scrollLeft = document.getElementById('calendar').scrollLeft + 5; }, 20); } else if (posx <= 100) { vm.rightInterval = setInterval( function() { document.getElementById('calendar').scrollLeft = document.getElementById('calendar').scrollLeft - 5; }, 20); } }; vm.eventResizeComplete = function(event, edge, minuteChunksMoved) { var minutesDiff = minuteChunksMoved * vm.dayViewSplit; var start = moment(event.startsAt); var end = moment(event.endsAt); if (edge === 'start') { start.add(minutesDiff, 'minutes'); } else { end.add(minutesDiff, 'minutes'); } delete event.tempStartsAt; vm.onEventTimesChanged({ calendarEvent: event, calendarNewEventStart: start.toDate(), calendarNewEventEnd: end.toDate() }); }; vm.eventResized = function(event, edge, minuteChunksMoved) { var minutesDiff = minuteChunksMoved * vm.dayViewSplit; if (edge === 'start') { event.tempStartsAt = moment(event.startsAt).add(minutesDiff, 'minutes').toDate(); } }; }) .directive('mwlCalendarDayReversed', function() { return { template: '<div mwl-dynamic-directive-template name="calendarDayViewReversed" overrides="vm.customTemplateUrls"></div>', restrict: 'E', require: '^mwlCalendar', scope: { events: '=', columns: '=', viewDate: '=', onEventClick: '=', onEventTimesChanged: '=', onTimespanClick: '=', onDateRangeSelect: '=', dayViewStart: '=', dayViewEnd: '=', dayViewSplit: '=', dayViewEventChunkSize: '=', dayViewEventWidth: '=', customTemplateUrls: '=?', cellModifier: '=', templateScope: '=' }, controller: 'MwlCalendarDayReversedCtrl as vm', bindToController: true }; });
#include "PCHEADER.h" bool GS_HML::LoadObject(const std::string &filename,Sprite &arg,ObjectFileArgs &OFA) { //Begin File checks before saving //--------------------------------------------------- fstream stream; //Open a stream //Check to see if the specified file already exists stream.open(filename.c_str(),ios::in|ios::binary|ios::_Nocreate); if(stream.is_open()) { //First Load the Arguments for the file type. (These Are set when the Object is Saved.) //------------------------------------------------ stream.read((char *) &OFA.hasTexture,sizeof(bool)); stream.read((char *) &OFA.isAnimate,sizeof(bool)); stream.read((char *) &OFA.isChangable,sizeof(bool)); stream.read((char *) &OFA.useDialog,sizeof(bool)); stream.read((char *) &OFA.useScripts,sizeof(bool)); stream.read((char *) &OFA.Scene_ID,sizeof(int)); //Begin Loading the various data for the given Sprite Object //-------------------------------------------------------------- if(OFA.hasTexture) { if(!arg.Texture.LoadTextureData(stream)) { stream.flush(); stream.close(); return false; } } if(!arg.pAudioSource.LoadAudioSourceData(stream)) { stream.flush(); stream.close(); return false; } if(OFA.isAnimate) { if(!arg.Animations->LoadAnimData(stream)) { stream.flush(); stream.close(); return false; } } if(!arg.LoadSpriteData(stream)) { stream.flush(); stream.close(); return false; } //Flush and Close our File //------------------------------------------ stream.flush(); stream.close(); return true; } return false; } ////////////////////////////////////////////////////////////////////////////////////////// bool GS_HML::SaveObject(const std::string &filename,Sprite arg, ObjectFileArgs &OFA) { //Begin File checks before saving //--------------------------------------------------- fstream stream; //Open a stream //Setup our file structure variables to signal whether certain things should be saved for that file OFA.hasTexture = arg.GetHasTexture(); if(arg.GetNumOfAnimations()>0) { OFA.isAnimate = true; } //Check to see if the specified file already exists stream.open(filename.c_str(),ios::in|ios::binary); if(stream.is_open()) { //If it does prompt the user to save over the file if(MessageBox(NULL,"The file already exist. Would you like to save anyways?", "Attention : File Already Exists!",MB_YESNO|MB_ICONQUESTION)==IDNO) { stream.close(); return false; } stream.close(); } //If the file didn't exist or the user decided to save over the current file, //Open our file for saving stream.open(filename.c_str(),ios::out|ios::binary|ios::beg); if(stream.is_open()) { //First Save the Arguments for the file type. (These Are set when the SaveObject Dialog GUI is called from the Editor) //------------------------------------------------ stream.write((char *) &OFA.hasTexture,sizeof(bool)); stream.write((char *) &OFA.isAnimate,sizeof(bool)); stream.write((char *) &OFA.isChangable,sizeof(bool)); stream.write((char *) &OFA.useDialog,sizeof(bool)); stream.write((char *) &OFA.useScripts,sizeof(bool)); stream.write((char *) &OFA.Scene_ID,sizeof(int)); //Begin Saving the various data for the given Sprite Object //-------------------------------------------------------------- if(OFA.hasTexture) { if(!arg.Texture.SaveTextureData(stream)) { stream.flush(); stream.close(); return false; } } if(!arg.pAudioSource.SaveAudioSourceData(stream)) { stream.flush(); stream.close(); return false; } if(OFA.isAnimate) { if(!arg.Animations->SaveAnimData(stream)) { stream.flush(); stream.close(); return false; } } if(!arg.SaveSpriteData(stream)) { stream.flush(); stream.close(); return false; } //Flush and Close our File //------------------------------------------ stream.flush(); stream.close(); return true; } return false; } ////////////////////////////////////////////////////////////////////////////////////////// GS_HML::GS_HML() { ID_REF_COUNT = 0; ActiveState_ID = 0; PreviousState_ID = 0; } ////////////////////////////////////////////////////////////////////////////////// GS_HML::~GS_HML() { } ////////////////////////////////////////////////////////////////////////////////// bool GS_HML::ADDSTATE(GAMESTATE * const State,bool Init) { //Make sure that there are not too many active states. Limit to MAX_NUM_OF_STATES. if(ID_REF_COUNT >= MAX_NUM_OF_STATES) { MessageBox(NULL,"There are too many Active Game States. You must \n" " Remove one from memory before adding a new one. Add Game state aborted.", "GAMESTATE OVERFLOW!",MB_OK|MB_ICONERROR); return false; } else { ID_REF_COUNT += 1; //ADD 1 to our reference count. } int b = 0; //Set the counter variable to one to access first scene cache location. //Check to see if the current location is empty. do { if(StateCache[b].Get_StateName() == "") { //Set the Active Level State Information StateCache[b] = *State; //Store the values of the State reference into our empty location StateCache[b].Set_State_ID(ID_REF_COUNT); //Set the states ID for referencing if(Init) //Should we set the state as the lead one? { StateCache[b].Set_isStarting(true); StateCache[b].Set_hasFocus(true); StateCache[b].Set_isPaused(false); break; } else //Otherwise don't { StateCache[b].Set_isStarting(false); StateCache[b].Set_hasFocus(false); break; } } else { b+=1; if(b>=(MAX_NUM_OF_STATES)) { //ADD A MESSAGE OF STATE COULD NOT BE ADDED TO //BE DISPLAYED USING CUSTOM MESSAGE FUNCTION //ONCE GUI IS FINISHED. MessageBox(NULL,"The State could not be added. Current States being managed are full.", " State Addition Error",MB_OK||MB_ICONERROR); return false; } } } while(b<=(MAX_NUM_OF_STATES - 1)); return true; } ///////////////////////////////////////////////////////////////////////////////////// bool GS_HML::REMOVESTATE(int STATE_ID) { // remove state. StateCache[STATE_ID].Remove(); ID_REF_COUNT --; return true; } ///////////////////////////////////////////////////////////////////////////////////// bool GS_HML::ResolveMessage(int STATE_ID, int MESSAGE) { if(STATE_ID < 0 || STATE_ID > ID_REF_COUNT) //The Id isn't valid. { MessageBox(NULL, "The STATE_ID is not valid. Please make sure it \n" " is in the appropriate range. The program will continue but \n" " the attempted state command could not be carried out.", " GS_HML::ResolveMessage() Error!",MB_OK||MB_ICONEXCLAMATION); return false; } //ID is valid, make sure message is valid. if(MESSAGE < 0 || MESSAGE > 9) { MessageBox(NULL, "The Message is not valid. Please make sure it \n" " is in the appropriate range. The program will continue but \n" " the attempted state command could not be carried out.", " GS_HML::ResolveMessage() Error!",MB_OK||MB_ICONEXCLAMATION); return false; } else { HML_DispatchMessage(STATE_ID,MESSAGE); return true; } return false; } ///////////////////////////////////////////////////////////////////////////////////// bool GS_HML::ResolveMessage(std::string StateName, std::string MESSAGE) { int g=0; int StateID=0; int MessageID=0; //Go through all the current game states and find the one with the qaulifing state name do { if(StateCache[g].Get_StateName() == StateName) { StateID = g; break; } if(g > 26) { MessageBox(NULL,"The specified state could not be found. \n" " Make sure the spelling of the state name and the message are correct. \n" " The program will continue but no action will be performed by the state.", "GL_HML::ResloveMessage() Error",MB_OK||MB_ICONEXCLAMATION); return false; break; } } while(StateID==0); //Find the corrisponding Message transform(MESSAGE.begin(),MESSAGE.end(),MESSAGE.begin(), (int(*)(int)) toupper); if(MESSAGE=="INITIALIZE") { HML_DispatchMessage(StateID,INITIALIZE_STATE); return true; } else if(MESSAGE=="REMOVE") { HML_DispatchMessage(StateID,REMOVE_STATE); return true; } else if(MESSAGE=="SETFOCUS") { HML_DispatchMessage(StateID,STATE_HASFOCUS); return true; } else if(MESSAGE=="REMOVEFOCUS") { HML_DispatchMessage(StateID,STATE_REMOVEFOCUS); return true; } else if(MESSAGE=="PAUSE") { HML_DispatchMessage(StateID,PAUSE_STATE); return true; } else if(MESSAGE=="UNPAUSE") { HML_DispatchMessage(StateID,UN_PAUSE_STATE); return true; } return false; } ///////////////////////////////////////////////////////////////////////////////////// int GS_HML::HML_DispatchMessage(int STATE_ID, int MESSAGE) { if(MESSAGE==REMOVE_STATE) { REMOVESTATE(STATE_ID); } else if(StateCache[STATE_ID].UpdateMessages(MESSAGE)) { return 1; } return 0; } ////////////////////////////////////////////////////////////////////////////////////// bool GS_HML::Push_State(int STATE_ID) { StateCache[STATE_ID].Set_hasFocus(false); return false; } ////////////////////////////////////////////////////////////////////////////////////// bool GS_HML::Pop_State(int STATE_ID) { StateCache[STATE_ID].Set_hasFocus(true); return true; }
Pixel Mechanic Voice Lines * "You didn't pay attention! Now you're suffering the dire consequences." * "Come to my halls full of music. Escape from this empty tune." * “Allow me to play my song. It's a shame you won't be around for long." * "Enjoy your stay before I rip the life out of your eyes." Trivia * Pixell's mechanic is similar to the Puppet from Five Nights at Freddy’s 2.
Appeal of Brush Electric Company et al. In an action to compel defendant to specifically perform a contract for the purchase of electric apparatus and supplies, and for damages suffered by the plaintiff on account of defendant’s refusal to buy such goods from the plaintiff as required by the contract, the basis on which the master found the loss of profits and damages, and his calculations, sustained. Note. — Contracts relating to things personal are ordinarily not enforced, unless it appears that there can be no adequate compensation in damages. In such case equity will enforce the agreement. Vail v. Osburn, 174 Pa. 580, 34 Atl. 315; Bald Eagle Valley R. Co. v. Nittany Valley R. Co. 171 Pa. 284, 29 L. R. A. 423, 50 Am. St. Rep. 807, 33 Atl. 239; Cumberland Valley R. Co. v. Gettysburg & H. R. Co. 177 Pa. 519, 35 Atl. 952; Cornwall & L. R. Co.’s Appeal, 125 Pa. 232, 17 Atl. 427. For authorities as to loss of profits as an element of damages, see the following editorial notes: Loss of profits as an element of damages for breach of contract, note to Wells v. National Life Asso. 53 L. R. A. 33; loss of profits of sale or purchase as damages, note to Guetzkow Bros. Co. v. Andrews, 52 L. R. A. 209. (Argued November 4, 1887. Decided November 11, 1887.) October Term, 1887, No. 222, W. D., before Gordon, Oh. J., Paxson, Steeeett, and Williams, JJ. Certiorari to the Court of Common Pleas No. 2 of Allegheny County to review a decree fixing the amount of loss and damage to be paid by the Allegheny County Light Company, defendant, to the Brush Electric Company and John E. Eidall, plaintiffs, in a suit in equity. Affirmed. The bill alleged, inter alidj that the plaintiffs, the Brush Electric Company (which was a corporation of the state of Ohio, engaged in making and selling electric light apparatus under patents controlled by it) and Eidall & Ingold, its exclusive sales agents in Pittsburgh and Allegheny, had granted to the defendant, the Allegheny County Light Company, the exclusive right of leasing or furnishing for use to others lights and electric lamps of the Brush apparatus, with all accessories appertaining thereto, and to connect such lamps with and operate them by electric generating apparatus furnished by the Brush company and its said agents, the same to be sold to defendant at regular card rates. That thereafter and on March 31, 1883, the parties entered into the following contract, supplementary to the original con-contract between them: “And whereas, Said Allegheny County Light Company has been using said Brush apparatus, purchasing the same together with supplies from said Eidall & Ingold, and it is desired that said light company shall continue said use, “Therefore, It is now agreed by and between the parties that on all future purchases of said Brush apparatus or supplies made by said light company from said Kidall & Ingold, or their assignees, there shall be allowed to said light company a discount of 10 per cent from the list price, or such higher rate of discount as may be from time to time allowed to light companies purchasing from agents of said Brush Electric Company. This discount shall be allowed on apparatus and supplies, and shall not include storage batteries or incandescent lamps. “In consideration thereof, it is agreed on the part of the Allegheny County Light Company, that said company shall not purchase, rent, sell, or use, any other electric arc light apparatus or carbons, except those provided by said Brush Electric Company. “In witness hereof, the said Kidall & Ingold have hereto set their hands and seals, and the president of said light company has hereto set his hand and affixed the seal of said company, the day and year aforesaid.” That the discount of 10 per cent named in said contract of March 31, 1883, was, as well known to the defendant, to come off from the commissions said agents, Kidall & Ingold, were entitled to receive from the said Brush Electric Company, and an important and the chief consideration, to the said agents and their said principal, was the special agreement of said defendant not to purchase, rent, sell, or use any otMr electric arc light, apparatus, or carbons, excepting those provided by the said Brush Electric Company, especially as the carbons therein mentioned were, in the use of the electric light, the great article of consumption and of consequent steady, continuous, and large supply. That after the execution of said contract of March 31, 1883, the said Ingold, of said film of Kidall & Ingold, assigned his interest in said firm, including his interests in said contracts, to his said partner John E. Kidall, a plaintiff herein. That each of the said parties, interested in said contracts, continued their business together, in relation to the subject-matter thereof in full and faithful observance and compliance with all the obligations and terms of said agreements, until some time in the summer or fall of the year 1884, when the defendant entered upon, and has ever since continued, and declares its purpose of continuing in, a course of direct violation and disregard of the terms of said agreements in this, “that since that time it has purchased all its carbons used for and in its business, comprising large amounts and frequent purchases, from others, and wholly of those made and provided and sold by others, than the-said Brush Electric Company or its said agent or agents; such others being competitors of said Brush Electric Company, in that part of its business, from whom said agreements were expressly designed and intended to protect it; and this defendant has so done, declaring and still declares that it intended and intends to purchase and use, and to continue purchasing and using such carbons from others than your said orators, notwithstanding the express terms, as aforesaid, of its agreement and obligation contained in said contract of March 31, 1883. And the said defendant has been and is engaged in constructing and making apparatus, machinery, or accessories, or parts thereof, covered by said contract of 1881, as above set forth, excluding' the defendant from such right, and it is particularly engaged in making commutators, which is a definite part of said machinery and apparatus, to the especial injury of said Brush Electric Company in its business, which is an important part thereof, of making and manufacturing all apparatus, machinery, accessories, and supplies (including carbons, etc.), in this bill and said contracts referred to.” The prayer of the bill asked: (1) That the contract referred to be declared binding upon the parties; (2) that the defendant specifically perform said contracts; (3) that defendant be enjoined from continuing to construct or malee commutators, apparatus, machinery, or accessories, or any parts thereof which are furnished by the Brush company; (4) that defendant be enjoined from purchasing or using any other electric arc light, apparatus, or carbons, except those provided by the Brush company; (5) that defendant make discovery of all carbons, arc lights, or apparatus bought from others than the plaintiffs, and of all commutators, apparatus, machinery,- or accessories, made by it or by others than the Brush company; (6) that an account be had, showing the profits or commissions of which plaintiffs have been deprived, and of the damages and losses they have sustained by the conduct of defendant, and that the same be paid to plaintiffs as their respective interests appear; and (7) general relief. The defendant answered, denying the equities of the bill. After evidence had been given before a master the bill was dismissed, but on appeal to the supreme court the decree of dismissal was reversed and the court below directed to proceed to a final decree. (See 114 Pa. 574, 7 Atl. 794.) Thereupon, on March 5, 1887, a decree was entered directing the defendant to “observe and comply with and specifically perform all of the terms of the contracts recited and referred to in the plaintiff’s bill filed in this suit, and by the said defendant, to be thereby observed, complied with and performed,” and enjoining the defendant “from purchasing or using any other carbons excepting those provided by the said Brush Electric-Company, and from otherwise violating tire terms of said agreement.” The case was referred back to the master, Thomas Harriott,. Esq., to state an account between the parties, who reported substantially as follows: The only question now before the master is the damage suffered by the plaintiffs under the contract of March 31, 1883, on account of defendant refusing to buy “carbons” from the plaintiffs as therein provided. After the execution of this contract the defendant bought from • other parties than the plaintiffs 458,359 carbons; of these 153,500 were one half inch carbons; and the balance, 294,859, -were seven sixteenths inch carbons. These’ purchases continued from December, 1884, to March,. 1887. The agreement of March, 1883, provides that the defendant shall buy all its carbons from the Brush company at the list price of this company, less a discount of 10 per cent or such higher rate of discount as may be given to other light companies buying from the agents of the Brush company. The master is not able to find from the testimony just wha.t the market price of carbons was. It is true that we have in evidence the price paid by the defendant for all the carbons bought from others than the Brush company; but there is no evidence to show that they were equal in quality to the Brush carbons, and the master does not think that the price' paid by the defendant was the market price. Mr. Duncan, the general manager, testifies that while the various manufacturers had a price list, that was nearly uniform, yet the salesmen did not regard this; and while the dealers all sent out price lists, he never found an agent that stuck to the price list. The testimony also shows that the defendant bought carbons as low as $8 per 1,000, while the cheapest carbon made by tire Brush company cost $9.40 per 1,000 to make it, and a very large proportion of the carbons from others than the Brush company were bought at a price much less than cost price for the manufacture of the Brush carbons. The master, therefore, cannot find that the prices in evidence of the purchases made as above are the market prices, and holds that the list prices of the Brush company are to govern in assessing the damages in this case. Notwithstanding the master is of the opinion that the price list of the Brush company shall govern in assessing the damages in this case, yet it is very difficult to arrive at the damage done. From December 1, 1884, to May 10, 1886, the list price of the Brush company for carbons was $34.38 for %-inch, and $27.50 for 7/16-inch carbons; and since May 10, 1886, the list price has been $23.75 for %-inch, and $21.25 for seven-sixteenths carbons. The cost of making these carbons was, from December 1, 1884, to June 1, 1885, $12.75 for seven-sixteenths inch per 1,000; and from June 1, 1885, to January 1, 1886, for same size, $10.85 per 1,000; and from January 1, 1886, to present time, for same size, $9.40 per 1,000. The ^-inah. carbons cost to make $2 per 1,000 more than the prices above during the period specified. If this were the only element the computation of the damage would be easy; it seems, however, that in making carbons there are from 10 to 20 per cent of the lumber put into the oven that come out slightly imperfect but still fit for use; these imperfect-carbons are called “seconds.” Up to the first of July, 1885, the Brush company did not sell these “seconds,” but ground them up and worked the material over again. These “seconds” are in. every way made the same and of the same material as the “firsts,” or perfect carbons, the only difference in the two being that a short piece at the end of the “seconds” may be burned too hard or may not be quite straight. After July 1, 1885, the Brush company commenced to sell these “seconds” at a price slightly above the cost of making the “firsts,” but never made a list price for “seconds.” It continued to sell these “seconds,” at from $11 to $14 per 1,000, until a short time after the interlocutory decree of March 5, 1887. From December, 1884, until the Brush company stopped selling “seconds,” after the said decree, the only carbons bought by the defendant from the Brush company were “seconds.” Seventeen thousand were thus bought during July and August, 1885, and the evidence does not show the number bought after the •decree of March 5, 1887, before the Brush company stopped selling “seconds.” The plaintiffs claim that the question of “seconds” should not enter into this case, as they were not sold •or considered in March, 1883, when the agreement was signed ■and there was no list price for “seconds,” while the defendant ■claims that it would have bought nothing but “seconds” during the period for which an account is ordered, and that the damage should be based on the profit on “seconds.” The testimony shows that the “seconds” bought by the defendant from the plaintiffs answered the purpose as well as “firsts,” except that they required a little more care in setting. The master thinks, therefore, that if the plaintiffs could have supplied the defendant with all the carbons it needed of the grado of “seconds,” the defendant had a right to order these after July 1, 18S5. The testimony is not full on this point; the plaintiffs show that “seconds” are an accidental product, varying from 10 to 20 per cent of the carbons put into the kiln, and that the demand for them, during the time they were sold was such that they could not accumulate any stock; they do not show, however, that at any time they were unable to fill any order for “seconds.” The master thinks that if he is right in ruling that the defendant can order “seconds,” the plaintiffs were bound to furnish this grade, then it is fair, under the evidence, to presume that the •supply would be sufficient, unless the plaintiffs show that it would not. ■ The master, therefore, thinks that damages should be assessed to July 1, 1885, according to the list price of the Brush company for “firsts,” and since that time according to the profits on “seconds.” Just what the profits are on “seconds” is not clear; the plaintiffs’ witness claims the difference between the value of “seconds,” at the price for which they are sold, and their value as material to grind over, but does not give the data from which this could be accurately ascertained; the •defendant claims that the profit, on “seconds” is the difference between the cost of “firsts” and the selling price of “seconds,” •claiming that it costs just the same to make “seconds” as “firsts.” This is the correct view of the cost price as given by plaintiff, based on both “firsts” and “seconds.” As the cost of making carbons decreased $1.90 per 1,000 at about the time the company commenced selling “seconds,” the master concludes that this decrease was probably due to the increased value of “seconds” as a marketable product over their value as material. The selling price of “seconds” varied from $11 to $14 per 1,000; but as the plaintiffs have not furnished exact data as to the number sold at each price, the master assumes that the lowest price was $11 per 1,000 for seven-sixteenths inch, and the highest price $14 per 1,000 for half inch; and as the difference in their cost is $2 per 1,000, tire variation in price would be $3. As the master has no means of knowing whether the number sold at the highest price was many or few, he here again rules that the burden of showing this is on the plaintiffs, and considers all the seven-sixteenths inch sold at $11, and all one half inch sold at $13 per 1,000. This, supposing the cost of making the half inch to be $2 per 1,000 more than the seven-sixteenths inch, would make the profit uniform per 1,000 on “seconds,” and damages are assessed on the number of carbons purchased by the defendant, as follows-: Carbons purchased. %inch. 7-16 inch. Profit per M. From Dec. 1, 1884, to June 1,1885 27,960 — $16.20 $452.95 “ “ “ “ — 71,899 12.00 862.78 Average interest from Mar. 1, ’85 — — 184.20 From June 1, ’85, to July 1, ’85.. 6,460 — 18.10 116.92 “ “ “ “ — 10,540 13.90 146.50 Interest since July 1, 1885...... — — 31.60 From July 1, ’85, to Jan. 1, ’86, seconds.................... 91,000 both sizes. 0.15 13.65 Average interest from Oct. 1, ’85 — —• — 1.43 From Jan. 1, 1886, to Meh. 5, 1887 250,500 both sizes. 1.60 400.S0 Average interest from Aug. 1, ’86 — — — 22.00' $2,232.83 Of this -sum one ninth is due to J. B. Ridall, $ 268.09 And the balance to the Brush Company, im.: 1,904.74 $2,232.83 The plaintiffs filed the following exceptions to the report: 1. That he does not find on the basis that the loss of profits to the Brush Electric Company, on the carbons bought of others, was the difference between the list prices of the Brush Electric Company and the cost, as proved and set out in the report, less the 10 per cent discount. 2. The master errs, in finding and basing his calculations upon the notion that the plaintiffs would have sold, and the defendant would have purchased, the carbons called “seconds.” 3. The master errs, in basing his calculation upon the purchase of “seconds,” especially as the evidence clearly indicates (there being nothing to the contrary) that of all the carbons, purchased of others, none of them were “seconds.” 4. The master errs in his calculation and estimates, and in his methods of ascertaining and allowing of profits and commissions on said “seconds.” 5. The master errs, in not taking such testimony as there was, to wit, that of Q-. W. Stokeley, as' to the cost and value of “seconds” to the Brush company, and thereby placing the cost and value to the said company at not exceeding one half of-the price at which they were sold. ■ 6.' The master erred, in not finding upon the basis of the contract, and in undertaking to find respecting an article not contemplated in the contract, and being an accidental product sold exceptionally and without any market or list price. The exceptions were overruled, and the report confirmed, and a final decree entered, making the above decree of March 5, 1887, final and perpetual, and directing the defendant to pay to the plaintiffs $2,232.83, loss and damages, with costs. ' From this decree the plaintiffs appealed, assigning as error the action of the court in not affirming the plaintiffs’ exceptions to the master’s report. D. D. Bruce and M. A. Woodward, for appellants. There are two main propositions contained in our assignments of error: First, that the master erred in not treating the ordinary and usual product of manufacture, and sale of carbons — such as'this contract of March 31, 1883, necessarily referred to, and of which alone we'evér had a “list price,” as the articles upon which our loss was to be based and estimated. (See errors, first, second, third, and sixth.) Second, that if the said departure from the contract, as adopted by the master, was proper, that then he ignored the best and only reliable evidence as to cost or value to us of “seconds” carbons, and gave us but a fraction of the loss which we were entitled to upon that basis. (See errors, fourth and fifth.) Upon the first question we ask'to stand upon our contract rights, and that the contract be construed in respect to the circumstances surrounding it when it was made. The imperfect carbons, called “seconds,” were then, as at all times, produced by us, but only as a part of the refuse of our manufacture. Others sold them, but we did not. With a fixed customer to whom we could always sell “firsts” at a larger profit,, it is equally unreasonable to assume that we would have sold “seconds.” And the best evidence that the defendant would not have bought them is the fact that it bought none from others, never buying any but the three lots bought from us, in July, 1885, during its default to us, and when we had first commenced the experiment, since abandoned, of selling them, in place of using them as raw material. To this error our third exception is directed. By no reasonable argument, therefore, can the master be sustained in admitting the new, uncertain, and complicated subject of “seconds” under our contract On the second question, the testimony shows that, if this basis is to govern, fully one half of the price at which the master finds “seconds” to have been sold, would be profit. At his basis of the average selling price of the 91,000 and 250,500 supposed “seconds” carbons, the profit at one half would have been $2,049.00, and the average interest to July, 1887, $112.69, amounting to $1,723.81 more than the amount allowed by the master. If we are right on the first proposition, then our profits are of easy calculation, and amount to $6,229.95, with an average interest to July 1, 1886, to be added of $466.17, or $4,253.83 in addition to the amount allowed by the master. W. B. Rodgers, for appellee. Before the master we contended that carbons had a market price, there being a large number of makers, and therefore the measure of damages was the difference between the cost to the Brush company and the market price. It was on this theory that a calculation was produced before the master, some items of which showed that the market price was less than the cost price as claimed by the Brush company. But another position taken by us he sustained, which was this: July 1, 1885, the Brush company began to put on the market two grades of carbons, firsts and seconds. Ridall began to sell these seconds, and so continued until some time in April, 1887, Avhen the Brush company ceased to furnish them. The only purchases made by us of Brush carbons between December, 1884, and March 5, 1887, Avere of these seconds. These purchases were two orders in July, 1885, and two in August. Then came the interlocutory decree of March 5, 1887, requiring us to buy none but the Brush carbons sold by Bidall. All our purchases from Bidall from that date, until the Brush company ceased to furnish them, a period of about two months, were of “seconds.” The master took our view and charged us with the profit of the Brush company based on “firsts” from December, 1884, to July, 1885, and on “seconds” from July 1, 1885, to the date of interlocutory decree. I come now to the argument on the other side. There are two propositions: First, the contract of March 31, 1883, confined the defendant to the purchase of firsts because they were the usual and ordinary product, and of Avhich alone there ever was a list price, and that the damages should have been fixed on that basis. Second, that if seconds Avers to be taken as the basis, then the master ignored the best evidence as to their cost and value. As to the first proposition, the defendant was not limited to, or bound to buy, “firsts” by the contract of March 31, 1883. The agreement in that paper on the part of the defendant is that it shall not purchase any other “carbons except those provided by the Brush company.” “Seconds” were provided by the Brush company; therefore we had the right to purchase them. But it is said the Brush company never had a list price of seconds; but it cannot be seriously contended that if the Brush company put one or more additional classes of carbons on the market, it could prevent us from buying them by refusing to make a list price. The term list price means the price at which the Brush company sells its goods. Besides that, the term as used in the contract refers merely to the basis upon which the discount to be allowed by Bidall is to be calculated, and nothing more. The second proposition, for the reasons already given, is radically unsound, either as a mathematical, business, or legal proposition. As a mathematical proposition, it ignores one of the main elements, the cost of the workmanship in the seconds. As a business proposition, it claims a false rule for profits; and as •a legal proposition, it asks the court to assume the value of the materials alone, but gives no certain data on which that value ■could be fixed, although that data was in the possession of the plaintiffs alone. The “best and the only reliable evidence” on this subject was that it cost as much to make a second as a first, .as the material and the labor were just the same. Per Curiam: An examination of this case fails to convince us that the de■cree of the court below was in any particular wrong-; hence, we refuse to sustain the appellants’ exception. The appeal is dismissed and the decree affirmed, at the costs ■of appellants.
How do I prove the following statement about the kernel? I have the following problem: Given a finite group $G$ and $p$ the smallest prime dividing $card(G)$. Let $H$ be a subgroup s.t. $card(G\setminus H)=p$ Let $X=G\setminus H$ and consider the action $$G\times X\rightarrow X; \,\,(g,xH)\mapsto gxH$$ Let $$\rho:G\rightarrow Bij(X,X);\,\,g\mapsto \rho(g):X\rightarrow X\,\,\text{given by}\,\,\rho(g)(xH)=gxH$$the associated morpism. I have just shown that $Stab_G(xH)=xHx^{-1}$ is the stabilizer. Now I need to deduce that $ker(\rho)=\bigcap_{x\in G} xHx^{-1}$. I'm somehow unsure since they wrote deduce I think we should use that $Stab_G(xH)=xHx^{-1}$. I wanted to take $g\in ker(\rho)$ and then show that for all $x\in G$ $gxH=xH$ But i somehow struggle. Could someone help me please? I'd say your definition of $\rho(g)$ is a bit off. Is $x\in G$? Or $x\in G/H$? Maybe $\rho(g)(xH) = gxH$ Sorry I edited it, now it should be okei right? Could you still help me? $\ker(\rho) \subset \bigcap xH x^{-1}$: If $g\in\ker(\rho)$ then $\rho(g) = id_{X}$ and we have $gxH = xH$ for all $x \in G$. Thus $g\in\text{Stab}_G(xH)=xHx^{-1}$ for every $x\in G$. $\ker(\rho)\supset \bigcap xH x^{-1}$: Let $g\in xHx^{-1}$ for every $x\in G$. Then let $y\in G$ and consider $\rho(g)(yH)= gyH$. Since $g\in yHy^{-1}$, we have an $h\in H$ such that $g=yhy^{-1}$ and can compute $$gyH = yhy^{-1}yH= yhH= yH.$$ So $\rho(g)(yH)=yH$. But $y$ was arbitrary hence $\rho(g)=Id_{X}$. ah wow perfect thanks but I think in the second direction it should be for every $x\in G$ Yes thank you, edited.
We determine the structure of two variations on the Temperley-Lieb algebra, both used for dealing with special kinds of boundary conditions in statistical mechanics models. The first is a new algebra, the `blob' algebra (the reason for the name will become obvious shortly!). We determine both the generic and all the exceptional structures for this two parameter algebra. The second is the periodic Temperley-Lieb algebra. The generic structure and part of the exceptional structure of this algebra have already been studied. Here we complete the analysis, using results from the study of the blob algebra.
import weightService from "./weight" test('lbToKgConversion returns correct value', () => { expect(weightService.lbToKgConversion(1)).toBe(0.45359237) }) test('lbToKgConversion returns correct value', () => { expect(weightService.lbToKgConversion(2)).toBe(2*0.45359237) }) test('kgToLbConversion returns correct value', () => { expect(weightService.kgToLbConversion(1)).toBe(2.20462262) }) test('kgToLbConversion returns correct value', () => { expect(weightService.kgToLbConversion(2)).toBe(2*2.20462262) }) test('rounded rounds down correctly', () => { expect(weightService.rounded(2.501)).toBe(2.5) }) test('rounded rounds up correctly', () => { expect(weightService.rounded(2.49)).toBe(2.5) }) test('rounded doesnt round unnecessarily', () => { expect(weightService.rounded(2.5)).toBe(2.5) })
Why is the plot of residuals against fitted values a horinzontal line when the dependent variable is linearly related to the indenpendent variable? In ordinary least squares regression (OLS), if the plot of the residuals against the fitted values form a horizontal line around 0, then we can say that the dependent variable is linearly related to the independent variable. I had thought that this is true because $E(y_i - \hat{y}_I)=0$ when the dependent variable is linearly related to the independent variable, see here. However, suppose: $y_i = \alpha + \sin(x_i) + \epsilon_i$. Then $E(y_i - \hat{y}_i)$ is still 0, see here but then the plot of its residuals against its fitted value is no longer a horizontal line around 0, as this R code shows: n <- 10^3 df <- data.frame(x=runif(n, 1, 10)) df$mean.y.given.x <- sin(df$x) df$y <- df$mean.y.given.x + rnorm(n) model <- lm(y ~ x, data=df) plot(predict(model, newdata=df), residuals(model)) abline(a=0,b=0,col='blue') So my question is, which assumption(s) of OLS that causes the plot of the residuals and the fitted value to be a horizontal line around 0 and why/how is it true? You've regressed $y$ on $x$, not on $\sin x$. @Scortchi Are you referring to the R code or the expression $E(y_i-\hat{y_i})=0$? Could you please show me how to regress it on $\sin x$? In the statement model <- lm(y ~ x, data=df) substitute sin(x) for x The R code - though I'd overlooked that the model equation you gave doesn't in fact include an coefficient for $\sin x$, as explained by @Glen_b. In any case, a correctly specified model is a necessary assumption for the residuals to behave as expected with OLS, & the point of plotting residuals vs fits is to look for signs of mis-specification. Scortchi and Peter Flom have both correctly pointed out that you didn't fit the model you specified. However, there's no coefficient on $\sin(x)$ in the model, so if you actually want to fit $y_i = \alpha + \sin(x_i) + \epsilon_i$ you should not regress on $\sin(x)$. In that model it's an offset, not a regressor. The correct way to specify the model $$y_i = \alpha + \sin(x_i) + \epsilon_i$$ in R is: model <- lm(y ~ 1, offset=sin(x), data=df) which produces the residual vs fitted plot: or as a residuals vs x plot: Alternatively, one could fit model2 <- lm( y-sin(x) ~ 1, data=df) which gives the same estimate for $\alpha$. The residual vs fitted plot is of no use in this case (because of the difference in the way the offset was brought into the model by modifying $y$), but the residuals vs x plot is identical to the second plot above. Gung is right to suggest in comments that it often makes sense to fit the offset as a regressor anyway (for example, to check that the offset-coefficient of 1 is reasonable); this is the model that Scortchi and Peter Flom were discussing in comments. Here's how you do that: model3 <- lm( y ~ sin(x), data=df) If we look at the summary (summary(model3)) we get: Coefficients: Estimate Std. Error t value Pr(>|t|) (Intercept) 0.000717 0.032050 0.022 0.982 sin(x) 1.069593 0.044947 23.797 <2e-16 *** --- Signif. codes: 0 ‘***’ 0.001 ‘**’ 0.01 ‘*’ 0.05 ‘.’ 0.1 ‘ ’ 1 Residual standard error: 0.9921 on 998 degrees of freedom Multiple R-squared: 0.362, Adjusted R-squared: 0.3614 F-statistic: 566.3 on 1 and 998 DF, p-value: < 2.2e-16 which has coefficients close to what we'd expect. Finally, you might do this: model4 <- lm( y ~ sin(x),offset=sin(x), data=df) but its effect is only to reduce the fitted coefficient of $\sin(x)$ by 1, so we can extract the same information from model3's output. thank you for your answer. Actually, I just completed my degree's introductory statistics course. Could you please recommend some textbooks that covers the concepts in your answer? E.g. I have never come across the term 'offset' before. The use of the term 'offset' is most commonly encountered when discussing generalized linear models (GLMs). The concept exists in ordinary regression, but since it's so easy to just adjust $y$ for fixed-coefficient terms instead (as I did with model2 above), people often don't run into it until they deal with GLMs (where you generally can't just adjust the data in that fashion) and almost any decent introduction to GLMs discuss the idea. Fox & Weisberg's An R Companion to Applied Regression does cover the use of an offset in regression in R. However, when dealing with a fixed-coefficient term, if the idea of moving from $y=\alpha+\sin(x)+\epsilon$ to $y^* =\alpha+\epsilon$ where $y^=y-\sin(x)$, and then using $\hat y = \hat y^+\sin(x)$ makes sense to you, you don't really need any special concepts to deal with offsets.
import {ButtonProps} from '@material-ui/core/Button'; import {DialogProps} from '@material-ui/core/Dialog'; import {clone, merge} from 'lodash'; export interface IButtonAttr { props?: ButtonProps; text: string | React.ReactNode; action: string; preventClose?: boolean; } export interface IModalOptions { dialogProps?: DialogProps; title?: string | React.ReactNode; description?: string | React.ReactNode; cancelText?: string | React.ReactNode | boolean; cancelProps?: ButtonProps; confirmText?: string | React.ReactNode | boolean; confirmProps?: ButtonProps; confirmPreventClose?: boolean; buttons?: IButtonAttr[]; force?: boolean; } export interface IModal extends IModalOptions { id: number; cancelProps?: any; confirmProps?: any; resolve: any; reject: any; cb?: any; open?: boolean; } interface IListener { resolve: any; reject: any; cb?: any; dialog: IModalOptions; id: number; } export default class ModalityService { public static getInstance() { if (!this.instance) { this.instance = new ModalityService(); } return this.instance; } private static instance: ModalityService; private uniqueId: number; private readonly listeners: { [key: number]: IListener; }; private newConfirmHandler: ((dialog: IModal) => void) | null; private readonly defaultOptions: IModalOptions; private constructor() { this.uniqueId = 0; this.listeners = {}; this.newConfirmHandler = null; this.defaultOptions = { cancelText: 'Cancel', confirmText: 'Confirm', }; } public open(options: IModalOptions, cb?: (action: 'confirm' | 'cancel' | string) => void): Promise<'confirm' | 'cancel' | string> { const data = clone(merge(clone(this.defaultOptions), clone(options))); this.uniqueId++; const id = this.uniqueId; (data as IModal).id = id; let internalResolve: any = null; let internalReject: any = null; const promise: Promise<'confirm' | 'cancel' | string> = new Promise((resolve, reject) => { internalResolve = resolve; internalReject = reject; }); (data as IModal).resolve = internalResolve; (data as IModal).reject = internalReject; (data as IModal).cb = cb; this.listeners[id] = { dialog: data, id, reject: internalReject, resolve: internalResolve, cb, }; if (this.newConfirmHandler) { this.newConfirmHandler(data as IModal); } return promise; } public setNewConfirmListener(fn: any) { this.newConfirmHandler = fn; } public removeListener(id: number) { delete this.listeners[id]; } }
Compile method, debug method, compile program and debug program ABSTRACT Language specification in each of at least two particular ranges set in an inputted program are decided in a partial language specification deciding step. It is judged if there is a difference between the language specifications in the particular ranges is judged in a judging step. At least a part of codes in one of the particular ranges is corrected when it is judged that there is the difference between the language specifications in the particular ranges in a partial code correcting step. BACKGROUND OF THE INVENTION 1. Field of the Invention The present invention relates to a compile method, a debug method, a compile program and a debug program for converting a program described in a high-level language into an object program. 2. Description of the Related Art In software development in recent years, a program scale has been increasing, and an object-oriented language having a high maintainability and a high reusability has been watched based on such situations. A typical example of the object-oriented language is the C++ language. The C++ language is becoming the focus of attention as a language that replaces the C language conventionally widely used for programming over the years, and the language in the programming shifts from the C language to the C++ language across the relevant technical field. In the language shift, an object code operated without any problem can be generated only through simply replacing a C language compiler with a C++ language compiler under normal conditions because the C++ language is a high-order compatible language of the C language. However, when a program recited in the C language is compiled in the C++ language, there causes a problem that a code size and an execution time are unfavorably increased. A conventional first solution for the problem is to use a linkage assignment with respect to a program as recited in the Literature (see “The linkage assignment in Chapter 7.4, “The Annotated C++ Reference Manual” written by M. A. Ellis, B. Stroustrup, translated by Takanori Adachi and Hiroshi Koyama) A second solution is to analyze a range of language specification in a program and automatically apply the language specification of a subset for the compilation as far as possible as recited in No. 2003-50700 of the Japanese Patent Applications Laid-Open. In the first solution, however, a programmer is forced to take the linkage assignment into account in the programming, which stand the way of easily shifting from the C language to the C++ language. In the second solution, a consistency is not taken account with respect to encoding symbols in the program (function name, variable name and the like) though compilation from the C++ language to the EC++ language that is the C++ subset specification for built-in use is possible. As a result, it becomes impossible to expect optimization such as the through reduction of the code size and the execution time based on the degeneracy of the language specification. SUMMARY OF THE INVENTION Therefore, a main object of the present invention is to provide a compile method wherein a program can be easily shifted by a programmer to an upward compatible program without being aware of any linkage assignment, and reduction of a code size and an execution time can be realized as a result of maximal degeneracy of a language specification, and a debug method capable of tuning easily. In order to solve the foregoing problem, a compile method according to the present invention is a compile method for converting an inputted program into an object program, including: a partial language specification deciding step for deciding language specification in at least two particular ranges set in the inputted program; a judging step for judging if there is a difference between the language specifications in the particular ranges; and a partial code correcting step for correcting at least a part of codes in one of the particular ranges when it is judged that there is the difference between the language specifications in the particular ranges. According to the foregoing method, since the codes are corrected so that they can be combined with each other, programs including the partially different language specification can be combined and compiled in an optimum language specification, the codes can be generated more efficiently. Preferably, the language specifications are decided based on facility of a programming language used in the particular ranges in the partial language specification deciding step. According to the foregoing method, since the programmer can combine programs including the partially different language specifications without correcting a source program, and the programs can be compiled in an optimum language specification, the codes can be generated more efficiently. As a more preferable mode of the foregoing method, the language specifications are decided based on a language specification control statement in the case where the language specification control statement is present in the inputted program in the partial language specification deciding step. According to the foregoing method, for example, the programmer describes the language specification control statement according to a #pragma instruction in the source program, and thereby the language specification can be freely selected without any influence from facilities originally described in the source program. More preferably, in the language specification deciding step, the language specifications are decided based on a language specification control instruction in the case where the language specification control instruction is given to a compile system for compiling the inputted program. According to the foregoing method, for example, the programmer provides the language specification control instruction based on an option of the compile system, and thereby the language specification can be freely selected without correcting the source program and without any influence from the facilities originally described in the source program. More preferably in the partial code correcting step, a function coding name having a language specification of a subset is changed into a function coding name having a language specification of a superset in the case where there is a difference between the language specifications in definition and reference in all of functions. According to the foregoing method, the programs can be combined even when there is any difference between the language specifications of the definition and the reference. As a more preferable mode of the foregoing method, in the partial code correcting step, when all of external functions that are multiple-defined are changed into an external function which has a singular function name and is not multiple-defined in the program, and the all of external functions that are multiple-defined are changed into the external function that has the singular function name and is not multiple-defined in the program, the partial language specification deciding step is executed again. According to the foregoing method, the C++ program that uses the multiple-defined functions can be compiled as the C program, which improves the efficiency in the code generation. As a more preferable mode of the foregoing method, in the partial code correcting step, when all of external functions which belong to a namespace, are changed into an external function that has a singular function name and does not belong to the namespace in the program, and the all of external functions which belong to the namespace are changed into the external function that has the singular function name and does not belong to the namespace in the program, the partial language specification deciding step is executed again. According to the foregoing method, the C++ program using the namespace can be compiled as the C program, which improves the efficiency in the code generation. As a more preferable mode of the foregoing method, in the partial code correcting step, when all of external functions that are generated from a template are changed into an external function that has a singular function name and is not generated from the template in the program, and the all of external functions that are generated from the template are changed into the external function that has the singular function name and is not generated from the template in the program, the partial language specification deciding step is executed again. According to the foregoing method, the C++ program using the namespace can be compiled as the C program, which improves the efficiency in the code generation. A debug method according to the present invention is a debug method for debugging an inputted program, including: a partial language specification acknowledging step for acknowledging language specifications in each of at least two particular ranges set in the inputted program; and a partial language specification displaying step for displaying the acknowledged language specifications in each of the particular ranges together with program source thereof. According to the foregoing method, the programmer can easily confirm which part of the program is compiled through which of the language specifications. As a result, the debugging and tuning operations can be efficiently performed. As a preferable mode of the foregoing method, the method further includes a template development displaying step for displaying a source program in which the template is developed in the case where an object to be analyzed is an instance generated from the template. According to the foregoing method, the programmer can easily grasp the source program after the template is developed, and the debugging and tuning operations can be efficiently performed. As a more preferable mode of the foregoing method, the method further includes a displaying step of a subset language specification violation part for explicitly showing contents of a part of the program deviating from a language specification of a subset in the case where the language specification of the subset is defined in the acknowledged language specification. According to the foregoing method, the programmer can easily grasp the part of the program deviating from the subset, and the tuning operation can be more efficiently performed. The present invention not only can realize the compile method and the debug method including these characteristic steps but also can realize a compile program and a debug program that make a computer execute these characteristic steps included in the compile method and the debug method, and a compile device and a debugger device which execute these characteristic steps included in the compile method and the debug method. Further, it is needless to say that the compiler and the debugger can be distributed via a recording medium such as CD-ROM (compact disc read-only memory) and a transmission medium such as Internet. A compile program according to the present invention is a compile program for converting an inputted program into an object program, the compile program making a computer execute: a partial language specification deciding facility for deciding language specification in each of at least two particular ranges set in the inputted program; a facility for judging if there is a difference in the language specifications between the particular ranges; and a partial code correcting facility for correcting at least a part of codes in one of the particular ranges in the case where it is judged there is the difference in the language specifications between the particular ranges. As a preferable mode of the foregoing program, the partial language specification deciding facility decides the language specification based on facility of a programming language used in the particular ranges. As a more preferable mode of the foregoing program, the partial language specification deciding facility decides the language specification based on a language specification control statement in the case where there is the language specification control statement in the inputted program. As a more preferable mode of the foregoing program, the partial language specification deciding facility decides the language specifications based on a language specification control instruction in the case where the language specification control instruction is provided to a compile system for compiling the inputted program. As a more preferable mode of the foregoing program, the partial code correcting facility changes a function-coding name having a language specification of a subset into a function-coding name having a language specification of a superset in the case where there is a difference between the language specifications of definition and reference in all of functions. As a more preferable mode of the foregoing program, the partial code correcting facility changes all of external functions which are multiple-defined into an external function which has a singular function name and is not multiple-defined in the program, and the partial language specification deciding facility is executed again when the all of external functions which are multiple-defined are changed into the external function which has the singular function name and is not multiple-defined in the program. As a more preferable mode of the foregoing program, the partial code correcting facility changes all of external functions that belong to a name space into an external function which has a singular function name and does not belong to the name space in the program, and the partial language specification deciding function is executed again when the all of external functions which belong to the name space are changed into the external function which has the function name and does not belong to the name space in the program. As a more preferable mode of the foregoing program, the partial code correcting facility changes all of external functions that are generated from a template into an external function which has a singular function name in the program and is not generated from the template, and the partial language specification deciding step is executed again when the all of external functions that are generated from the template are changed into the external function that has the singular function name in the program and is not generated from the template. A debug program according to the present invention is a debug program for debugging an inputted program and for making a computer execute the following facilities: a partial language specification acknowledging facility for acknowledging language specifications in each of at least two particular ranges where the inputted program is set; and a partial language specification displaying facility for displaying the acknowledged language specifications in each of the particular ranges together with program source thereof. As a preferable mode of the foregoing program, it is the debug program for further making the computer execute a template development displaying facility for displaying a source program in which the template is developed in the case where an object for analysis is an instance generated from the template. As a more preferable mode of the foregoing program, it is the debug program for further making the computer execute a subset language specification violation displaying step for explicitly showing contents of a part of the program deviating from a language specification of a subset in the case where the language specification of the subset is defined in the acknowledged language specification. According to the compile method of the present invention, the programs including the partially different language specifications can be combined easily and compiled in an optimum language specification. As a result, the codes can be more efficiently generated. Further, according to the debug method of the present invention, the programmer can easily grasp by which of the language specifications the part of the program is compiled and which part of the source program should be corrected when it is changed into the language specification of the subset. As a result, the debugging and tuning operations can be efficiently executed. The compile method and the debug method according to the present invention can be effectively applied to a compile method, a debug method and the like for an embedded device which demands an object code having a small code size such as a mobile telephone and PDA (personal digital assistant). BRIEF DESCRIPTION OF THE DRAWINGS These and other objects as well as advantages of the invention will become clear by the following description of preferred embodiments of the invention. A number of benefits not recited in this specification will come to the attention of those skilled in the art upon the implementation of the present invention. FIG. 1 is a flow chart of processing steps executed by a compiler according to a preferred embodiment of the present invention. FIG. 2 is a flow chart showing details of a partial language specification deciding step. FIG. 3 is a flow chart showing details of a partial code correcting step. FIG. 4 is a flow chart of processing steps executed by a debugger according to the preferred embodiment. FIG. 5 is a flow chart showing details of a partial language specification displaying step. FIG. 6 shows an example of a source program memorized in a program memorizing unit used in a specific example 1. FIG. 7 shows an example of intermediate code information according to the prior art. FIG. 8 shows an example of a source program corrected by a conventional method. FIG. 9 shows an example of intermediate code information after the source program used in the specific example 1 is applied to the partial language specification deciding step. FIG. 10 shows an example of intermediate code information after the source program used in the specific example 1 is applied to the partial language code correcting step step. FIG. 11 shows an example of a source program memorized in a program memorizing unit used in a specific example 2. FIG. 12 shows an assembly code in which the source program used in the specific example 2 is compiled with the C++ language and an assembly code in which the source program used in the specific example 2 is compiled with the C language. FIG. 13 shows an example of the intermediate code information after the source program used in the specific example 2 is applied to a first partial language specification deciding step. FIG. 14 shows an example of the intermediate code information after the source program used in the specific example 2 is applied to a first partial code correcting step. FIG. 15 shows an example of the intermediate code information after the source program used in the specific example 2 is applied to a second partial language specification deciding step. FIG. 16 shows an assembly code in which the method according to the present invention is not applied to the source program used in the specific example 2 and an assembly code in which the relevant method is applied thereto. FIG. 17 shows a debugger monitor that displays a source program used in a specific example 3. FIG. 18 shows a debugger monitor that displays language specification information of the source program used in the specific example 3. FIG. 19 shows a debugger monitor that displays a source program in which a template of the source program used in the specific example 3 is developed. FIG. 20 shows a debugger monitor that displays contents of a part of the program deviating from a C-language specification in the source program used in the specific example 3. FIG. 21 is a flow chart of a conventional program converting method (second solution). DETAILED DESCRIPTION OF THE INVENTION Hereinafter, a compile method according to a preferred embodiment of the present invention is described referring to the drawings. FIG. 1 is a flow chart of processing steps executed by a compiler. The compiler reads a header file and a source program memorized in a program memorizing unit D1 (Step S1). The compiler analyzes the syntax of the read source program, and generates a symbol table and a syntax tree (Step S2). Next, the compiler generates an intermediate code based on the generated syntax tree (Step S3). After that, the compiler executes various optimizing processes to the generated intermediate code (Step S4). Further, the compiler allocates hardware resources such as a register and a memory to all of variables included in the optimized intermediate code (Step S5), and then, converts the resource-allocated intermediate code into an object code and outputs an object program thereof to a generated code memorizing unit D2 (Step S6). The optimizing step S4 includes a partial language specification deciding step S10, a partial code correcting step S11, and else optimizing step S12. In the partial language specification deciding step S10, the intermediate code is analyzed, and the language specification in a partial range of each program is decided. In the partial code correcting step S11, the language specification in the partial range of each program that was decided in the partial language specification deciding step S10 is analyzed, and the codes in the respective ranges are corrected so that they can be combined depending on differences in the language specifications in the respective ranges. The details of the partial language specification deciding step S10 and the partial code correcting step S11 will be described later. The else optimizing step S12 is not described since it is a conventional optimizing step and not a subject matter of the present invention. The source program inputting step S1, the syntax analyzing step S2, the intermediate code generating step S3, the else optimizing step S12, the resource allocating step S5 and the object outputting step S6 are not described in detail because they are similar to the conventional steps and not the main subject matters of the present invention. A description is given below of the partial language specification deciding step S10 and the partial code correcting step S11, that are the subject matters of the present invention. FIG. 2 is a flow chart showing details of the partial language specification deciding step S10. As a loop processing L1, steps S21-25 are executed with respect to each of intermediate codes corresponding to at least two particular ranges set in an inputted program. The processing advances to the Step S21 in the case where there is an object for analysis, while the processing advances to the Step S11 in the absence of the object for analysis. In the Step S21, the language specification is decided based on facility of a programming language used in the range of the object for analysis, language specification information is recorded in an intermediate language, and the processing advances to the Step S22. Here, the language specification information is information capable of judging at least with which of the language specifications it complies (for example, such a character string as “C++” when it comes to the C++ language). In the Step S22, it is judged whether or not there is a language specification control statement in the ranges of the object for analysis. The language specification control statement recited here, is the one where a particular operation is designated to the compiler by describing in the source program so that, for example, as a #pragma instruction. The processing advances to the step S23 when a result of the judgment is true, while advancing to the Step S24 when the result is not true. In the Step S23, the language specification are decided based on a language specification control instruction, the language specification information is recorded in the intermediate language, and then, the processing advances to the Step S24. In the Step S24, it is judged if the language specification control instruction is provided to a compile system. The language specification control instruction recited here is the one that it is directly given to the compile system so that a particular operation is designated to the compiler as a command line option. The processing advances to the Step S25 when a result of the judgment is true, while advancing to the loop processing L1 when the result is not true. FIG. 3 is a flow chart showing details of the partial code correcting step S11. In a loop processing L2, Steps S31-S38 are executed by each of the intermediate codes of all of functions. The processing advances to the Step S31 in the presence of the object for analysis, while advancing to the Step S39 in the absence of the object for analysis. In the Step S31, it is judged if there is a difference between the language specifications in a definition and a reference of the relevant function. The difference between the language specifications denotes such a case that the language specification in the definition of the function is recorded “C++” in the intermediate code, and the language specification in the reference of the function is recorded “C” in the intermediate code. The processing advances to the Step S32 when a result of the judgment is true, while advancing to the Step S33 when the result is not true. In the Step S33, a function coding name having the language specification of a subset is changed into a function coding name having the language specification of a superset, and the processing advances to the Step S33. In the Step S33, it is judged if the relevant function is an external function that is multiple-defined. The multiple definition denote such a case that functions are defined so that function names thereof are identical and arguments thereof are different as in f(void) and f(int). The processing advances to the Step S34 when a result of the judgment is true, while advancing to the Step S35 when the result is not true. In the Step S34, the relevant function is changed into an external function that has a singular function name and is not multiple-defined in the program. Then, the processing advances to the Step S35. In the Step S35, it is judged if the relevant function is an external function belonging to a name space. The external function belonging to the name space is a function declared in the scope of the namespace as namespace S{int f(void)}. The processing advances to the Step S36 when a result of the judgment is true, while advancing to the Step S37 when the result is not true. In the Step S36, the relevant function is changed into an external function that has a singular function name and does not belong to the name space in the program. Then, the processing advances to the Step S37. In the Step S37, it is judged if the relevant function is an external function generated from a template. The external function generated from the template here is a function that is instantiated from a function template defined as in template <class T> Tf(T a). The processing advances to the Step S38 when a result of the judgment is true, while advancing to the loop processing L2 when the result is not true. In the Step S38, the relevant function is changed into a external function that has a singular function name and is not generated from the template in the program. Then, the processing advances to the L2. In the Step S39, it is judged if any of the judgment results of the Steps S33, S35 and S37 is true. The processing advances to the Step S10 when the judgment result is true, while advancing to the Step S12 when the result is not true. As described above, the optimizing step S4 including the partial language specification deciding step S10 and the partial code correcting step S11 is executed to the intermediate code, and thereafter the resource allocating step (Step S5) and the object program outputting step (Step S6) are executed to the optimized intermediate code. Thereby, a programmer can shift the program easily to an upward compatible program without being aware of any linkage assignment, and a code size and an execution time can be reduced as a result of the degeneracy of the language specification at a maximum level. The present invention is not limited to the analysis of the intermediate code, and any type of data can be analyzed as far as it is data representing the syntax analyzing result of the source program. Next, a debug method according to the present preferred embodiment is described referring to the drawings. FIG. 4 is a flow chart of processing steps executed by a debugger. The debugger reads a program memorized in the program memorizing unit D1 (Step N1). The debugger analyzes the read program, and thereafter shifts to the command input waiting Step N2. Next, the debugger judges if a partial language specification display command is inputted (Step N3). When a result of the judgment is true, Step N10 for partially displaying the language specification is executed. After that, Step N11 for displaying other source information is executed, and the processing advances to the command input waiting step N2. When the result is not true, another debugging step (Step N4) is executed, and the processing advances to the command input waiting step N2. The details of the partial language specification displaying step N10 will be described in detail later. Since the other source displaying Step N11 is a step for displaying a general source information, and is not a subject matter of the present invention, it is not described. In addition, the program inputting step N1, the command input waiting step N2, and the another debugging step N4 are not described in detail because they are processing steps similar to the conventional steps and are not subject matters of the present invention. Hereinafter, the partial language specification displaying step N10, that is a key structure of the present invention, is described. FIG. 5 is a flow chart showing details of the partial language specification displaying step N10. In a loop processing L3, Steps N21-N26 are executed by each of the debug information corresponding to the particular ranges of the inputted program. The processing advances to the Step N21 in the presence of the object for analysis, while advancing to the Step N11 in the absence of the object for analysis. In the Step N21, it is judged if there is the language specification information. The processing advances to the Step N22 when a result of the judgment is true, while advancing to the loop processing L3 when the judgment is not true. In the Step N22, the language specification of each range is displayed together with its program source based on the language specification information recorded in the debug information, and the processing advances to the Step N23. In the Step N23, it is judged if the object for analysis is an instance generated from the template. The processing advances to the Step N24 when a result of the judgment is true, while advancing to the Step N25 when the result is not true. In the Step N24, the source program in which the template is developed is displayed, and the processing advances to the Step N25. In the Step N25, it is judged if the language specification of the subset is defined to the acknowledged language specification. The processing advances to the Step N26 when a result of the judgment is true, while advancing to the loop processing L3 when the result is not true. In the Step N26, contents of a part of the program deviating from the language specification of the subset are clearly described. When the debugging process including the partial language specification displaying step N10 described above is executed, the programmer can easily grasp with which language specification each part of the program is compiled and which part of the source program should be corrected when it is changed to the language specification of the subset. As a result, the debugging and tuning operations can be efficiently executed. Hereinafter, the compile method and the debug method according to the present invention are described in more detail referring to a specific example in which the C++ language is used as the language specification of the superset and the C language is used as the language specification of the subset. SPECIFIC EXAMPLE 1 FIG. 6 shows an example of the source program memorized in the program memorizing unit D1. A description is given below of the compile method in the case where the source program is inputted. It is assumed that a user designates that <main. cpp> is compiled as the C++ language and <sub.c> is compiled as the C language. The language specification can be designated in such a manner that #pragma is described in the source program, which is, however, similar to the case where the language specification is designated based on a command line, so it is omitted in this description. The source file, <main. cpp>, shown in FIG. 6 refers to the external function f. The source file, <sub. c>, defines the external function f. Referring to FIG. 7, a problem to be solved is mentioned in advance. FIG. 7 shows a part of information in the intermediate codes respectively on the reference side and the definition side of the function f that are generated in the intermediate code generating step S3. In the case of compiling the specific example 1 according to the conventional method, linkage is carried out without taking the language specification of the intermediate code of the object for analysis into account. Therefore, the function coding names are different in the specific example 1 (_f_Fv and _f), which results in generation of an error at the time of the linkage. Therefore, as shown in FIG. 8, it is necessary, to consolidate the function coding name in such a manner that the user corrects the program so that a linkage assignment is explicitly shown (see “C”). Thus, the correcting step is conventionally necessary to use the source in the C language also in the C++ language, which deteriorates a degree of efficiency in the development process. Next, a specific example of the present preferred embodiment is shown. In the present specific example, analysis is performed in a range of a function scope. In the partial language specification deciding step S10, the loop processing L1 shown in FIG. 2 (Steps S21-S25) is executed by each of the intermediate codes that correspond to a range of the respective function scopes with respect to the intermediate codes generated in the intermediate code generating step. In the present specific example, it is assumed that the user designates that <main. cpp> is compiled with the C++ language, and <sub.c> is compiled with the C language. Therefore, the judgment result of the Step S24 is true, and the Step S25 is executed. As a result, the information indicating that the language specification is the C++ language is recorded in the intermediate code on the reference side, while the information indicating that the language specification is the C language is recorded in the intermediate code on the definition side (see FIG. 9). Next, the partial code correcting step S11 is executed. First, the loop processing L2 (Steps S31-S38) shown in FIG. 3 is executed. In the present specific example, there is a difference between the language specification, information of the intermediate codes respectively on the definition side and the reference side (C++ language and C language). Therefore, the judgment result of the Step S31 is true, and the Step S32 is executed. Comparing the C++ language and the C language to each other, the C++ language is the language specification of the superset, while the C language is the language specification of the subset. Therefore, the function coding name of the intermediate code on the definition side having the language specification of the subset is changed into the function coding name of the intermediate code on the reference side having the language specification of the superset (FIG. 10, consolidate to _f_Fv). In the Steps S33-S38 thereafter, it is judged that all of the steps are not true. As described above, by considering the language specification of the intermediate code of the object for analysis so that the same function coding name is consolidated, and it is allowed to link without the explicit assignment of the linkage. Therefore, the step of correcting the source, which was conventionally necessary to use the source in the C language also as the one in the C++ language, becomes unnecessary. As a result, the development process can achieve an improved efficiency. SPECIFIC EXAMPLE 2 FIG. 11 shows an example of a source program using a multiple-definition function memorized in the program memorizing unit D1. The source program in which the namespace and the template are used is processed in a manner similar to the specific example in the case of the multiple-definition function, the description of which is omitted. A description is given below of the compile method in the case where the relevant source program is inputted. It is assumed that <test.cpp> is compiled without the designation of the language specification from the user. Referring to FIG. 12, problem area is pointed out in advance. In the case of compiling the source program <test. cpp> in the present specific example according to the conventional method, all of the functions are compiled as the functions of the C++ language because the multiple-definition function, which is the facility of the C++ language, is used in the source program. Because the functions are compiled as the functions of the C++ language even though only the facility of the C language is used in the bodies of the functions, the code may be redundantly generated. FIG. 12 shows a code generating result in the case where “void f (void) of “test.cpp” is compiled as the function of the C++ language and a code generating result in the case where it is compiled as the function of the C language. As it is clear from FIG. 12, the code is redundantly generated in the code generating result compiled as the function of the C++ language in comparison to the one compiled as the function of the C language. Next, another specific example of the present preferred embodiment is shown. In a manner similar to the specific example 1, analysis is executed to the function scope in the present specific example. In the partial language specification deciding step S10, the loop processing L1 shown in FIG. 2 (Steps S21-S25) is executed by each of the intermediate codes that correspond to the respective function scopes with respect to the intermediate codes generated in the intermediate code generating step. In the present specific example, since the language specification is not designated through the user, the judgment results of the Steps S22 and S24 are both not true, and the Steps S23 and S25 are not executed. Therefore, the language specification information decided based on the facility of the programming language used in the range of the object for analysis is recorded in the intermediate code (Step S21). In the present specific example, the functions, “void f(void)” and “void f(int)” are the multiple-definition functions having the same function name, and are the facilities (functions) of the C++ language. Therefore, the information that the language specification is the C++ language is recorded in the intermediate codes in the range in which these facilities are used (C++ language in all of the intermediate codes shown in FIG. 13). Next, the partial code correcting step S11 is executed. First, the loop processing L2 shown in FIG. 3 (Steps S31-S38) is executed. In the present specific example, since there is no difference between the language specification information of the intermediate code on the definition side and the language specification information of the intermediate code on the reference side (both in the C+language), the Step S32 is not executed, while the Step S34 is executed because the relevant function is the multiple-definition function. In the present specific example, as shown in FIG. 14, the function name is changed into such function names as _ _L1, _ _L2, that are external function names having singular function names and are not multiple-defined in the program, and thereafter each of the language specifications and the function coding names are retained. The multiple-definition function is recited as an example in the present specific example, however, the function name may be changed into an external function having a singular function name and is not multiple-defined in the program in the program in which the namespace and the template are used in a manner similar to the multiple-definition function (Steps S36 and S38). As a result that the Step S34 is executed, the judgment result of the Step S39 is true, and the partial language specification deciding Step S10 is executed again. At the second execution of the Step S10, since the function names of the multiple-definition functions are changed into _ _L1, _ _L2. Therefore, the relevant functions are not acknowledged as the multiple-definition functions. Then, the information indicating that the language specification is the C language is recorded in the Step S21, and the function coding names as the C language are given to the relevant functions (FIG. 15). no particular change is made in the execution of the second Step S1, and the processing advances to the processing steps on and after the Step S12. FIG. 16 shows a code generating result in the case where the present invention is not applied to the present specific example and a code generating result in the case where the present invention is applied thereto. As clear from FIG. 16, any redundant code is deleted in the case where the present invention is applied in comparison to the case where the present invention is not applied. SPECIFIC EXAMPLE 3 A specific example of the debug method is shown. It is assumed in the present specific example that the language specification information recorded in the intermediate code according to the compile method of the present invention is also recorded in debug information. FIG. 17 shows a monitor that executes a debug program. When the partial language specification display command is inputted, the partial language specification displaying step N1 is executed. In the present specific example, the Step N22 is executed because it is assumed that the language specification information is recorded in the debug information, and the language specifications in each of at least two ranges set in the inputted program are displayed together with their program sources based on the language specification information recorded in the debug information. FIG. 18 shows an example of the language specification display. The display example merely shows an example, and any manner is acceptable as far as the information concerning the language specification is displayed. Next, it is judged where or not the object for analysis is an instance generated from the template (Step N23). In the present specific example, in the case where the object for analysis is A<int>obj, a result of the judgment is true because the instance is generated from the template, and the source program where the template is developed is displayed as shown in FIG. 19. The display example merely shows an example, and any manner is acceptable as far as the information concerning the source program where the template is developed is displayed. Next, it is judged whether or not the language specification of the subset is defined to the acknowledged specification (Step N25). In the present specific example, since the C language, that is the subset specification of the C++ language, is defined, a result of the judgment is true. Then, as shown in FIG. 20, contents of a part deviating from the language specification of the C language (template facility) in the range where the language specification is judged to be the C++ language, are explicitly described. The display example merely shows an example, and any manner is acceptable as far as the information concerning the contents of the part deviating from the language specification of the subset is displayed. As is clear from FIGS. 18-20, when the debug method according to the present invention is used, the programmer can easily grasp which part of the program is compiled based on which language specification and which part of the source program should be corrected when it is changed into the language specification of the subset. As a result, the debugging and tuning operation can be efficiently executed. While preferred embodiments of this invention has been described in detail, it will be understood that various modifications may be made therein, and it is intended to cover in the appended claims all such modifications as fall within the true spirit and scope of this invention. 1. A compile method for converting an inputted program into an object program, including: a partial language specification deciding step for deciding language specification in each of at least two particular ranges set in the inputted program; a judging step for judging if there is a difference between the language specifications in the particular ranges; and a partial code correcting step for correcting at least a part of codes in one of the particular ranges when it is judged that there is the difference between the language specifications in the particular ranges. 2. The compile method as claimed in claim 1, wherein the language specifications are decided based on facility of a programming language used in the particular ranges in the partial language specification deciding step. 3. The compile method as claimed in claim 1, wherein the language specifications are decided based on a language specification control statement in the case where the language specification control statement is present in the inputted program in the partial language specification deciding step. 4. The compile method as claimed in claim 1, wherein the language specifications are decided based on a language specification control instruction in the case where the language specification control instruction is provided to a compile system for compiling the inputted program in the partial language specification deciding step. 5. The compile method as claimed in claim 1, wherein a function coding name having a language specification of a subset is changed into a function coding name having a language specification of a superset in the case where there is a difference between the language specifications of definition and reference in all of functions in the partial code correcting step. 6. The compile method as claimed in claim 1, wherein in the partial code correcting step, all of external functions that are multiple-defined are changed into an external function that has a singular function name and is not multiple-defined in the program, and the partial language specification deciding step is executed again when the all of external functions that are multiple-defined are changed into the external function which has the singular function name and is not multiple-defined in the program. 7. The compile method as claimed in claim 1, wherein in the partial code correcting step, all of external functions which belong to a namespace into an external function which has a singular function name and does not belong to the namespace in the program, and the partial language specification deciding step is executed again when the all of external functions which belong to the namespace are changed into the external function which has the singular function name and does not belong to the namespace in the program. 8. The compile method as claimed in claim 1, wherein in the partial code correcting step, all of external functions which are generated from a template are changed into an external function which has a singular function name and is not generated from the template in the program, and the partial language specification deciding step is executed again when the all of external functions which are generated from the template are changed into the external function which has the singular function name and is not generated from the template in the program. 9. A debug method for debugging an inputted program, including: a partial language specification acknowledging step for acknowledging language specifications in each of at least two particular ranges set in the inputted program; and a partial language specification displaying step for displaying the acknowledged language specifications in each of the particular ranges together with program source thereof. 10. The debug method as claimed in claim 9, further including a template development displaying step for displaying a source program where the template is developed in the case where an object for analysis is an instance generated from the template. 11. The debug method as claimed in claim 9, further including a step for displaying violation part to a subset language specification where contents of a part of the program deviating from a language specification of a subset is explicitly shown in the case where the language specification of the subset is defined in the acknowledged language specification. 12. A compile program for converting an inputted program into an object program, the compile program making a computer execute: a partial language specification deciding facility for deciding language specification in each of at least two particular ranges set in the inputted program; a facility for judging if there is a difference between the language specifications in the particular ranges; and a partial code correcting facility for correcting at least a part of codes in one of the particular ranges in the case where it is judged there is difference between the particular ranges. 13. The compile program as claimed in claim 12, wherein the partial language specification deciding facility decides the language specification based on facility of a programming language used in the particular ranges. 14. The compile program as claimed in claim 12, wherein the partial language specification deciding facility decides the language specifications based on a language specification control statement in the case where there is the language specification control statement in the inputted program. 15. The compile program as claimed in claim 12, wherein the partial language specification deciding facility decides the language specifications based on a language specification control instruction in the case where the language specification control instruction is provided to a compile system for compiling the inputted program. 16. The compile program as claimed in claim 12, wherein the partial code correcting facility changes a function coding name having a language specification of a subset into a function coding name having a language specification of a superset in the case where there is a difference between the language specifications of definition and reference in all of functions. 17. The compile program as claimed in claim 12, wherein the partial code correcting facility changes all of external functions which are multiple-defined into an external function which has a singular function name and is not multiple-defined in the program, and the partial language specification deciding facility is executed again when the all of external functions which are multiple-defined are changed into the external function which has the singular function name and is not multiple-defined in the program. 18. The compile program as claimed in claim 12, wherein the partial code correcting facility changes all of external functions which belong to a namespace into an external function which has a singular function name and does not belong to the namespace in the program, and the partial language specification deciding function is executed again when the all of external functions which belong to the namespace are changed into the external function which has the singular function name and does not belong to the namespace in the program. 19. The compile program as claimed in claim 12, wherein the partial code correcting facility changes all of external functions which are generated from a template into an external function which has a singular function name and is not generated from the template in the program, and the partial language specification deciding step is executed again when the all of external functions which are generated from the template are changed into the external function which has the singular function name and is not generated from the template in the program. 20. A debug program for debugging an inputted program, the debug program making a computer execute: a partial language specification acknowledging facility for acknowledging language specifications in each of at least two particular ranges set in the inputted program; and a partial language specification displaying facility for displaying the acknowledged language specifications in each of the particular ranges together with program source thereof. 21. The debug program as claimed in claim 20, wherein the program further makes the computer execute a template development displaying facility for displaying a source program in which the template is developed in the case where an object for analysis is an instance generated from the template. 22. The debug program as claimed in claim 20, wherein the program further makes the computer execute a subset language specification violation displaying step for explicitly showing contents of a part of the program deviating from a language specification of a subset in the case where the language specification of the subset is defined in the acknowledged language specification.
Anwar ARMSTRONG, et al., Plaintiffs, v. NAVIENT SOLUTIONS, LLC, Defendant. Civil Action No. 16-2212 (RDM) United States District Court, District of Columbia. Signed March 2, 2018 Tyler Jay King, Franklin Square Law Group, Washington, DC, for Plaintiffs. Timothy J. McEvoy, Pro Hac Vice, Sean Patrick Roche, Cameron/McEvoy, PLLC, Fairfax, VA, for Defendant. MEMORANDUM OPINION AND ORDER RANDOLPH D. MOSS, United States District Judge In 2014, Plaintiff Ann Holiday received a statement from Navient Solutions, Inc. ("NSI"), a loan servicing company, reflecting a balance of $45,000 for multiple student loans purportedly made to her son, Plaintiff Anwar Armstrong. Asserting that they took out a single student loan for only $9,000, Plaintiffs allege that they contacted NSI to dispute the account balance and that NSI failed to correct the error. Instead, according to Plaintiffs, NSI reported the incorrect loan balance to consumer reporting agencies, which, in turn, used the faulty information from NSI to generate credit reports that negatively affected Armstrong's ability to secure employment. In response, Plaintiffs filed suit in the Superior Court of the District of Columbia, alleging that NSI violated the District of Columbia Consumer Protection Procedures Act ("CPPA") by making a "misleading statement ... regarding their credit worthiness and credit balances." Dkt. 2-2 at 4 (Compl. ¶ 14). They also asserted common law claims for negligence and breach of contract, alleging that NSI "breached its duty of care by failing to safeguard [Plaintiffs'] information, and failing to correct errors in their account balances," id. at 6 (Compl. ¶ 26), and breached its contract with Plaintiffs by "not revers[ing] or adjust[ing] erroneous charges ... and [by] assert[ing] an invalid account balance," id. at 5 (Compl. ¶ 20). NSI removed the action to this Court, and now moves to dismiss under Federal Rule of Civil Procedure 12(b)(6). See Dkt. 5. NSI asserts that all three of Plaintiffs' claims are preempted by the federal Fair Credit Reporting Act ("FCRA"), 15 U.S.C. § 1681 et seq. , and, alternatively, that Plaintiffs have failed to state a claim upon which relief can be granted. For the reasons explained below, the Court concludes that the FCRA preempts Plaintiffs' CPPA and negligence claims in part. In addition, the Court concludes that any non-preempted portion of Plaintiffs' negligence claims fails to state a claim, as does Plaintiffs' claim for breach of contract. With respect to the non-preempted portion of Plaintiffs' CPPA claim, it is unclear whether Plaintiffs have alleged facts sufficient to sustain Article III standing. The Court will, accordingly, grant NSI's motion to dismiss in part and deny it in part, and will direct that the parties show cause why the remaining portion of Plaintiffs' CPPA claim should not be dismissed (or remanded) for want of jurisdiction. I. BACKGROUND For purposes of the pending motion to dismiss, the following facts taken from Plaintiffs' complaint are accepted as true. See Am. Nat'l Ins. Co. v. FDIC, 642 F.3d 1137, 1139 (D.C. Cir. 2011). Plaintiffs Holiday and Armstrong allege that they "entered an agreement for student loan credit" with NSI, Dkt. 2-2 at 5 (Compl. ¶ 18) for the amount of $9,000, id. at 3 (Compl. ¶ 2 (Statement of Facts) ). In December 2014, however, Holiday received a statement from NSI reflecting a balance "due [of] over $45,000" for "multiple private student loans" allegedly made to Armstrong. Id. at 3 (Compl. ¶ 1). Plaintiffs alleged that they requested proof of the claimed loans but that NSI failed to "produce any executed promissory notes evidencing the loans that [it] claimed were owed," id. (Compl. ¶¶ 3-4), and that NSI failed to credit $8,460 in loan payments already made by Plaintiffs, id. at 4 (Compl. ¶ 6). Plaintiffs further allege that NSI failed to correct the erroneous loan balance and "continue[d] to falsely claim that" Plaintiffs "owe[d] amounts that they [did] not owe." Id. at 3 (Compl. ¶¶ 2, 5). Although the complaint is not a model of clarity, Plaintiffs also appear to allege that NSI reported the incorrect balance-owed to consumer reporting agencies, which, in turn, used the information they received from NSI to generate negative credit reports about Armstrong. Id. at 4 (Compl. ¶ 14) (NSI "misled" Plaintiffs and "others" about the loan balance). "[B]ecause of [those] credit report[s]," Plaintiffs allege, Armstrong was not offered two jobs that he applied for in 2015. Id. at 4 (Compl. ¶¶ 7-10). Plaintiffs filed suit against NSI in the Superior Court of the District of Columbia in 2016. Dkt. 2-1 at 2. NSI removed the case to this Court pursuant to 28 U.S.C. §§ 1332, 1441, and 1446, Dkt. 2 at 1-3. On November 14, 2016, NSI moved to dismiss the complaint, arguing that "[a]ll of the Plaintiffs' claims ... [we]re preempted by the [FCRA]," and, alternatively, that they "fail[ed] to state a claim upon which relief c[ould] be granted." Dkt. 5 at 1. II. LEGAL STANDARD A motion to dismiss brought under Federal Rule of Civil Procedure 12(b)(6) is designed to "test[ ] the legal sufficiency of a complaint." Browning v. Clinton , 292 F.3d 235, 242 (D.C. Cir. 2002). In evaluating such a motion, the Court "must first 'tak[e] note of the elements a plaintiff must plead to state [the] claim' to relief, and then determine whether the plaintiff has pleaded those elements with adequate factual support to 'state a claim to relief that is plausible on its face.' " Blue v. District of Columbia , 811 F.3d 14, 20 (D.C. Cir. 2015) (quoting Ashcroft v. Iqbal , 556 U.S. 662, 675, 678, 129 S.Ct. 1937, 173 L.Ed.2d 868 (2009) ) (alterations in original) (citation omitted). Although "detailed factual allegations" are not necessary to withstand a Rule 12(b)(6) motion, Bell Atl. Corp. v. Twombly , 550 U.S. 544, 555, 127 S.Ct. 1955, 167 L.Ed.2d 929 (2007), "a complaint must contain sufficient factual matter, [if] accepted as true, to 'state a claim to relief that is plausible on its face,' " Iqbal , 556 U.S. at 678, 129 S.Ct. 1937 (quoting Twombly , 550 U.S. at 570, 127 S.Ct. 1955 ). A plaintiff can survive a Rule 12(b)(6) motion even if "recovery is very remote and unlikely," but the facts alleged in the complaint "must be enough to raise a right to relief above the speculative level." Twombly , 550 U.S. at 555-56, 127 S.Ct. 1955 (quotation marks omitted). III. ANALYSIS A. Preemption 1. The Fair Credit Reporting Act "Congress enacted the FCRA in 1970 to ensure fair and accurate credit reporting, promote efficiency in the banking system, and protect consumer privacy." Safeco Ins. Co. of Am. v. Burr , 551 U.S. 47, 52, 127 S.Ct. 2201, 167 L.Ed.2d 1045 (2007). The FCRA imposes duties on consumer reporting agencies ("CRAs") and on entities that furnish information about borrowers to CRAs. See 15 U.S.C. § 1681s-2. Although the FCRA does not define "furnisher," courts have noted that "[t]he most common ... furnishers of information are credit card issuers, auto dealers, department and grocery stores, lenders, utilities, insurers, collection agencies, and government agencies." Himmelstein v. Comcast of the Dist., LLC , 931 F.Supp.2d 48, 52 (D.D.C. 2013) (citing Chiang v. Verizon New Eng., Inc. , 595 F.3d 26, 35 n.7 (1st Cir. 2010) ). Here, NSI is a "furnisher" within the meaning of the FCRA because it allegedly provided information about Plaintiffs' loan balances and "creditworthiness" to CRAs, which then used that information to generate credit reports. See Dkt. 2-2 at 4 (Compl. ¶ 14); Dkt. 5 at 12; Dkt. 9 at 9. It is a violation of the FCRA to "furnish any information relating to a consumer to any [CRA] if [the furnisher] knows or has reasonable cause to believe that the information is inaccurate" or to "furnish information relating to a consumer to any [CRA] if (i) the [furnisher] has been notified by the consumer ... that the specific information is inaccurate[,] and (ii) the information is, in fact, inaccurate." 15 U.S.C. § 1681s-2(a)(1). In addition, one who "regularly and in the ordinary course of business furnishes information to one or more [CRAs] about [the furnisher's] transactions or experiences with any consumer" have a duty to correct and to update any credit reporting information that the furnisher "determines is not complete or accurate," and, "[i]f the completeness or accuracy of any [such] information ... is disputed ... by a consumer, the [furnisher] may not furnish the information to any [CRA] without notice that such information is disputed by the consumer." Id. § 1681s-2(a)(2) & (3). If a consumer notifies a CRA that she disputes the accuracy of an item, the FCRA requires the CRA to notify the furnisher, who, in turn, must "conduct an investigation with respect to the disputed information;" "report the results of the investigation to the [CRA];" and, "if the investigation finds that the information is incomplete or inaccurate, report those results to all other [CRAs] to which the person furnished the information and that compile and maintain files on consumers on a nationwide basis." Id. § 1681s-2(b). When it was enacted in 1970, the FCRA expressly preempted state common law claims "in the nature of defamation, invasion of privacy, or negligence with respect to reporting of information against any ... person who furnishes information to a [CRA]." 15 U.S.C. § 1681h(e), Pub. L. 91-508, title VI, § 601, 84 Stat. 1131. The scope of that provision, however, was limited by a proviso, which allowed claims to proceed if the information was "furnished with malice or willful intent to injure [the] consumer." Id. In 1996, however, Congress returned to the question of preemption under the FCRA. Although Congress did not repeal the preemption provision adopted in 1970, it added a new and more comprehensive one. That provision, which is codified at 15 U.S.C. § 1681t(b), provides in relevant part: No requirement or prohibition may be imposed under the laws of any State-(1) with respect to any subject matter regulated under- ... (F) section 1671s-2 of this title, relating to the responsibilities of persons who furnish information to consumer reporting agencies .... Pub. Law 104-208, Div. A, Title II, § 2419, 110 Stat. 3009 ("Preemption of State Law"). A number of courts have puzzled over the interplay between the 1970 and 1996 preemption provisions. See, e.g., Macpherson v. JPMorgan Chase Bank, N.A. , 665 F.3d 45, 47-48 (2d Cir. 2011) ; Purcell v. Bank of Am. , 659 F.3d 622, 625-26 (7th Cir. 2011) ; Himmelstein , 931 F.Supp.2d at 56-60 ; Manno v. Am. Gen. Fin. Co. , 439 F.Supp.2d 418, 423-29 (E.D. Pa. 2006) ; Johnson v. Citimortgage , 351 F.Supp.2d 1368, 1373-75 (N.D. Ga. 2004) ; Vazquez-Garcia v. Trans Union De Puerto Rico , 222 F.Supp.2d 150, 159-63 (D.P.R. 2002). Some have held that the 1996 preemption provision "only preempts state claims arising after a furnisher of information receives notice of a dispute;" others have held that the 1996 preemption provision applies to statutory claims, while the 1970 provision applies to common law claims; and still others have held that 1996 provision "preempts all related state-law causes of action against furnishers, even willful violations of state common law." Himmelstein , 931 F.Supp.2d at 57-59. The last of these readings of the statute gives the FCRA the broadest preemptive sweep, and it has garnered the support of the two most recent circuit court opinions addressing the issue, Macpherson , 665 F.3d at 48 ; Purcell , 659 F.3d at 625-26, as well as recent opinions from this Court, Himmelstein , 931 F.Supp.2d at 60 ; Rivera v. JPMorgan Chase Bank , 140 F.Supp.3d 88, 93 (D.D.C. 2015). For the reasons explained below, the Court is persuaded by the reasoning in those decisions. First, as explained in Himmelstein , the first theory-which posits that the 1996 provision preempts only state law claims arising after a furnisher of information receives notice of a dispute-is untenable. 931 F.Supp.2d at 57-58. It finds no support in the text of the statute, and it would "perversely afford[ ] 'a furnisher of information more protection from exposure to liability for acts committed after receiving notice of dispute than for acts committed before such notice." Id. (quoting Johnson , 351 F.Supp.2d at 1375 ). The second theory-which posits that the 1996 provision preempts state statutory law but not state common law-is "slightly more attractive." Id. at 58. According to those courts that have adopted this approach, the phrase "requirement[s] or prohibition[s] ... imposed under the laws of any State" is most naturally construed to refer to state statutory law; the phrase "would be an awkward, roundabout way of forbidding state courts" from interpreting or applying state tort law. Manno , 439 F.Supp.2d at 426. They add, moreover, that the absence of any mention by Congress of § 1681h(e) at the time it enacted § 1681t suggests that it viewed the two provisions as distinct-one governing tort claims, and the other governing statutory claims. Id. The problem with this theory is that the Supreme Court considered and rejected a similar argument in Cipollone v. Liggett Group, Inc. , 505 U.S. 504, 112 S.Ct. 2608, 120 L.Ed.2d 407 (1992). In that case, the Court was called upon to construe the preemption provision contained in the Public Health Cigarette Smoking Act of 1969. Using language similar to the language Congress used in the 1996 FCRA preemption provision, the Public Health Cigarette Smoking Act provided: No requirement or prohibition based on smoking and health shall be imposed under State law with respect to the advertising or promotion of any cigarettes the packages of which are labeled in conformity with the provisions of [the Public Health Cigarette Smoking Act]. 15 U.S.C. § 1334(b) (emphasis added). Construing this language, the Cipollone plurality first rejected the contention that "common-law damages actions do not impose 'requirement[s] or prohibition[s]." 505 U.S. at 521, 112 S.Ct. 2608. As the plurality explained: "The phrase '[n]o requirement or prohibition' sweeps broadly and suggests no distinction between positive enactments and common law; to the contrary, those words easily encompass obligations that take the form of common-law rules." Id. The plurality then rejected the further argument that "the phrase 'imposed under State law' " refers to statutory law, and not common law. Id. at 522, 112 S.Ct. 2608. To the contrary, as the Supreme Court observed, "[a]t least since Erie R. Co. v. Tompkins , 304 U.S. 64, 58 S.Ct. 817, 82 L.Ed. 1188 (1938), [the Court] ha[s] recognized the phrase 'state law' to include common law as well as statutes and regulations." Id. Providing a majority on this portion of the plurality's analysis, Justices Scalia and Thomas agreed that "the language of the Act plainly reaches beyond such [positive] enactments;" that "the general tort-law duties petitioner invokes against the cigarette companies can, as a general matter, impose 'requirement[s] or prohibition[s]' within the meaning of [the Act];" and that "the phrase 'State law' as used in that provision embraces state common law." Id. at 548-49, 112 S.Ct. 2608 (Scalia, J., concurring in the judgment in part and dissenting in part). Significantly, the Supreme Court decided Cipollone before Congress enacted the 1996 FCRA preemption provision. That is important for two related reasons. First, congressional intent is the ultimate touchstone of preemption, Wyeth v. Levine , 555 U.S. 555, 565, 129 S.Ct. 1187, 173 L.Ed.2d 51 (2009), and courts generally assume that Congress is aware of the judicial precedent against which it legislates, see Merck & Co., Inc. v. Reynolds , 559 U.S. 633, 648, 130 S.Ct. 1784, 176 L.Ed.2d 582 (2010). Second, as the Supreme Court wrote in Riegel v. Medtronic , "Congress is entitled to know what meaning [the courts] will assign to terms regularly used in its enactments." 552 U.S. at 324, 128 S.Ct. 999. The Supreme Court, accordingly, has admonished that, "[a]bsent other indication, references to a State's 'requirements' " should be construed to "include[ ] ... common-law duties." Id. ; see also Bates v. Dow Agrosciences LLC , 544 U.S. 431, 443, 125 S.Ct. 1788, 161 L.Ed.2d 687 (2005) ; Williamson v. Mazda Motor of Am., Inc. , 562 U.S. 323, 328-29, 131 S.Ct. 1131, 179 L.Ed.2d 75 (2011). Returning to the FCRA, the only "other indicia" of congressional intent is the fact that Congress did not repeal or mention the 1970 preemption provision at the time it adopted the 1996 provision (and, indeed, made an unrelated amendment to the 1970 provision, see Omnibus Consolidated Appropriations Act, 1997, Pub. L. 104-208, 110 Stat. 3009 (1996) ). That fact, however, is far from compelling. As the Seventh Circuit has explained, we do not perceive any inconsistency between the two statutes. Section 1681h(e) preempts some state claims that could arise out of reports to credit agencies; § 1681t(b)(1)(F) preempts more of these claims. Section 1681h(e) does not create a right to recover for wil[l]fully false reports; it just says that a particular paragraph does not preempt claims of that stripe. Section 1681h(e) was enacted in 1970. Twenty-six years later, in 1996, Congress added § 1681t(b)(1)(F) to the United States Code. The same legislation also added § 1681s-2. The extra federal remedy in § 1681s-2 was accompanied by extra preemption in § 1681t(b)(1)(F), in order to implement the new plan under which reporting to credit agencies would be supervised by state and federal administrative agencies rather than judges. Reading the earlier statute, § 1681h(e), to defeat the later-enacted system in § 1681s-2 and § 1681t(b)(1)(F), would contradict fundamental norms of statutory interpretation. .... [T]he statutes are compatible: the first-enacted statute preempts some state regulation of reports to credit agencies, and the second-enacted statute preempts more.... This understanding does not vitiate the final words of § 1681h(e), because there are exceptions to § 1681t(b)(1)(F). When it drops out, § 1681h(e) remains. But, even if our understanding creates some surplusage, courts must do what is essential if the more recent enactment is to operate as designed. Purcell , 659 F.3d at 625. The Court, accordingly, has little difficulty concluding that the 1996 FCRA preemption provision applies to both statutory and common law claims. 2. Plaintiffs' Inaccurate Reporting Claims Plaintiffs' CPPA claims challenge, in part, NSI's failure to accurately report their loan balance to CRAs. Plaintiffs allege, for example, that NSI's "representations ... regarding the amount of the account balance" misled Plaintiffs "and others "-presumably the CRAs that generated negative credit reports about Armstrong-"regarding [Plaintiffs'] creditworthiness and credit balances," and that Armstrong lost employment opportunities "because of his credit report." Dkt. 2-2 at 4-5 (Compl. ¶¶ 7-10, 14-15). These allegations implicate § 1681s-2's requirement that NSI "provide accurate credit information" and that it "investigate, report, and correct inaccurate information upon notice of a dispute." Ihebereme v. Capital One, N.A. , 933 F.Supp.2d 86, 98 (citing 15 U.S.C. § 1681s-2(a) - (b) ). And, as a result, they trigger the 1996 preemption provision, which bars claims "with respect to any subject matter regulated under ... [S]ection 1681s-2 ... relating to the responsibilities of persons who furnish information to" CRAs. 15 U.S.C. § 1681t(b)(1)(F). Accordingly, to the extent Plaintiffs' CPPA claims challenges NSI's disclosure of allegedly inaccurate information to one or more CRAs, it is preempted by the FCRA. See Ihebereme , 933 F.Supp.2d at 107 (dismissing as preempted the plaintiffs' claim that "defendants violated the [CPPA] by disseminating false information to credit bureaus"). For similar reasons, Plaintiffs' breach of contract and negligence claims are also preempted to the extent they challenge NSI's disclosure of inaccurate information to CRAs. Both of those claims allege that NSI failed to maintain accurate records regarding Plaintiffs' loans and loan balances, Dkt. 2-2 at 5-6 (Compl. ¶¶ 20, 26), and both suggest that, as a result of NSI's errors, Armstrong lost employment opportunities, Dkt. 2-2 at 4-6 (Compl. ¶¶ 8, 10, 21, 27). For the same reasons described above, those allegations implicate § 1681s-2, see 15 U.S.C. § 1681s-2(a)(1)(A) ("A person shall not furnish any information relating to a consumer to any [CRA] if the person knows or has reasonable cause to believe that the information is inaccurate."); id. § 1681s-2(a)(1)(B) ("A person shall not furnish information relating to a consumer to any [CRA] if ... the person has been notified by the consumer ... that specific information is inaccurate; and ... the information is, in fact, inaccurate."). These claims are, therefore, preempted by § 1681t(b)(1)(F) as well. B. Failure To State a Claim Without the preempted claims, all that remains of Plaintiffs' complaint is the allegation that NSI misreported information to the Plaintiffs themselves. For the reasons described below, that factual allegation fails to state a claim under D.C. common law or the CPPA. 1. Breach of Contract The complaint alleges that Plaintiffs "entered into an agreement for student loan credit" with NSI and that NSI "breached [its] contract [with Plaintiffs] because it did not reverse or adjust erroneous charges and/or payments and/or adjustments and asserts an invalid account balance." Dkt. 2-2 at 5 (Compl. ¶¶ 18, 20). Because Plaintiffs' credit reporting claims are preempted, all that remains of this claim is their contention that NSI breached some contractual duty it owed them by misstating their account balance in correspondence with them. As currently pled, Plaintiffs' complaint fails to state a claim. Under D.C. law, "[t]o prevail on a claim of breach of contract, a party must establish (1) a valid contract between the parties; (2) an obligation or duty arising out of the contact; (3) a breach of that duty; and (4) damages caused be [the] breach." Brown v. Sessoms , 774 F.3d 1016, 1024 (D.C. Cir. 2014) (quoting Tsintolas Realty Co. v. Mendez , 984 A.2d 181, 187 (D.C. 2009) ). Plaintiffs' complaint fails to allege any contractual obligation or duty that protects Plaintiffs from incurring the "loss of time and convenience" stemming from contacting NSI to dispute an inaccurate account balance. Dkt. 2-2 at 4-5 (Compl. ¶¶ 11, 21). Nor have Plaintiffs alleged any recoverable damages flowing from NSI's alleged breach. For example, Plaintiffs claim that their injuries include "loss in creditworthiness" and "lost income and job opportunities," id. at 4 (Compl. ¶ 11), but those injuries could be caused only by NSI's reporting of allegedly inaccurate information to CRAs. And Plaintiffs' additional damages allegations-their claims for "emotional distress" and "punitive damages," see id. -are not cognizable under D.C. contract law. Given Plaintiffs' failure to "plead[ ] th[e] elements" of a contract claim under D.C. law "with adequate factual support to 'state a claim to relief that is plausible on its face,' " Blue , 811 F.3d at 20 (quoting Iqbal , 556 U.S. at 678, 129 S.Ct. 1937 ) (alterations in original), the Court will dismiss this portion of Plaintiffs' breach of contract claim without prejudice. 2. Negligence Plaintiffs also allege that NSI was negligent because it failed "to exercise reasonable care in dealing with" Plaintiffs, failed "to disclose all material facts," and failed to "correctly report their account balance." Dkt. 2-2 at 6 (Compl. ¶ 24). To allege a negligence claim, a plaintiff must allege facts sufficient to show "(1) a duty, owed by the defendant to the plaintiff, to conform to a certain standard of care; (2) a breach of this duty by the defendant; and (3) an injury to the plaintiff proximately caused by the defendant's breach." Findlay v. CitiMortgage, Inc. , 813 F.Supp.2d 108, 120 (D.D.C. 2011) (quoting District of Columbia v. Fowler , 497 A.2d 456, 463 n.13 (D.C. 1985) ). The tort, moreover, "must exist in its own right independent of [any] contract, and any duty upon which the tort is based must flow from considerations other than [a] contractual relationship." Carter v. Bank of Am., N.A. , 888 F.Supp.2d 1, 15 (D.D.C. 2012) (quoting Nugent v. Unum Life Ins. Co. of Am. , 752 F.Supp.2d 46, 53-54 (D.D.C. 2010) ). In other words, "[t]he tort must stand as a tort even if the contractual relationship did not exist." Id. (quoting Nugent , 752 F.Supp.2d at 54 ). Plaintiffs have failed to identify any authority recognizing a lender or servicer's non-contractual duty "to disclose all material facts" and "correctly [to] report ... account balance[s]" outside a contractual relationship, and the Court can discern no basis for imposing such a duty. Findlay , 813 F.Supp.2d at 120 ("The relationship between a debtor and a creditor is ordinarily a contractual one ...."). Accordingly, Plaintiffs have failed to allege a claim of negligence against NSI independent from their contractual relationship, and the Court will dismiss this portion of Plaintiffs' negligence claim without prejudice. 3. CPPA Finally, Plaintiffs assert that NSI violated the CPPA by "sending [an] inaccurate billing statement[ ] to Plaintiffs," thereby "misrepresent[ing] to them the amounts owed for private student loans." Dkt. 9 at 6; see also Dkt. 2-2 at 4 (Compl. ¶ 14) ("[NSI]'s representations ... regarding the amount of the account balance ... constitute a misleading statement ...."). It is not clear from the present record, however, whether Plaintiffs satisfy the requirements of Article III standing with respect to this claim. See Steel Co. v. Citizens for a Better Env't , 523 U.S. 83, 94-95, 118 S.Ct. 1003, 140 L.Ed.2d 210 (1998). Plaintiffs must plead, as an "irreducible constitutional minimum," that they have suffered "an injury in fact" that is both "concrete and particularized" and "actual or imminent, not conjectural or hypothetical." Lujan v. Def. of Wildlife , 504 U.S. 555, 560, 112 S.Ct. 2130, 119 L.Ed.2d 351 (1992) (internal quotation marks omitted); see also Hancock v. Urban Outfitters, Inc. , 830 F.3d 511, 514 (D.C. Cir. 2016) ("[A]n asserted injury to even a statutorily conferred right 'must actually exist,' and must have 'affect[ed] the plaintiff in a personal and individual way.' " (quoting Spokeo, Inc. v. Robins , --- U.S. ----, 136 S.Ct. 1540, 1543, 1548, 194 L.Ed.2d 635 (2016) ). "[A] lawsuit under the CPPA does not relieve a plaintiff of the requirement to show a concrete injury-in-fact to himself." Silvious v. Snapple Beverage Corp. , 793 F.Supp.2d 414, 417 (D.D.C. 2011) (quoting Grayson v. AT & T Corp. , 15 A.3d 219, 244 (D.C. 2011) ); see also Mann v. Bahi , 251 F.Supp.3d 112, 119 (D.D.C. 2017) ("[I]f [the plaintiff] had only alleged that [the defendant] violated the CPPA without also alleging that he was misled or that he suffered any harm, that would not be sufficient."). Here, Plaintiffs allege that they were injured as a result of NSI's "misleading statement" because they were "in fact misled," and, as a result, suffered "emotional distress" and "loss of time and convenience," for which they seek compensatory damages, statutory damages, and the costs of suit and attorney's fees. Dkt. 2-2 at 4-5 (Compl. ¶¶ 11, 14, 16). Plaintiffs, however, offer no factual allegations that plausibly support the conclusion that they were misled about their loan balances; indeed, they allege that they brought NSI's error to the company's attention. Id. at 3 (Compl. ¶ 2 (Statement of Facts) ). Moreover, the only relevant injuries that they allege are emotional distress and loss of time and convenience. Yet, "a plaintiff can ... establish Article III injury in fact based on emotional harm" only if "that alleged harm stems from the infringement of some legally protected, or judicially cognizable, interest that is either recognized at common law or specifically recognized as such by the Congress." Al-Aulaqui v. Obama , 727 F.Supp.2d 1, 25 (D.D.C. 2010) (internal quotation marks and citations omitted). And relevant precedent also counsels that a plaintiff's loss of time and convenience is not generally sufficient to establish standing, see US Ecology, Inc v. U.S. Dep't of Interior , 231 F.3d 20, 25 (D.C. Cir. 2000), nor is the cost of bringing suit, see Steel Co. , 523 U.S. 83 at 107, 118 S.Ct. 1003, 140 L.Ed.2d 210. Notwithstanding these serious issues, neither party has addressed the question of whether Plaintiffs have alleged facts sufficient to establish injury in fact, and, until the Court is satisfied that it has Article III jurisdiction, it cannot reach the merits of Plaintiffs' CPPA claim. Accordingly, the Court will deny NSI's motion to dismiss as to the non-preempted portion of Plaintiffs' CPPA claim and will order that the parties show cause why that claim should not be dismissed for lack of jurisdiction. CONCLUSION For the reasons explained above, Defendant's Motion to Dismiss, Dkt. 5, is hereby GRANTED in part and DENIED in part. It is hereby ORDERED that the parties show cause on or before March 30, 2018 why the remaining portion of Plaintiffs' CPPA claim should not be dismissed (or remanded) for want of jurisdiction. It is further ORDERED that the parties shall appear before the Court on April 17, 2018 at 2:00 p.m., in Courtroom 21 to address the standing issue and discuss appropriate next steps. "The term 'consumer reporting agency' means any person which, for monetary fees, dues, or on a cooperative nonprofit basis, regularly engages in whole or in part in the practice of assembling or evaluating consumer credit information or other information on consumers for the purpose of furnishing consumer reports to third parties, and which uses any means or facility of interstate commerce for the purpose of preparing or furnishing consumer reports." 15 U.S.C. § 1681a(f). Although at times, there may be good reason to give the phrase a narrower construction, the Court concluded that this history of the Public Health Cigarette Smoking Act counseled in favor of giving the phrase its generally accepted meaning. Cipollone , 505 U.S. at 522-23, 112 S.Ct. 2608. Although a plurality gave the term "requirement" a narrower construction in Medtronic, Inc. v. Lohr , 518 U.S. 470, 116 S.Ct. 2240, 135 L.Ed.2d 700 (1996), it did so only in light of the "implausib[ility]" that Congress would have intended to "preclude[ ] state courts from affording state consumers any protection from injuries resulting from a defective medical device." Id. at 487, 116 S.Ct. 2240. And, in any event, "five justices concluded that common-law causes of action for negligence and strict liability do impose 'requirement[s]' and would be pre-empted by federal requirements specific to a medical device." Riegel v. Medtronic, Inc. , 552 U.S. 312, 323-24, 128 S.Ct. 999, 169 L.Ed.2d 892 (2008) (citing 518 U.S. at 512, 116 S.Ct. 2240 ); id. at 503-05, 116 S.Ct. 2240 (Breyer, J., concurring in part and concurring in the judgment). The Court does not construe Plaintiffs' complaint as seeking a declaratory judgment establishing that they owe NSI only the $9,000 they allege they borrowed (less the $8,460 in payments they allege they made), rather than the "over $45,000" balance on their account statement. Dkt. 2-2 at 3-4 (Compl. ¶¶ 1-2, 6 (Statement of Facts) ). To the extent that Plaintiffs seek such relief, they must seek leave to amend their complaint. See Bond v. U.S. Dep't of Justice , 828 F.Supp.2d 60, 80 (D.D.C. 2011) ("Damages for emotional harm stemming from any breach [are] not recoverable under District of Columbia law." (citing Asuncion v. Columbia Hosp. for Women , 514 A.2d 1187, 1190 (D.C. 1986) ); Fed. Fire Protection Corp. v. J.A. Jones/Tompkins Builders, Inc. , 267 F.Supp.2d 87, 91 (D.D.C. 2003) ("Where the basis of [a] complaint is a breach of contract, it is the general rule in the District of Columbia that punitive damages are not recoverable, even if it is proved that the breach was willful, wanton, or malicious." (citing Sere v. Group Hospitalization, Inc., 443 A.2d 33, 37 (D.C. 1982) ; Den v. Den , 222 A.2d 647, 648 (D.C. 1966) ).
With no explanation, label text_A→text_B with either "not-entailed" or "entailed". text_A: Kevin heard that last week i was diagnosed with insomnia and now i m just so tired of it text_B: Kevin did not hear a pun not-entailed.
Talk:GAME Update V2/@comment-2601:647:4901:210F:701E:313E:163C:68D4-20180826012229 Is it better to spend your saved materials and gold to get top 1-3 once or get top 4-20 twice?
import java.util.Arrays; import java.util.Collection; import java.util.HashMap; import java.util.Map; public class FutureUniverse implements TurnStartListener, TurnDoneListener { private static final Logger log = Logger.getLogger(FutureUniverse.class); private static final OwnerPredicate ENEMY = new OwnerPredicate(Player.ENEMY); private static final OwnerPredicate MINE = new OwnerPredicate(Player.ME); private static final Function<Planet,Integer> GET_GROWTH_RATE = new Function<Planet,Integer>() { @Override public Integer apply(final Planet input) { return input.getGrowthRate(); } }; private static final Function<FuturePlanet,Integer> GET_SHIP_COUNT = new Function<FuturePlanet,Integer>() { @Override public Integer apply(final FuturePlanet input) { return input.getShipCount(); } }; private static final Reduce<Integer> SUM = new Reduce<Integer>(Reduce.IntegerOperation.Add); private final Universe universe; private final FutureUniverse previous; private FutureUniverse next; private final Map<Planet,FuturePlanet> planets; private final int turnDelta; private int turn; private final Memoize<Planets> enemyPlanets; private final Memoize<Integer> enemyShipCount; private final Memoize<Integer> enemyPlanetShipCount; private final Memoize<Integer> enemyProduction; private final Memoize<Planets> myPlanets; private final Memoize<Integer> myShipCount; private final Memoize<Integer> myPlanetShipCount; private final Memoize<Integer> myProduction; private final Memoize<Fleets> fleets; private final Collection<Memoize<?>> caches; public FutureUniverse(final Universe universe, final Collection<FuturePlanet> futurePlanets) { this(universe, futurePlanets, null); } public FutureUniverse(final FutureUniverse previous) { this( previous.universe, new Mapper<FuturePlanet,FuturePlanet>(previous.planets.values(), new Function<FuturePlanet,FuturePlanet>() { @Override public FuturePlanet apply(final FuturePlanet input) { return Planet.create(input); } }), previous ); } private FutureUniverse(final Universe universe, final Iterable<FuturePlanet> futurePlanets, final FutureUniverse previous) { super(); this.universe = universe; this.planets = new HashMap<Planet,FuturePlanet>(); for (final FuturePlanet futurePlanet : futurePlanets) { this.planets.put(futurePlanet.getPlanet(), futurePlanet); } this.previous = previous; if (previous != null) { if (previous.next == null) { previous.next = this; } this.turnDelta = previous.turnDelta + 1; this.turn = previous.turn + 1; } else { this.turnDelta = 0; this.turn = 0; } final Function<Planet, Integer> getShipCount = Functions.compose(GET_SHIP_COUNT, Functions.forMap(planets)); enemyPlanets = new Memoize<Planets>() { @Override protected Planets calculate() { final Planets enemyPlanets = new Planets(); for (final FuturePlanet futurePlanet : new Filter<FuturePlanet>(planets.values(), ENEMY)) { enemyPlanets.add(futurePlanet.getPlanet()); } return enemyPlanets; } }; enemyShipCount = new Memoize<Integer>() { @Override protected Integer calculate() { return getEnemyPlanetShipCount() + getFleets().getEnemyShipCount(); } }; enemyPlanetShipCount = new Memoize<Integer>() { @Override protected Integer calculate() { return SUM.reduce(new Mapper<Planet,Integer>(enemyPlanets.get(), getShipCount)); } }; enemyProduction = new Memoize<Integer>() { @Override protected Integer calculate() { return SUM.reduce(new Mapper<Planet,Integer>(enemyPlanets.get(), GET_GROWTH_RATE)); } }; myPlanets = new Memoize<Planets>() { @Override protected Planets calculate() { final Planets myPlanets = new Planets(); for (final FuturePlanet futurePlanet : new Filter<FuturePlanet>(planets.values(), MINE)) { myPlanets.add(futurePlanet.getPlanet()); } return myPlanets; } }; myShipCount = new Memoize<Integer>() { @Override protected Integer calculate() { return getMyPlanetShipCount() + getFleets().getMyShipCount(); } }; myPlanetShipCount = new Memoize<Integer>() { @Override protected Integer calculate() { return SUM.reduce(new Mapper<Planet,Integer>(myPlanets.get(), getShipCount)); } }; myProduction = new Memoize<Integer>() { @Override protected Integer calculate() { return SUM.reduce(new Mapper<Planet,Integer>(myPlanets.get(), GET_GROWTH_RATE)); } }; fleets = new Memoize<Fleets>() { @Override protected Fleets calculate() { return universe.findFleets( new Predicate<Fleet>() { public boolean apply(final Fleet obj) { return obj.getTurnsRemaining() > turnDelta; } } ); } }; caches = Arrays.<Memoize<?>>asList( enemyPlanets, enemyShipCount, enemyPlanetShipCount, enemyProduction, myPlanets, myShipCount, myPlanetShipCount, myProduction, fleets ); } public FuturePlanet get(final Planet planet) { return planets.get(planet); } public Planets getEnemyPlanets() { return enemyPlanets.get(); } public int getEnemyShipCount() { return enemyShipCount.get(); } public int getEnemyPlanetShipCount() { return enemyPlanetShipCount.get(); } public int getEnemyProduction() { return enemyProduction.get(); } public Planets getMyPlanets() { return myPlanets.get(); } public int getMyShipCount() { return myShipCount.get(); } public int getMyPlanetShipCount() { return myPlanetShipCount.get(); } public int getMyProduction() { return myProduction.get(); } public int getTurn() { return turn; } public int getTurnDelta() { return turnDelta; } @Override public void turnDone() { if (next != null) { next.turnDone(); } for (final Memoize<?> cache : caches) { cache.reset(); } } @Override public void turnStart(final int turn) { this.turn = turn + turnDelta; for (final FuturePlanet futurePlanet : planets.values()) { futurePlanet.setTurn(this.turn); futurePlanet.resolve(previous != null ? previous.planets.get(futurePlanet.getPlanet()) : null); } if (next != null) { next.turnStart(turn); } handleShipCount(); } private Fleets getFleets() { return fleets.get(); } private void handleShipCount() { final Planets myPlanets = getMyPlanets(); final Planets enemyPlanets = getEnemyPlanets(); final Planets myFrontLinePlanets = new Planets(); if (!myPlanets.isEmpty()) { for (final Planet planet : enemyPlanets) { myFrontLinePlanets.add(myPlanets.findNearest(planet)); } } for (final FuturePlanet futurePlanet : planets.values()) { if (Player.ME.equals(futurePlanet.getOwner())) { handleMine(futurePlanet, enemyPlanets, myFrontLinePlanets.contains(futurePlanet.getPlanet())); } else if (Player.NEUTRAL.equals(futurePlanet.getOwner())) { handleNeutral(futurePlanet); } else { handleEnemy(futurePlanet); } } } private void handleMine(final FuturePlanet futurePlanet, final Planets enemyPlanets, final boolean frontLine) { final Planet planet = futurePlanet.getPlanet(); final Planet closestEnemy = enemyPlanets.findNearest(planet); final int distance = closestEnemy != null ? planet.distance(closestEnemy) : 1; //final int minimumShipCount = planet.getGrowthRate() * (frontLine ? (int) Math.ceil(Math.log(Planet.getMaxDistance()) / Math.log(distance)) : 1); final int shipCount = futurePlanet.getShipCount(); if (next != null) { final FuturePlanet nextFuturePlanet = next.planets.get(planet); if (!Player.ME.equals(nextFuturePlanet.getOwner())) { // deficit futurePlanet.setRequiredShipCount(shipCount + nextFuturePlanet.getShipCount()); } else { // surplus, but how much? final int surplusShipCount = Math.min(shipCount, nextFuturePlanet.getAvailableShipCount()); int closestEnemyShipCount = 0; if (frontLine && closestEnemy != null && closestEnemy.getActualShipCount() > distance * planet.getGrowthRate()) { closestEnemyShipCount = closestEnemy.getActualShipCount() - distance * planet.getGrowthRate(); } futurePlanet.setRequiredShipCount(Math.max(shipCount - surplusShipCount, closestEnemyShipCount)); } } else { // surplus, farthest known turn in the future int closestEnemyShipCount = 0; if (frontLine && closestEnemy != null && closestEnemy.getActualShipCount() > distance * planet.getGrowthRate()) { closestEnemyShipCount = closestEnemy.getActualShipCount() - distance * planet.getGrowthRate(); } futurePlanet.setRequiredShipCount(Math.max(0, closestEnemyShipCount)); } } private void handleEnemy(final FuturePlanet futurePlanet) { futurePlanet.setRequiredShipCount(0); } private void handleNeutral(final FuturePlanet futurePlanet) { futurePlanet.setRequiredShipCount(0); if (next != null) { final Planet planet = futurePlanet.getPlanet(); final FuturePlanet nextFuturePlanet = next.planets.get(planet); if (Player.ME.equals(nextFuturePlanet.getOwner())) { futurePlanet.setRequiredShipCount(nextFuturePlanet.getRequiredShipCount() + futurePlanet.getShipCount()); } else { futurePlanet.setRequiredShipCount(nextFuturePlanet.getRequiredShipCount()); } } } private static class OwnerPredicate implements Predicate<FuturePlanet> { private final Player owner; private OwnerPredicate(final Player owner) { super(); this.owner = owner; } @Override public boolean apply(final FuturePlanet obj) { return owner.equals(obj.getOwner()); } } private static abstract class Memoize<T> { private boolean cached; private T value; public T get() { if (!cached) { value = calculate(); cached = true; } return value; } protected abstract T calculate(); public void reset() { cached = false; value = null; } } }
Doliolum krohni, n. sp. These may be divided into two sections : — (l) "With numerous stigmata extending along the greater part of the lengtli of the branchial sac, including the first six species ; and (2) with only a few stigmata at the posterior end of the lirauchial sac, including the remaining three species. Doliolum denticulatum, Q. and G., Huxley, Phil. Trans. 1851, part ii. p. 595. Not Doliolum denticidatum, Q. and G., Krohn, Arch. f. Naturgesch., 1852, p. 57. Not Doliolum denticulatum, Q. and G., Keferstein and Ehlers, Zoologische Beitriige, p. 65, 1861. Not Doliolum denticulatum, Q. and G., Grobben, Arb. zool. Inst. Wien, Bd. iv. p. 55, 1882. Not Doliolum denticulatum, Q. and G., Herdman, Trans. Koy. Soc. Edin., vol. xxxii. part i. p. 93, 1883. This species, the oldest member of the genus, was found in 1827 on the surface of the Pacific Ocean, near the islands of Yanikoro and Amboiua, by the French naturalists Quoy and Gaimard, during the voyage of the " Astrolabe." They described and figured it in the official account of the expedition published in 1835. The diagnosis given is as follows : — The short description which follows the diagnosis adds nothing of importance, but merely shows that the observers mistook the branchial sac for a pair of plume-like internal gills, the muscle bands for vessels, and the endostyle for an aorta. The figures ^ show two views of the entire animal, a representation of the (svijaposed) plume-like gills, and an end view of the anterior extremity. There appear from the figures to have been ten branchial lobes, if Quoy and Gaimard are correct in designating the end which they figure with denticulations as the anterior. The position of the endostyle in relation to the muscle bands would rather lead to the oj)posite conclusion, but probably it is represented too far back in the body. The second species of Doliolum described by Quoy and Gaimard, Doliolum caudatum, is an asexual form with nine muscle bands and a well-marked dorsal outgrowth. Possibly it is the Blastozooid belonging to the present species. During the voyage of H.M.S. " Eattlesnake " in 1849, Huxley found specimens of Doliolum in the South Pacific, a little to the northward of Sydney, N.S.W., between Sydney and New Zealand, and in consideralile numbers just at the entrance of the Bay of Islands. These he identified with the species Doliolum denticidatum described by Quoy and Gaimard, and he published a detailed anatomical account with figures in the
Thread:Scottlukaswilliams/@comment-<IP_ADDRESS>-20181105193711/@comment-1036300-20181106122727 I haven't read the new comics yet but I would like to. Do I know you? I can send a postcard if you'd like.
Rows order in SQLite Database (iOS) I have a database with a table called 'connection', for simplicities' sake, let's say I only have one column which is called 'rowName'. Now let's say I add a row with rowName = a; now I add a row with rowName = q, and lastly I add a row with rowName = w (letters are completely random). Now, I irritate thru the results with the statement: NSString * queryStatements = [NSString stringWithFormat:@"SELECT rowName, FROM tableName"]; and using the code: NSMutableArray * rows = [[NSMutableArray alloc] init]; //create a new array sqlite3_stmt * statement; if(sqlite3_prepare_v2(databaseHandle, [queryStatements UTF8String], -1, &statement, NULL) == SQLITE_OK){ while (sqlite3_step(statement) == SQLITE_ROW){ NSString * rowName = [NSString stringWithUTF8String : (char*) sqlite_column_text(statement, 1)]; [rows addObject : connection]; } sqlite3_finalize(statement_; } In the array rows, will the object at index 0 be rowName = a, and at index 1 rowName=q, and at index 2 rowName = w? or will it be random? Is there a way to make it not-random? Also, if i delete a row, will it have any affect on the other rows order? Never depend on a sort order from your database. Always specify one if it is required. SELECT rowName FROM tableName order by rowName gives you the data sorted by rowName. If you need a different order, you need another column. You can also sort your NSArray if need be. What sort order are you looking for? can i put a row of type DATE in the order by argument? Yes. Technically SQLite doesn't have a date data type, but assuming you did the appropriate conversion when populating the database, you should be fine. If you are relying on sqlite in your app, you may want to check out FMDB on github by Gus Muller. It provides a nice objective c wrapper around all those sqlite c functions. and btw - totally agree with @terry wilcox on the "order by suggestion" - try to avoid using column indexes when possible.
Talk:Ensemble Stars! Anime Adaption/@comment-30992005-20180505173952 Valkyrie, Switch, MaM, and Eden will be in the anime Also a website for the anime too
User talk:Lumenos/Archive1 WELCOME LUMENOS! I finally beat Ace in greetering someone. 23:18, 15 August 2009 (UTC) Hi! Welcome to my letter arrangements. Lumenos (talk) 03:59, 19 August 2009 (UTC) Sysopship * 1) Don't block people that aren't sysops, even if they are vandalising, use the vandal break. * 3) Don't protect or delete a page without discussing it on the talk page first. * Wiping with sticks. An excellent idea.&mdash; Unsigned, by: Lumenos / talk / contribs Lumenist naming conventions PI saith, "He seems to name everything after himself " * Get over yourself and relax already. 08:03, 19 August 2009 (UTC) You broke my Wiki! I have half a mind to vandal bin and/or permaban you for doing so! Lol. 06:40, 8 June 2010 (UTC) * Touché! 03:43, 10 June 2010 (UTC) * You never know. 20:01, 16 June 2010 (UTC) As discussed Dear Trall suspect Grammar * Only because I was wondering what peculiar thing you were up to now. 00:53, 26 June 2010 (UTC) * ~ That blind moron (talk) 01:08, 26 June 2010 (UTC) I did a funny Copyright * Reply to Bob: * 1) The source of the material is credited by the links, no? I think I did this consistently. * ~ That nutter who&#39;se always re-editting their comments (talk) 21:22, 26 June 2010 (UTC) So, basically, what you should do is put this at the top of each page copied from here: * That would be my interpretation as well.--BobSpring is sprung! 11:49, 27 June 2010 (UTC)
Talk:Bugged/@comment-25289363-20170708193243/@comment-32210650-20181114050903 It looks like that Buster, Arthur, Binky and Francine are going to kill brain!
from tests.core.fixtures.checks.valid_check_2 import ValidCheck OUTPUT = 'valid_check_1' class InheritedCheck(ValidCheck): def check(self, instance): return OUTPUT
Board Thread:Off Topic/@comment-6069202-20140127223312/@comment-27783182-20140227164126 granted, however the internet is so fast that you can't use it. i wish for a glass bow.
Toggle nav Here are all of the cool things you can learn through OpenGrasp! Click a topic to view related lessons, sources, and reviews for each of them.
Can't import css style in React application Hello I have problem with import css in React application and I don't know what is wrong in my code: Error: 133:1 Uncaught Error: Module parse failed: D:\src\styles\styles.css Unexpected token (1:0) You may need an appropriate loader to handle this file type. |*{ | color:red; | } at eval (133:1) at Object.<anonymous> (bundle.js:979) at __webpack_require__ (bundle.js:20) at eval (app.js?bd9c:4) at Object.<anonymous> (bundle.js:737) at __webpack_require__ (bundle.js:20) at Object.<anonymous> (bundle.js:405) at __webpack_require__ (bundle.js:20) at bundle.js:63 at bundle.js:66 styles.css (only test file): *{ color:red; } webpack config.js: const path = require('path'); module.exports={ entry: './src/app.js', output: { path: path.join(__dirname, 'public'), filename: 'bundle.js' }, module:{ rules: [{ loader: 'babel-loader', test: /\.js$/, exclude: /node_modules/ }, { test: /\.css$/, use: [ 'style-loader', 'css-loader' ] }] }, devtool: 'cheap-module-eval-source-map', devServer:{ contentBase: path.join(__dirname, 'public') } }; package.json: { "name": "", "version": "1.0.0", "main": "index.js", "author": "", "license": "", "scripts": { "serve": "live-server public/", "build": "webpack", "dev-server": "webpack-dev-server" }, "dependencies": { "babel-cli": "6.24.1", "babel-core": "6.25.0", "babel-loader": "7.1.1", "babel-plugin-transform-class-properties": "6.24.1", "babel-preset-env": "1.5.2", "babel-preset-react": "6.24.1", "css-loader": "0.28.4", "live-server": "^1.2.0", "react": "16.0.0", "react-dom": "16.0.0", "react-modal": "2.2.2", "style-loader": "0.18.2", "validator": "8.0.0", "webpack": "3.1.0", "webpack-dev-server": "2.5.1" } } .babelrc : { "presets":[ "env", "react" ], "plugins":[ "transform-class-properties" ] } app.js: import React from 'react'; import ReactDOM from 'react-dom'; import './styles/styles.css'; ReactDOM.render(<App/>, document.getElementById('app')); I dont have idea what is wrong ;? How are you running the app, does it run successfully if you remove the css import ? I use yarn run dev-server in console, yes if i delete css code it run successfully Try this. const path = require('path'); module.exports={ entry: './src/app.js', output: { path: path.join(__dirname, 'public'), filename: 'bundle.js' }, module:{ rules: [{ loader: 'babel-loader', test: /\.js$/, exclude: /node_modules/ }, { test: /\.css$/, include: /node_modules/, use: ['style-loader', 'css-loader'], }] }, devtool: 'cheap-module-eval-source-map', devServer:{ contentBase: path.join(__dirname, 'public') } }; Also, try changing your css. body { background-color: red; } Just to see if it works. Try using this rule: module: { rules: [ { test: /\.css$/, use: [ { loader: 'style-loader' }, { loader: 'css-loader', options: { modules: true } } ] } ] } could you explain why would that fix it ? Thanks for answers. I don't know why but after restrting my computer, my first version suddenly it's working... I think that I need to clean my computer :) Couple ways you can solve this. Add to your "loaders" in package.json { test: /\.css$/, loader: "style-loader!css-loader" } Add to your "rules" in package.json { test: /\.css$/, use: [ 'style-loader', 'css-loader' ]}
Feature Request: traditional "mousekeys" for small adjustments Might be helpful to have press-and-hold keybindings for making small movements up/down/left/right when you get near a target. Keyboardio had this & it was pretty helpful. Thanks again! I never really used mousekeys when I had my keyboardio (I ended up selling it) so I'm not sure what the implementation looks like. If it is anything like the ergodox implementation (which I only briefly played with) then it attempts to provide full mouse emulation (acceleration and all), which is a bit more involved and probably not very useful for small distances. I understand the frustration of the occasional overshoot when warping which I suspect may be better remedied by a discrete movement system (which is also easier to implement) and intend to experiment with this at some point later today. Yeah just thinking linear movement is fine, in the four cardinal directions On Tue, May 12, 2020, 3:10 PM rvaiya<EMAIL_ADDRESS>wrote: I never really used mousekeys when I had my keyboardio (I ended up selling it) so I'm not sure what the implementation looks like. If it is anything like the ergodox implementation (which I only briefly played with) then it attempts to provide full mouse emulation (acceleration and all), which is a bit more involved and probably not very useful for small distances. I understand the frustration of the occasional overshoot when warping which I suspect may be better remedied by a discrete movement system (which is also easier to implement) and intend to experiment with this at some point later today. — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/rvaiya/warp/issues/5#issuecomment-627536513, or unsubscribe https://github.com/notifications/unsubscribe-auth/AAD3MAHHUTS3QJDN5VOWDLLRRGNIZANCNFSM4M6N2Y6A . I've implemented a basic version of this that seems to work reasonably well. Default bindings for manual movement are 'wasd' (see -m) which allows for fairly refined movement in conjunction with the quadrant approach with a bit of practice (you can also use them to shift the entire quadrant in the event that you overshoot the target during one of the recursive movements). Let me know how you get along with it. I got around to installing XFCE (all 2+GB of it!) and did a bit more testing. It turns out there is another issue which prevents certain programs from working properly. X uses a model in which only one client can have control of the keyboard at a given time. Unfortunately since warp needs access to the keyboard in order to process input events this effectively means that pointer actions which result in programs attempting to grab the keyboard will fail (and conversely any attempts by warp to grab the pointer if another program has already done so will also fail). This is most notable on the XFCE top bar but also appears to be true of many graphical program menus which silently fail when they can't grab the keyboard to implement menu shortcuts (interestingly Chrome does not suffer from this issue). Since I principally use chrome, urxvt and a fairly spartan window manager I have managed to remain blissfully ignorant of these issues :P. This is obviously a major problem for anyone who uses GUI oriented setup (and who would consequently most likely benefit from a program like this). It probably also explains the conspicuous lack of such programs already in existence. Potential Solutions As far as I can tell there doesn't exist a way to properly implement something like this in X. There was an attempt to introduce an extension for accessibility related key interception some time ago (XEviE) but it appears to have been abandoned. The following are potential workarounds. Yield the keyboard grab upon click (stop drop and roll). Pros Trivial to implement Allows for popup boxes to be opened (but not subsequently navigated with warp) Potentially solves many use cases. Possible to achieve with pure X. Cons Inconsistent behaviour (clicks which trigger pointer grabs will close the movement session, others will not) Cannot be used while dialog boxes (or anything else which grabs the keyboard) are open. Encourages a hybrid approach (warp to select the menu item/menu item shortcuts for selection). May affect other things like DE widgets which grab the keyboard. Use uinput Pros Doesn't have any of the above limitations since it operates below the X layer. Basic functionality works everywhere (i.e like keyboard firmware). Key interception logic can be partially recycled from a keymapper I wrote. Cons Requires root Involves a complete rewrite + server/client model to realize graphical features. Less elegant :P Resources https://freedesktop.org/wiki/Software/XEvIE/ https://cgit.freedesktop.org/xorg/xserver/commit/?id=f4036f6ace5f770f0fe6a6e3dc3749051a81325a Conclusion Presently I favour the first option but if your main use case involves interacting with a lot of graphical program menus then I am willing to undertake the second (I may eventually end up implementing it as an exercise anyway). Let me know how you want to proceed. will look this over in the next day or so On Thu, May 14, 2020 at 7:19 PM rvaiya<EMAIL_ADDRESS>wrote: I got around to installing XFCE (all 2+GB of it!) and did a bit more testing. It turns out there is another issue which prevents certain programs from working properly. X uses a model in which only one client can have control of the keyboard at a given time. Unfortunately since warp needs access to the keyboard in order to process input events this effectively means that pointer actions which result in programs attempting to grab the keyboard will fail (and conversely any attempts by warp to grab the pointer if another program has already done so will also fail). This is most notable on the XFCE top bar but also appears to be true of many graphical program menus which silently fail when they can't grab the keyboard to implement menu shortcuts (interestingly Chrome does not suffer from this issue). Since I principally use chrome, urxvt and a fairly spartan window manager I have managed to remain blissfully ignorant of these issues :P. This is obviously a major problem for anyone who uses GUI oriented setup (and who would consequently most likely benefit from a program like this). It probably also explains the conspicuous lack of such programs already in existence. Potential Solutions As far as I can tell there doesn't exist a way to properly implement something like this in X. There was an attempt to introduce an extension for accessibility related key interception some time ago (XEviE) but it appears to have been abandoned. The following are potential workarounds. Yield the keyboard grab upon click (stop drop and roll). Pros Trivial to implement Allows for popup boxes to be opened (but not subsequently navigated with warp) Potentially solves many use cases. Possible to achieve with pure X. Cons Inconsistent behaviour (clicks which trigger pointer grabs will close the movement session, others will not) Cannot be used while dialog boxes (or anything else which grabs the keyboard) are open. Encourages a hybrid approach (warp to select the menu item/menu item shortcuts for selection). May affect other things like DE widgets which grab the keyboard. Use uinput Pros Doesn't have any of the above limitations since it operates below the X layer. Basic functionality works everywhere (i.e like keyboard firmware). Key interception logic can be partially recycled from a keymapper I wrote. Cons Requires root Involves a complete rewrite + server/client model to realize graphical features. Less elegant :P Resources https://freedesktop.org/wiki/Software/XEvIE/ https://cgit.freedesktop.org/xorg/xserver/commit/?id=f4036f6ace5f770f0fe6a6e3dc3749051a81325a Conclusion Presently I favour the first option but if your main use case involves interacting with a lot of graphical program menus then I am willing to undertake the second (I may eventually end up implementing it as an exercise anyway). Let me know how you want to proceed. — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/rvaiya/warp/issues/5#issuecomment-628936057, or unsubscribe https://github.com/notifications/unsubscribe-auth/AAD3MABQRTD7YXLV7E7SXTLRRR4A7ANCNFSM4M6N2Y6A . -- Pete Fein | wearpants.org | @wearpants I've refactored most of the code and made several improvements (including a new hints mode). Let me know what you think. You should reread the readme anyway since the configuration mechanism has changed (it is not file based). It is likely I've introduced several bugs in the process, but I'm sure you will let me know as you find them :P. I also recommend reading through #3 if you haven't already. This work great! Though I find myself wanting a mousekeys mode that moves the cursor from it's current position, instead of recentering first - often I'm clicking on two UX elements in succession that are near each other, and it's slow to start over from the center each time I'm not sure how this would work. How large should the grid be if it is activated when the pointer is not in the center of the screen? I suppose one approach might be to take the shortest distance to a screen edge and then use that as the boundary but that introduces other usability problems. What does kaleidoscope do? P.S I strongly recommend trying out hints mode. I actually find it superior to a real mouse in some ways and it seems like it might address a few of the problems you have. I was thinking it'd just move the mouse independent from the grid (and I guess exiting the grid if it was active). Would be definitely useful - sometimes I miss my target by a few pixels when clicking (see also #12) and on many webpages I find I have repeated actions in the same vertical or horizontal plane as current cursor position, and I'd like to just slide over Will play around with hint mode too. So zoom mode looks really neat, but won't work too well with my twiddler hardware, which has a very limited number of accessible keysyms without multi-key chording (or caps-lock style modal sifts for symbols in my config)... and my layout is optimized by letter & bigram frequency, it doesn't logically relate to QWERTY or alphabetical order at all. Too bad for me, I like hint mode quite a lot I've added a parameter called grid_activation_timeout which can be set to 0 to disable automatic deactivation after initial warping (old default behaviour). If you really want a dedicated key to control the pointer locally I can add one but at that point you might want to supplement warp with the default X implementation of mouse keys. I've also added a hint_characters config option which allows you specify which characters are in the hint set. It is still character driven but a little more flexible. I am not sure how your input system works but I am prepared to make the mode more configurable if you think you would benefit from it. 20 individual keys is more than sufficient to reach most screen locations but if you have some of them mapped to digraphs then you may need a 3 key hint system. mmm, I tried native X11 mousekeys, and it's really awkward setup for me - takes up several keysyms in a different modal state than the keys I use for warp, and leaves me with two different sets of buttons :-( I think what I'm after is similar to Keyboardio's behaviour (they can always move the mouse directly after detecting screen size through USB magic): an activation key that turns on mouse mode, enabling clicks and cardinal directional mouse movement, starting at the current pointer position a second activation key that turns on grid mode (like it does now) starting with the pointer at center. Hitting a directional mouse key turns off the grid and enters mouse mode from wherever the grid ended up. That seems more intuitive to me than the current approach of directionally moving the grid - it enables rapid coarse hops with the grid and small precise movements with directional. Make sense? an activation key that turns on mouse mode, enabling clicks and cardinal directional mouse movement, starting at the current pointer position Yes, this is what I originally understood. I imagine the main use case for what you describe would be the ability to select items close to the pointer after a long period of inactivity. Personally I can't think of too many instances in which this would be useful (although I admittedly rarely interact with multiple GUI elements). If you are sure you want a dedicated activation key for this it should be fairly easy to add I just want to confirm you've made full use of the existing local movement functionality. a second activation key that turns on grid mode (like it does now) starting with the pointer at center. Hitting a directional mouse key turns off the grid and enters mouse mode from wherever the grid ended up. That seems more intuitive to me than the current approach of directionally moving the grid - it enables rapid coarse hops with the grid and small precise movements with directional. Make sense? The current movement keys (wasd) move the grid by a fixed configurable interval which is the same thing as moving the pointer by a fixed interval. If I've understood you correctly then you simply want to preserve the existing functionality but hide the grid decorations, right? Personally I find using wasd in conjunction with uijk quite useful since I can often get away with making fewer adjustments for overshoot while I am still narrowing in on the item of interest than after shrinking the target area down to the size of pointer (though this is probably because I am using two hands on a traditional keyboard). This is facilitated by the fact that the grid is still accessible if I choose to use it, however I could just as easily use 'wasd' exclusively to move the pointer (at the center of the grid) to the target (which seems to be what you describe). an activation key that turns on mouse mode, enabling clicks and cardinal directional mouse movement, starting at the current pointer position I imagine the main use case for what you describe would be the ability to select items close to the pointer after a long period of inactivity. Sometimes it's that, and sometimes I miss the UI element by a few pixels from grid mode and then need to start over from the center. a second activation key that turns on grid mode (like it does now) starting with the pointer at center. Hitting a directional mouse key turns off the grid and enters mouse mode from wherever the grid ended up. That seems more intuitive to me than the current approach of directionally moving the grid - it enables rapid coarse hops with the grid and small precise movements with directional. Make sense? The current movement keys (wasd) move the grid by a fixed configurable interval which is the same thing as moving the pointer by a fixed interval. If I've understood you correctly then you simply want to preserve the existing functionality but hide the grid decorations, right? Hiding the grid would help... I'm somewhat agnostic on the current behaviour of directionally moving the entire grid, I think my instinct would be to get close with the regular grid and then move directionally after. There's a few different use cases for me here, but the common features are: directional mouse movement starting from current position, instead of recentering first hiding the grid during directional movement switching from grid mode to directional mode when a directional key is pressed As for activation, I could see two separate activation keys (for grid and directional).... ...or a single activation key that enters directional mode, followed by a key to switch to grid mode and/or jumping immediately like trigger_mods does (I use a 3x3 layout and would expect the central key to give me full screen grid - could probably do that by overloading it with the switch-to-grid key in config). This seems really nice, hope I described it ok. UX is complicated. :-) I imagine the main use case for what you describe would be the ability to select items close to the pointer after a long period of inactivity. Sometimes it's that, and sometimes I miss the UI element by a few pixels from grid mode and then need to start over from the center. This the what the movement keys are for. E.G As you can see the initial warp to the upper quadrant (u) overshoots the target at which point I move the cursor/grid down one unit (s) until the target is covered and then use grid keys to select it. Hiding the grid would help... I'm somewhat agnostic on the current behaviour of directionally moving the entire grid, I think my instinct would be to get close with the regular grid and then move directionally after. You can do this with the existing functionality. E.G There's a few different use cases for me here, but the common features are: directional mouse movement starting from current position, instead of recentering first hiding the grid during directional movement switching from grid mode to directional mode when a directional key is pressed As for activation, I could see two separate activation keys (for grid and directional).... I've implemented this in the latest commit. There is now an additional mode called 'discrete' mode which provides simple directional cursor movement without warping the pointer. It can be mapped to a dedicated activation key (as can hint mode) and used in conjunction with grid mode. P.S Some config names changed (notably the direction keys). See the readme/latest commit message for details. Will try, thanks!
What is the reason of this errror: "PermissionError: [WinError 5] Access is denied" I try to run this example code on Pycharm2018.3.3, it didn't work out. But the same code can run on the IDLE without any error. My environment is Python3.7 + windows10. from multiprocessing import Process, Queue def f(q): q.put([42, None, 'hello']) if __name__ == '__main__': q = Queue() p = Process(target=f, args=(q,)) p.start() print(q.get()) # prints "[42, None, 'hello']" p.join() Process Process-1: Traceback (most recent call last): File "C:\Users\WYM\AppData\Local\Programs\Python\Python37-32\lib\multiprocessing\process.py", line 297, in _bootstrap self.run() File "C:\Users\WYM\AppData\Local\Programs\Python\Python37-32\lib\multiprocessing\process.py", line 99, in run self._target(*self._args, **self._kwargs) File "E:\Spider2\ControlNode\test.py", line 4, in f q.put([42, None, 'hello']) File "C:\Users\WYM\AppData\Local\Programs\Python\Python37-32\lib\multiprocessing\queues.py", line 82, in put if not self._sem.acquire(block, timeout): PermissionError: [WinError 5] Access is denied. It would be good to know what the access was to. This issue may be a known bug with Python 3.7.2 I was experiencing the same issue and fixed it by creating a new virtual environment with Python 3.7.0 I have tested it with Python 3.7.4 and it works too.
Samuel Newby Curle Samuel Newby Curle FRSE (1930-1989) was a British mathematician. He served as Professor of Applied Mathematics at St Andrews University from 1967 until 1989. St Andrews University created the Curle Lecture in his memory. Life He was born in Sunderland, County Durham, on 18 June 1930 the son of Samuel Curle and his wife, Edith Newby Holmes. He attended Barnes School in Sunderland and then Bede Grammar School (now Sunderland College). He then attended Manchester University graduating BSc in 1951 and MSc in 1952. He received a PhD in 1955. In 1954 he began working at the National Physical Laboratory then in 1961 moved to the University of Southampton as a Reader in Mathematics. In 1967 he was given the Gregory Chair of Mathematics at St Andrews University. He was elected a Fellow of the Royal Society of Edinburgh in 1977. His proposers were Andrew G Mackie, Norrie Everitt, Douglas Samuel Jones and Donald C Pack. He died of heart disease on 27 June 1989 in Settle, North Yorkshire in England, and was buried in Kirkcaldy in Fife. Family In 1956 he married Shirley Kingsford Campion. The Curle Lecture In 2006 St Andrews University instituted the Curle Lecture in his memory. This is a biennial lecture on an entertaining way of presenting mathematical concepts. The first lecture 1089 and All That was given by Dr David Acheson of Jesus College, Oxford. Publications * The Laminar Boundary Layer Equations (1962) reprinted 1965
Method of fast coloring a polygon ABSTRACT A method of fast coloring a polygon for calculating areas of figures, the method comprising steps of resetting record values of all coordinate points to zero in the memory space; inputting the coordinate value of the ordered nodes defining the polygon into the memory space; connecting the nodes according to a predetermined order to determine connection points between the nodes; assigning a record value of each of the nodes as a flag value or a line value; assigning a record value of each of the connection points as a flag value or a line value; and calculating the area of the polygon by processing record values of the coordinate points in the memory space according to the record values of the nodes and the record values of the connection points. The method of fast coloring a polygon of the present invention is implemented with ordinary instead of complicated hardware configurations. BACKGROUND OF THE INVENTION 1. Field of the Invention The present invention generally relates to a method of fast coloring, in particular, to method of fast coloring a polygon. 2. Description of Prior Art In a digitized world, image processing has becoming a popular area along with the development of computer technologies. Image processing refers to the processing image data of images from digital cameras or from scanned photos taken by optical cameras. Similar to prior art optical processing in a dark room, computer applications are installed in a computer to process image data and generate desired image results. Such image processing applications are also called digital dark room, where users are allowed to receive desirable image results with simple procedures at a lower cost without the need to hide in the dark, have frequent contacts with chemical materials. More to that, image processing applications is advantageous in creating complicated image effects, providing massive duplications in hard copies, as well as directly applicable to printing, and convenient to distribute on the internet. The technology for area calculating is a critical part of image processing technologies. Typically, image data is massive and large demand in computing capability and processing time. Under the limitation of cost and hardware process accelerating capacity, software can be an effective approach to increase the processing speed. The processing time of the conventional technology for area calculating is in effective in terms of time and required memory for image processing, which is not suited for implementing at the hardware level. Therefore, it is advantageous to an image process user by offering a rapid method of fast coloring a polygon via increasing calculating speed implemented with a resource limited hardware configuration such as human machine interface instead of complicated hardware configuration. SUMMARY OF THE INVENTION Therefore it is an object of the present invention to provide a method to overcome the disadvantages of the aforementioned prior art such as limited calculation speed. The method according to the present invention can be implemented with ordinary hardware configuration instead of a complicated hardware configuration. To realize the above object, the major steps of the method of fast coloring a polygon comprise: resetting record values of all coordinate points to zero in said memory space; inputting the coordinate value of said ordered nodes defining said polygon into the memory space; connecting said nodes according to a predetermined order to determine connection points between said nodes; assigning a record value of each of said nodes as a flag value or a line value; assigning a record value of each of said connection points as a flag value or a line value; and calculating said area of said polygon by processing record values of said coordinate points in said memory space according to said record values of said nodes and said record values of said connection points. The method for fast coloring a polygon according to the present invention can be implemented with ordinary hardware configuration instead of a complicated hardware configuration. DETAILED DESCRIPTION OF THE INVENTION Drawings from FIG. 7 to FIG. 11 are perspective views illustrating embodiments using the method of fast coloring a polygon according to an embodiment of the present invention. The purpose of the method of fast coloring a polygon of the present invention is to calculate colored area of polygons shown from FIG. 7 to FIG. 11. The area of the polygons from FIG. 7 to FIG. 11 is defined by ordered nodes. The method of fast coloring a polygon of the present invention is further described in details in the following. FIG. 1 is a flow chart of the method of fast coloring a polygon according to the present invention. First, the record values of all coordinate points in a memory space pattern (the size of the memory space is by R rows multiple by C columns) is reset to zero at step S10 so as to proceed to the following step. FIG. 12 is a perspective view illustrating the memory space according to the present invention. The memory space comprises matrix memorizing units a plurality of coordinate points (each record value representing a pixel). Given the memory space comprises coordinate points of C rows (at X axis)×R(at Y axis) columns, the coordinate points of the row varies as the index of a row varies. The following steps of the present invention adapt the means to process by rows, it should be noted that the means to process by columns are also applicable to an embodiment according to the scope of the present invention. More to that, the coordinate points increment follow the sequence from left to right at X axis and from top to bottom at Y axis in the memory space shown in the FIG. 12. In the FIG. 1, the coordinates of nodes of a polygon (N-sided shape) is input into the memory space as (X₁, Y₁) . . . (X_(N), Y_(N)) at step S20. Following that, the nodes are connected by a predetermined order to determine the connection points between two nodes at step S30. For examples, node (X₁, Y₁) is connected to node (X₂, Y₂) and node (X₂, Y₂) is connected to node (X₃, Y₃) . . . till the last node (X_(N), Y_(N)) is connected to the first node (X₁, Y₁). At step S40, the record value of each node is assigned as a flag value or a line value (described in the following). At step S50, the record value of connecting nodes is assigned as a flag value or a line value (described in the following). At step S60, calculate the area of the polygon by processing record values of the coordinate points in the memory space according to the record values of the nodes and the record values of the connection points (described in the following). In the FIG. 4, each memory unit is called a coordinate point in the arrays of the memory space. The polygon is defined by a plurality of ordered nodes, which are labeled as points m, m+1. In addition, the coordinate points defined by the connected line between two nodes are called connection points for calculating the area of the polygon. The connection points are the points marked with shadowed background in the FIG. 4. FIG. 2 is a flow chart of assigning a record value of each of the nodes as a flag value or a line value according to the present invention and used for describing details of that the record value of each node is assigned as a flag value or a line value at step S40. Given that the coordinates of the m node, the current node, under the process is (X_(m), Y_(m)), the coordinates of the previous node is (X_(m−1), Y_(m−1)), and the coordinates of the next node is (X_(m+1), Y_(m+1)). XY coordinates are the conventional XY coordinate system. If m node is the first node, then the coordinates for the precious node (X_(m−1), Y_(m−1)) of the first node is (X_(N), Y_(N)). At step S420, the method determines if the current node, the precious node and the next node form a reflection by the following conditions: (Y _(m) >=Y _(m+1) and Y _(m) <Y _(m−1))   (1) (Y _(m) <Y _(m+1) and Y _(m) >=Y _(m−1))   (2) Given that either of the above conditions is true, that means the current node, the precious node and the next node does not form a reflection. The method moves to step S440 to assign the record value of the current node (X_(m), Y_(m)) as a flag value. If not true, that means the precious node and the next node form a reflection. The method moves to step S460 to assign the record value of the current node (X_(m), Y_(m)) as a line value. The flow of the node process characteristics can be expressed in the program language as the following:  if((Y_(m) >= Y_(m+1) and Y_(m) < Y_(m−1)) or (Y_(m) < Y_(m+1) and Y_(m) >= Y_(m−1))) {  Pattern[X_(m)][Y_(m)] := FLAG; } else {  Pattern[X_(m)][Y_(m)] := LINE; } FIG. 3 is a flow chart of assigning a record value of each of the connection points between two nodes as a flag value or a line value according to the present invention and used for describing details of that the record value of each connection point is assigned as a flag value or a line value at step S50. In other words, the method processes the record value of the connection point at the connected line between the current node and the next node (X_(m+1), Y_(m+1)). First, at step S510, the method determines if the coordinate Ym of the current node does not equals to the coordinate Y_(m+1) of the next node. If not true, that means the row coordinate Ym equals to the row coordinate Y_(m+1), the method moves to step S590 and terminates the process. If true, that means the current node and the next node is not in the same row, the method moves to step S520 to determine if the row coordinate Ym is more than the row coordinate Y_(m−1). If true, the method moves to step S530 to process the (Y_(m))−1 row to the (Y_(m−1))+1 row. If not true, then the method moves to step S540 to process the (Y_(m+1))−1 row to the (Y_(m))+1 row. The purpose of the above flow is to process the record value of the connection point on the connected line between the node (X_(m), Y_(m)) and the node (X_(m+1), Y_(m+1)), but not including the nodes (X_(m), Y_(m)) and (X_(m+1), Y_(m+1)). Then at step S550, the method determines if the record value of the first connection point of the next row is a flag value. In other words, if the connection point between (X_(m), Y_(m)) and (X_(m+1), Y_(m+1)) of the next row, which is the record value of the first connection point of the next row, is a flag value. If true, then the method moves to step S560 to assign the record value of the first connection point of the next row is a line value. If not true, the method moves to step S570 to assign the record value of the first connection point of the next row as a flag value. The purpose of the flow is to assign the record value of the first connection point of the next row as a flag value. However, if the record value of the first connection point of the next row is found already a flag value, then the method change the record value of the first connection point of the next row as a line value. At step S580, the method assigns the record values of the rest of the connection points of the next row as line values. If every connection point between two nodes of the polygon is processed by the steps demonstrated in the FIG. 3, the result is equivalent to the result after performing the step S50 demonstrated in the FIG. 1, the step to assign record values of connection points. FIG. 4 is a perspective view illustrating assigning a record value of each of the connection points as a flag value or a line value according to an embodiment of the present invention and used for giving further details of the above process. Each section shown in the drawing represents a location of the memory, where the characteristics of the connection points between node (X_(m), Y_(m)) and node (X_(m+1), Y_(m+1)) is determined. In the FIG. 4, notion m represents node (X_(m), Y_(m)), notion m+1 represents node (X_(m+1), Y_(m+1)), notion F represents a flag value, and notion L represents a line value. As shown in the drawing, the record value of the first connection point of the connection points on the connected line between (X_(m), Y_(m)) to (X_(m+1), Y_(m+1)) of the next row is assigned as a flag value F, and the record values of the rest of the connection points are assigned as line value L. FIG. 5 is a perspective view illustrating assigning a record value of each of the connection points as a flag value or a line value according to another embodiment of the present invention. The purpose of the drawing is to illustrate that the record value of the first connection point of the next row is assigned as a line value at step S560. The details are described in the flow used in the FIG. 4. It should be noted that the record value of the first connection point of the fourth row is a flag value F in the FIG. 4. Nonetheless, the record value of the first connection point of the next row is a flag value at step S550 in the FIG. 5, the method moves to step S560 to assign the record value of the first connection point of the next row as a line value L, which is different from the record value of the first connection point of the next row. After the above flow of assigning the record value of the connection points between the nodes as a flag value or a line value, the method move to calculate the area formed by the lines connected between nodes by a predetermined order (i.e. the colored area of the polygons from FIG. 7 to FIG. 11). FIG. 6 is a flow chart of processing the record values of a row of the polygon according to the present invention and used for describing in details of calculating the area of the polygon by processing record values of the coordinate points in the memory space according to the record values of the nodes and the record values of the connection points. For examples, calculate the colored area by calculating the sum of the record values when the record values are a specific value (such as logic value 1). The process method is to start the processing from each coordinate point of the first row of the memory, then move to process the second row, followed by processing the third row. At the same time, when processing each row, the process follows a left to right order on each coordinate. In the FIG. 6, the steps of the method performed on each row are started from step S610, then move to step S620 to determine if record value of the current coordinate point is a flag value. If true, then the method moves to step S630 to start coloring (or starting when the record value of the coordinate point is 1). The process method is to start the processing from each coordinate point of the first row of the memory, then move to process the second row, followed by processing the third row. At the same time, when processing each row, the process follows a left to right order on each row. In the FIG. 6, the steps of the method performed on each row are started from step S610, then move to step S620 to determine if record value of the current coordinate point is a flag value. If true, then the method moves to step S630 to start coloring (coloring can be implemented, for example, by continuously assigning logic value 1 to the record value of the current coordinate point). The processing of coloring (namely, assigning logic value 1 to the record value) continues till another flag value is found in the remaining coordinate points of the same row. When coloring is stopped, logic value 1 is no longer assigned to the successive coordinate points for the same row. Then the process moves back to step S610 to start again. On the other hand, if none of the record value of the remaining coordinate points of the same row is flag value, then the method moves to step S640 to determine if it is the end of the row processing. If yes, then the method moves to step S650 to terminate the row processing. If not, then the method returns to step S610 to start again. Accordingly, as all the coordinate points by each row in the memory are processed, the area of the polygon is calculated by summarizing the total quantity of colored coordinate points in all columns. The flow of processing values of the polygon according to the present invention can be expressed in the following program language (NC is the width of the memory space and NR is the length of the memory space): for(j:=0;j< NR;j++) {  DRAW_FLAG := 0;  for(i:=0;i<NC;i++)  {   if(Pattern[i][j] == FLAG)   {    DRAW_FLAG := abs(DRAW_FLAG −1);   }   if(DRAW_FLAG == 1 AND Pattern[i][j] == 0)   {    Pattern[i][j] = DRAW;   }  } } As mentioned above, FIG. 7 to FIG. 11 are perspective views of embodiments, which illustrate the method of fast coloring a polygon according to the present invention. The nodes are connected according to the order of the numbers. The colored area is calculated by following the method of fast coloring a polygon according to the present invention. As the skilled person will appreciate, various changes and modifications can be made to the described embodiments. It is intended to include all such variations, modifications and equivalents which fall within the scope of the invention, as defined in the accompanying claims. BRIEF DESCRIPTION OF DRAWING The features of the invention believed to be novel are set forth with particularity in the appended claims. The invention itself, however, may be best understood by reference to the following detailed description of the invention, which describes an exemplary embodiment of the invention, taken in conjunction with the accompanying drawings, in which: FIG. 1 is a flow chart of the method of fast coloring a polygon according to the present invention; FIG. 2 is a flow chart of assigning a record value of each of the nodes as a flag value or a line value according to the present invention; FIG. 3 is a flow chart of assigning a record value of each of the connection points between two nodes as a flag value or a line value according to the present invention; FIG. 4 is a perspective view illustrating assigning a record value of each of the connection points as a flag value or a line value according to an embodiment of the present invention; FIG. 5 is a perspective view illustrating assigning a record value of each of the connection points as a flag value or a line value according to another embodiment of the present invention; FIG. 6 is a flow chart of processing the record values of a row of the polygon according to the present invention; FIG. 7 is a perspective view illustrating the method of fast coloring a polygon according to an embodiment of the present invention; FIG. 8 is a perspective view illustrating the method of fast coloring a polygon according to an embodiment of the present invention; FIG. 9 is a perspective view illustrating the method of fast coloring a polygon according to an embodiment of the present invention; FIG. 10 is a perspective view illustrating the method of fast coloring a polygon according to an embodiment of the present invention; FIG. 11 is a perspective view illustrating the method of fast coloring a polygon according to an embodiment of the present invention; and FIG. 12 is a perspective view illustrating the memory space according to the present invention 1. A method of fast coloring a polygon used in a memory space for calculating an area of the polygon in the memory space, wherein the memory space comprises a plurality of coordinate points matrix in an X-Y coordinate system, the polygon is defined by a plurality of ordered nodes, said method comprising: (A) resetting record values of all coordinate points to zero in said memory space; (B) inputting the coordinates of said ordered nodes defining said polygon into the memory space; (C) connecting said nodes according to a predetermined order to determine connection points between said nodes; (D) assigning a record value of each said node as a flag value or a line value; (E) assigning a record value of said connection points as a flag value or a line value; and (F) calculating said area of said polygon by processing record values of said coordinate points in said memory space according to said record values of said nodes and said record values of said connection points. 2. The method of fast coloring a polygon of claim 1, wherein said step (D) further comprises: (D1) assigning a flag value to the record value of the current node if the Y coordinate value of the current node is not less than the Y coordinate value of the next node and Y coordinate value of the current node is less than the Y coordinate value of the previous node. 3. The method of fast coloring a polygon of claim 2, wherein said step (D) further comprises: (D2) assigning a flag value to the record value of the current node if the Y coordinate value of the current node is less than the Y coordinate value of the next node and Y coordinate value of the current node is not less than the Y coordinate value of the previous node. 4. The method of fast coloring a polygon of claim 3, wherein said step (D) further comprises: (D3) assigning a line value to the record value of the current node which is not assigned a flag value. 5. The method of fast coloring a polygon of claim 4, wherein said step (E) further comprises: (E1) processing the connection point between two nodes by rows; and (E2) assigning a flag value to the first connection point if the record value of the first connection point of the following row is not a flag value. 6. The method of fast coloring a polygon of claim 5, wherein said step (E) further comprises: (E3) assigning a line value to the first connection point if the record value of the first connection point of the following row is a flag value. 7. The method of fast coloring a polygon of claim 6, wherein at said step (F) the process order follows an order of from left to right and from top to bottom, and said step (F) further comprises: (F1) initiating coloring if the record value of a connection point of said processing row is a flag value, and stop coloring if the record value of the following connection point is also a flag value. 8. The method of fast coloring a polygon of claim 7, wherein said step (F) further comprises: (F2) determining the ending point of said processing row if the record values of the coordinate points of said processing row are not flag values; and (F3) terminating processing of the processing row if processing the end point of said processing row. 9. The method of fast coloring a polygon of claim 8, wherein said step (F) further comprises: (F4) calculate said area of said polygon by summarizing the total quantity of specific record values of coordinate points in said memory space. 10. The method of fast coloring a polygon of claim 9, wherein said specific record values are record values corresponding to logic value 1
import { runCommand } from '../utils'; import { ROOT_PATH } from '../consts'; /** * Get the commands that should be run * * @export * @param {InstallObject} webpackSetup - Webpack install object * @param {InstallObject} dependencySetup - Dependency install object * @param {string} packageManager - Package manager that should be used (npm or yarn) * @returns {Object} Install and install dependency commands */ function getCommands(webpackSetup, dependencySetup, packageManager) { if (packageManager === 'yarn') { return { install: `yarn add ${webpackSetup} ${dependencySetup}`, installDeps: 'yarn', }; } return { install: `npm install ${webpackSetup} ${dependencySetup}`, installDeps: 'npm install', }; } /** * Installs the Webpack and dependency modules * * @export * @param {InstallObject} webpackSetup - Webpack install object * @param {InstallObject} dependencySetup - Dependency install object * @param {string} packageManager - Package manager that should be used (npm or yarn) * @param {Object} options - Runner options * @returns {Promise} A promise indicating the installation success */ export default async function (webpackSetup, dependencySetup, packageManager, options) { const commands = getCommands(webpackSetup, dependencySetup, packageManager); // IMPORTANT: ROOT_PATH needs to contain a package.json file // If it doesn't exist, npm will search for the closest package.json and install into this folder const stdout = await runCommand(commands.install, ROOT_PATH, {}, options); if (stdout.indexOf(webpackSetup.toLocalName()) === -1 || stdout.indexOf(dependencySetup.toLocalName()) === -1) { const error = new Error(); error.err = ['Expected versions not in dependency tree', stdout]; throw error; } return runCommand(commands.installDeps, dependencySetup.installLocation, {}, options); }
package common import ( "fmt" ) type errScope string const ( errUnknown errScope = "" errConf = "conf_error" errDB = "db_error" errLogin = "login_error" errAccount = "account_error" errTushare = "tushare_error" errMinio = "minio_error" ) var ( // Conf error // DB error ErrDBEmptyUsername = makeError(errDB, "missing Username") ErrDBEmptyPassword = makeError(errDB, "missing Password") ErrDBEmptyHost = makeError(errDB, "missing Host") ErrDBEmptyPort = makeError(errDB, "missing Port") ErrDBEmptyDatabase = makeError(errDB, "missing Database") // Login error ErrorInvalidUsernameOrPassowrd = makeError(errLogin, "invalid username or password") // Account error ErrorEmptyAccountUsernameOrPassowrd = makeError(errAccount, "username or password is empty") // Tushare error ErrTushareURL = makeError(errTushare, "url is empty.") ErrTushareToken = makeError(errTushare, "token is empty.") ErrTushareCodeEmpty = makeError(errTushare, "code is empty.") ErrTushareStockFieldsLen = makeError(errTushare, "length of stock fields is not equal to headers.") ErrTushareDailyParamsEmpty = makeError(errTushare, "daily params is empty.") ErrTushareDailyFieldsUnknown = makeError(errTushare, "unknown daily field.") ErrTushareConceptDetailFieldsUnknown = makeError(errTushare, "unknown concept detail field.") // Minio error ErrMinioAccessKeyEmpty = makeError(errMinio, "access_key is empty") ErrMinioSecretKeyEmpty = makeError(errMinio, "secret_key is empty") ErrMinioEndPointEmpty = makeError(errMinio, "endpoint is empty") ) func makeError(scope errScope, msg ...string) error { return fmt.Errorf("[%s]: %s", scope, msg) }
Liquid cooled hybrid ABSTRACT An inflator that includes a quantity of gas generant housed within a chamber. A liquid and a piston are also housed within another chamber. This chamber is sealed by a burst disk. The piston includes an opening. During deployment of the inflator, the burst disk is ruptured and the piston moves towards the distal end of the inflator. The movement of the piston hydraulically expels the liquid through the opening in the piston such that the liquid cools and contacts the gas formed by combustion of the gas generant. BACKGROUND OF THE INVENTION Airbags and airbag systems have been known and used for decades and have been credited with saving many lives and preventing many injuries. Airbag systems are used to deploy an airbag during a crash into a vehicle interior so that the vehicle occupant will impact the deployed airbag rather than the harder surface(s) of the vehicle interior. Airbag systems generally include an inflator. An inflator is a device that is designed to produce and/or channel a quantity of inflation gas into the airbag during the deployment process. This influx of gas into the airbag inflates and deploys the airbag into the proper position. A variety of different inflators are known in the industry. One type of known inflator is the so-called “pyrotechnic” inflator. Pyrotechnic inflators include a quantity of gas generant that is combusted to produce the inflation gas necessary to deploy the airbag. Accordingly, pyrotechnic inflators generally include the quantity of gas generant housed within a housing (i.e., a pressure vessel). Pyrotechnic inflators also generally include a filter. The filter is designed to filter the gas to remove solids entrained in the gas stream. The filter is also designed to cool the gas. However, the filter is often expensive and adds significant cost and weight to the inflator. The automotive industry continues to demand airbag systems that are smaller, lighter, and less expensive to manufacture. In fact, as vehicles become smaller and more compact, such changes to the airbag systems are necessary in order to meet the constraints of these smaller cars. The airbag inflator is a significant component of the airbag system. Accordingly, reducing the size, weight, and/or cost of the inflator would result in significant size, weight, and/or cost savings in the overall airbag system. Accordingly, the present embodiments relate to a new type of inflator that does not include a filter, but may still operate to cool the gas and/or remove entrapped particles from the gas stream. The removal of the filter from the airbag inflator makes this device smaller, lighter, and less expensive to manufacture. BRIEF SUMMARY OF THE INVENTION The inflator comprises a quantity of gas generant housed within a gas generant chamber. The gas generant chamber has a constant volume. The inflator also comprises an initiator to ignite the gas generant and form gas during deployment. A chamber is also provided to house a piston and a liquid, wherein the chamber is sealed by a burst disk or a seal, wherein during deployment the burst disk or seal is unsealed and the piston moves and hydraulically expels the liquid through an opening in the piston such that the liquid contacts and cools the gas formed by combustion of the gas generant. In some embodiments, the inflator may be fully or partially filterless. Other embodiments may be designed in which a diffuser is provided, the gas formed by combustion of the gas generant flows through the diffuser. In some embodiments, the burst disk will be ruptured whereas in other embodiments, the seal is press-fit against the chamber and this seal is unsealed by the movement of the piston. In some embodiments, one or more baffles may be used. These baffles may be provided on the piston. In other embodiments, the gas formed by combustion of the gas generant and the liquid move into the interior of the piston, wherein an exit opening is provided to allow passage to the exterior of the piston. In some embodiments, the liquid hydraulically expelled through the opening vaporizes and mixes with the gas formed by combustion of the gas generant. Additional embodiments may be constructed in which the gas formed by combustion of the gas generant flows through a tortuous path prior to exiting the inflator. This tortuous path may have right angle turn(s) for capturing particulates entrained in the gas. In further embodiments, as the gas flows through the tortuous path, any particulates entrained in the gas are removed from the gas and deposited within the inflator. The particulates may be deposited at a capture area. A method of cooling gas formed in an inflator may also be disclosed. For this method, the inflator comprises a quantity of gas generant housed within a gas generant chamber, an initiator, and a piston chamber that houses a piston and a liquid. The method comprises the step of igniting the gas generant to form gas. An additional step of rupturing a burst disk used to seal the piston chamber is also performed. An additional step of moving the piston may also be performed. The movement of the piston operates to hydraulically expel the liquid through an opening in the piston such that the liquid contacts and cools the gas formed by combustion of the gas generant. The present embodiments relate to inflators that include a quantity of gas generant housed within a housing. The gas generant may be ignited to produce a quantity of inflation gas. This gas may then be channeled into an airbag to deploy the airbag. The inflator also includes gas flow openings in the housing. As will be explained herein, when the inflator is deployed, gas may flow out of the gas flow openings so that it may be channeled into the airbag. The inflator will further comprise a piston that is housed within a chamber. (The chamber is within the housing). Also housed within the chamber is a quantity of liquid. The piston may also have an opening that is sealed by a burst disk. When the opening is sealed, the liquid cannot escape through the opening. The combustion of the gas generant produces inflation gas that flows through the diffuser and contacts a piston that has been added to the inflator. This piston is housed within a chamber. When the gas enters this chamber, it pressurizes the chamber. At the same time, some of the gas may also begin to exit the inflator. The chamber housing the piston also includes a quantity of liquid. When this chamber is pressurized by the influx of gas into the chamber, liquid begins to flow through an opening in the piston. This liquid will then mix with the gas. When the liquid contacts the gas, at least some of the liquid is vaporized into a gas stream. Obviously, this vaporization process is endothermic and operates to cool the gas. Thus, by using a system that has evaporating liquid, the inflation gas may be cooled without the use of an expensive filter. Further, this vaporization of the liquid increases the amount of gas within the chamber. Thus, by using the liquid, the amount of gas generant necessary to produce sufficient inflation gas is reduced, further reducing the size and cost of the inflator. Thus, by using the liquid injection technique described above, the present embodiments do not require the use of a filter to cool the gas. The pressure required to inject the liquid into the gas stream may be provided by the combustion pressure of the inflator. In other embodiments, the particulates in the gas produced by ignition of the gas generant may be removed without the use of a filter. Specifically, at least part of the particulate removal function of the filter is accomplished by turning the gas flow significantly prior to exiting the inflator. The gas produced by vaporization or decomposition of the liquid contributes to the airbag inflation. It should also be noted that the present embodiments do not necessarily have to be filterless. A filter may also be used. However, the system may be “partially” filterless as the size and type of the filter needed may be reduced. BRIEF DESCRIPTION OF THE SEVERAL VIEWS OF THE DRAWINGS In order that the manner in which the above-recited and other features and advantages of the invention are obtained will be readily understood, a more particular description of the invention briefly described above will be rendered by reference to specific embodiments thereof which are illustrated in the appended drawings. Understanding that these drawings depict only typical embodiments of the invention and are not therefore to be considered to be limiting of its scope, the invention will be described and explained with additional specificity and detail through the use of the accompanying drawings in which: FIG. 1 is a sectional view of an embodiment of an inflator, the inflator being shown prior to deployment; FIG. 2 is a sectional view of the embodiment of FIG. 1, the inflator being shown as the initiator is being actuated and the burst disk ruptures; FIG. 3 is a sectional view of the embodiment of FIG. 1, the inflator being shown as the liquid injection is occurring; FIG. 4 is a sectional view of the embodiment of FIG. 1; the inflator being shown as the gas escapes the inflator; FIG. 5 is a sectional view of the embodiment of FIG. 1, the inflator being shown near the completion of the liquid injection; FIG. 6 is a sectional view of the embodiment of FIG. 1, the inflator being shown after the inflator has been fully deployed; FIG. 7 is a sectional view of another embodiment of an inflator; FIG. 8 is a sectional view of another embodiment of an inflator; FIG. 9 is a sectional view of another embodiment of an inflator; FIG. 10 is a sectional view showing a baffle that may be used as part of the embodiment of FIG. 9; FIG. 11 is a sectional view of another embodiment of an inflator; FIG. 11A is an enlarged version of the piston shown in FIG. 11; and FIGS. 12 through 14 are sectional views that show the stages of deployment of the inflator of FIG. 11. DETAILED DESCRIPTION OF THE INVENTION The presently preferred embodiments of the present invention will be best understood by reference to the drawings, wherein like parts are designated by like numerals throughout. It will be readily understood that the components of the present invention, as generally described and illustrated in the figures herein, could be arranged and designed in a wide variety of different configurations. Thus, the following more detailed description of the present embodiments, as represented in the Figures, is not intended to limit the scope of the invention, as claimed, but is merely representative of presently preferred embodiments of the invention. Referring now to FIG. 1, a sectional view of an inflator 100 is illustrated. The inflator includes a quantity of gas generant 104 housed within a housing 108. An initiator 112 is also added to the inflator 100. The initiator 112 is used to ignite the gas generant 104. When the gas generant 104 is ignited, a quantity of inflation gas is formed. This gas may then be channeled into an airbag (not shown) to deploy the airbag. Those skilled in the art will appreciate that initiators 112 and gas generant 104 are known in the art and that a variety of different features may be used for these components. The generant 104 is housed within a chamber 116 that is sealed by a burst disk 120. The chamber 116 is sometimes referred to as a “gas generant chamber.” The gas generant chamber 116 has a constant volume. Before, during, and after deployment of the inflator, the volume of the gas generant chamber 116 remains the same. A diffuser 124 is also positioned in the chamber 116. Specifically, when the initiator 112 is activated, it will produce hot gas that will flow through holes 128. This hot gas will then contact the generant 104 and ignite the generant 104. In turn, this ignition of the generant 104 creates a supply of gas that will pass through the holes 132 to the interior of the diffuser 124 and will then contact the burst disk 120. The increase in pressure within the chamber 116 caused by ignition of the generant 104 will rupture the burst disk 120 and allow the gas to exit the chamber 116. Referring still to FIG. 1, the inflator 100 will further comprise a piston 136 that is housed within a chamber 140. (The chamber 140 is within the housing 108). Also housed within the chamber 140 is a quantity of liquid 144. The piston 136 may also have an opening 148 that is sealed by a burst disk 152. When the opening 148 is sealed, the liquid 144 cannot escape through the opening 148. The liquid 144 may be any liquid that remains a liquid between −35 degrees Celsius and 85 degrees Celsius. The liquid must also be capable of vaporizing endothermically and, when vaporized, produce a gas that is within acceptable effluent limits associated with airbags. Also, the liquid may be non-corrosive so that it may be stored in a simple chamber. Any liquid that will meet these criteria may be used as the liquid 144. An example of a liquid that meets such criteria includes water mixed with CaCl₂. The inflator 100 of FIG. 1 also includes gas flow openings 156 and vent holes 157 in the housing 108. As will be explained herein, when the inflator 100 is deployed, gas may flow out of the gas flow openings 156 so that it may be channeled into the airbag (not shown). The gas flow openings 156 may or may not be sealed by burst disks prior to deployment. As noted above, FIG. 1 shows the inflator 100 prior to deployment. With reference to FIGS. 1 through 6, the deployment of the inflator 100 will now be described. FIG. 2 is a sectional view of the embodiment of FIG. 1 which shows actuation of the initiator 112. When the initiator 112 is actuated, hot gas will be produced and allowed to pass through the holes 128, thereby contacting the gas generant 104. Such contact with the gas generant 104 ignites and com busts the gas generant 104 into a quantity of inflation gas. FIG. 3 shows the inflator 100 as the gas generant 104 is combusted. As explained above, the combustion of the gas generant 104 produces inflation gas 160 (represented graphically by arrows) that flows through the holes 132 in the diffuser 124 and contacts the burst disk 120. The production of the gas 160 pressurizes the chamber 116 and causes the burst disk 120 to rupture. Once ruptured, the gas 160 will exit the chamber 116 and will enter the chamber 140. When the gas 160 is in the chamber 140, it will contact the piston 136 and pressurize the interior of the chamber 140. Some of the gas 160 may also begin to exit the inflator 100 via the openings 156. However, the pressurizing of the chamber 140 will, in turn, cause the burst disk 152 (shown in FIG. 1) to rupture. Once ruptured, liquid 144 begins to be injected through the opening 148 (which is no longer sealed by the disk 152) and mix with the gas 160. As shown in FIG. 3, the inflator 100 has an impact area 179 which is the area of the piston 136 that contacts the liquid 144. The inflator 100 also has a drive area 181, which is the area of the piston 136 that the gas 160 contacts. The impact area 179 is smaller than the drive area 181. The pressure of the liquid 144 is amplified due to the differences in areas 181 and 179. The liquid pressure is approximately equal to the pressure of the gas 160 times the ratio of the area 181 to area 179. This pressure differential causes the liquid 144 to inject into the inside region of the piston 136 and thereby interact with the gas 160. FIG. 4 shows the deployment of the inflator 100 as the liquid 144 is being injected through the opening 148 and mixing with the gas 160. The pressure within the chamber 140 pushes against the piston 136 and causes the piston to move towards the liquid 144. This hydraulic pressure on the liquid 144 forces more of the liquid 144 through the opening 148. Of course, while this is occurring, gas 160 continues to enter the chamber 140 from the chamber 116. Some of the gas also continues to exit via the openings 156. It should be noted that when the liquid 144 contacts the gas 160, at least some of the liquid 144 is vaporized into a gas stream. Obviously, this vaporization process operates to cool the gas 160. (Specifically, the heat required to vaporize or decompose the liquid 144 is removed from the gas stream and significantly cools the exiting gas 160). It should be noted that a filter is often used to cool the inflation gas. However, by using the liquid 144, the gas 160 may be cooled without the use of an expensive filter. Further, this vaporization of the liquid 144 increases the amount of gas within the chamber 140. Thus, by using the liquid 144, the amount of gas generant 104 necessary to produce sufficient inflation gas is reduced, further reducing the size and cost of the inflator 100. FIG. 5 shows the inflator 100 after the piston 136 has completely moved. Specifically, the piston 136 is moved to the distal end 164 of the chamber 140, thereby forcing all of the liquid 144 to pass through the opening 148. Again, as noted above, this liquid 144 is vaporized into a gas. However, even after the liquid 144 has been fully injected, gas 160 is still exiting the inflator 100 via the openings 156. FIG. 6 shows the inflator 100 after the deployment process is complete. The liquid 144 has been fully converted into gas and has been emptied out of the chamber 140. The gas 160 produced by ignition of the gas generant 104 has also been fully discharged. FIG. 7 is a sectional view of an additional embodiment of an inflator 200. The inflator 200 is similar to the embodiment of FIG. 1. For purposes of brevity, this discussion will not be repeated. The main difference between the inflator 200 and the inflator 100 is that the inflator 200 does not include a diffuser 124. Rather, the inflator 200 simply has openings 232 that the gas 260 will pass through, after it has been formed from ignition of the gas generant 104. When the gas 260 passes through the openings 232, it will flow, as indicated by the arrows, through a tortuous path. More specifically, the gas 260 will flow past the corner 264 such that the gas flow path will bend. When the gas flow bends in this manner, the entrained particulates and other solids found within the gas 260 will be separated from the gas 260 and will be deposited proximate the corner 264. Thus, the entrained particulates are removed from the gas 260 flow without the use of an expensive filter or diffuser. As shown in FIG. 7, a piston 236 is used in the inflator 200. The piston 236 is hollow and is within a chamber 240. Accordingly, the gas 260 leaving the chamber 116 will flow into the interior of the piston 236. The piston 236 includes exit openings 268 that will allow the gas to flow to the exterior of the piston 236 and then exit the inflator 200 via openings 156. The piston 236 also includes an opening 148 that may or may not be sealed by a burst disk 152. Again, the pressure caused by the gas will move the piston 236 towards the distal end 164 and will inject the liquid 144 through the opening 148. Once injected, the liquid 144 will vaporize and cool the gas 260 in the manner described above. During deployment of the inflator 200, the piston 236 may be fully displaced so that all of the liquid 144 may be fully forced through the opening 148. As shown in FIG. 7, the inflator 200 has an impact area 179 which is the area of the piston 136 that contacts the liquid 144. The inflator 200 also has a drive area 181, which is the area of the piston 136 that the gas 160 contacts. The impact area 179 is smaller than the drive area 181. Referring now to FIG. 8, a sectional view of an inflator 300 is illustrated. The inflator 300 is similar to the embodiments discussed above. For purposes of brevity, this discussion will not be repeated. Like the embodiment shown above, the inflator 300 does not include a diffuser. Rather, the embodiment of FIG. 8 includes openings 332 through which the gas 360 (produced by ignition of the generant 104) may exit the chamber 116. As with the embodiment discussed above, the gas 360, upon exiting the chamber 116, will engage in a tortuous path, passing two or more corners 364. Such corners 364 are right angle turns that will receive the entrained particulates. In other words, when the gas 360 turns at the corner 364, the entrained particulates will be separated out of the gas and deposited at a capture area 366. Generally, this capture area 366 is a corner or uneven surface that facilitates deposition. Thus, the entrained particulates are removed from the gas 360 without the use of an expensive filter or diffuser. The inflator 300 also includes a piston 336. The piston 336 is hollow and is within a chamber 340. Accordingly, the gas 360 leaving the chamber 116 will flow into the interior of the piston 336. The inflator 300 also includes openings that will allow the gas to flow to the exterior of the piston 336 and then exit the inflator 300 via openings 156. The piston 336 also includes an opening 148 that may or may not be sealed by a burst disk 152. Again, the pressure caused by the gas will move the piston 336 towards the distal end 164 and will inject the liquid 144 through the opening 148. Once injected, the liquid 144 will vaporize and cool the gas 360 in the manner described above. During deployment of the inflator 300, the piston 336 may be fully displaced so that all of the liquid 144 may be fully forced through the opening 148. The inflator 300 has an impact area 179 and a drive area 181. The impact area 179 is smaller than the drive area 181. Referring now to FIG. 9, another embodiment of an inflator 400 is illustrated. The inflator 400 is similar to the embodiments discussed above. For purposes of brevity, this discussion will not be repeated. As with the embodiment of FIG. 1, the inflator 400 includes a diffuser 124. As described above, gas 460 produced by the ignition of the generant 104 will flow through the holes 132 and rupture the burst disk 120 and then exit the chamber 116. Upon exiting the chamber 116, the gas will be allowed to exit the inflator 400 via the openings 156. The gas will also access the interior of the piston 436 and may move the piston 436, thereby forcing the liquid 144 through the opening 148 in the manner described above. The inflator 400 has an impact area 179 and a drive area 181. The impact area 179 is smaller than the drive area 181. FIG. 10 is a cutaway view of the embodiment of FIG. 9. As shown in FIG. 10, the inflator 400 may further include one or more baffles 470 that operate to guide/direct the gas 460 and the injected liquid 144. Those skilled in the art will appreciate that other configurations and/or shapes for the baffles 470 are also possible. In fact, the baffles 470 may be shaped, as necessary, to adjust the gas flow. The baffles 470 may have additional openings 472. It should be noted that in some embodiments, the direction of the gas leaving the gas generant chamber 116 is directly opposite (or substantially opposite) to the direction of the liquid 144 being injected. As a result, these two streams can push against each other and hinder proper flow. In some instances, the entrained particulates may accumulate and clog the opening 148 (FIG. 9), thereby preventing the liquid 144 from cooling the gas 460. Accordingly, the baffles 470 may be added to direct the gas flow so that it is not flowing in a direction that frustrates or prevents the flow of the liquid 144. FIG. 11 is a sectional view of another embodiment of an inflator 500 according to the present embodiments. FIG. 11A is an enlarged version of the piston portion of FIG. 11. Referring now to FIGS. 11 and 11A, the inflator 500 includes generant 104 housed within a chamber 116. As with the previous embodiments, an initiator 112 is capable of igniting the gas generant 104 into a quantity of inflation gas. A diffuser 124 is positioned within the chamber 116. The diffuser 124 includes holes 132 that will allow the gas produced by combustion of the generant 104 access to the interior of the diffuser 124 and can escape the chamber 116. The inflator 500 will further comprise a piston 136. Adjacent the piston 136 is a chamber 140 that includes a liquid 144. As shown in FIG. 11A, the liquid 144 is sealed within the chamber 140 with a seal 552. This seal 552 may be a cup that is press fit around the piston 136 to seal the chamber 140. In other embodiments, the seal 552 may be a coating that is added to the piston 136 (or the chamber 140) to seal the chamber 140. Those skilled in the art will appreciate how to seal the chamber 140 via the seal 552. It should be noted that the inflator 500, unlike some of the prior embodiments, does not have a burst disk. Rather, this embodiment has a seal 552 that is used to seal the chamber 140. Interior of the piston 136 is a mixing chamber 560. When gas produced by the combustion of the generant 104 exits the chamber 116, it may impact the piston 136 and fill the mixing chamber 560. As the piston 136 advances into chamber 140, the gas produced can exit the inflator via openings 156. The piston 136 has an impact area 179 and a drive area 181. The impact area 179 is smaller than the drive area 181. Referring now to FIGS. 12, 13, and 14, the stages of deployment of the inflator 500 are illustrated. When the generant 104 is combusted, a quantity of gas is produced. (This gas is illustrated by the arrow 160). This gas exits the chamber 116 via the diffuser 124 and may contact the piston 136. Some of this gas 160 may also begin to exit the inflator 500 via the openings 156. As the gas contacts the piston 136, the piston begins to move/displace towards the chamber 140. In turn, this displacement un seals the seal 552. For example, the press fit seal 552 is displaced such that it is no longer capable of sealing. (This may be that the piston 136 displaces past a press fit zone—i.e., an area that is press fit so that there is no longer a seal). In some embodiments, the piston 136 may have a depression 566 or other feature that, when moved towards the chamber 140, operates to ensure that there is a passage through which liquid 144 may flow. One or more vent holes 571 may also be added. These vent holes may operate to relieve pressure (such as “back pressure”) in the device. Once the seal 552 has been unsealed, liquid 144 will begin to flow out of the chamber 140. This liquid 144 may flow through openings 570 in the piston 136. Again, the movement of the piston 136 hydraulically expels the liquid 144 in the chamber 140 thereby causing the liquid 144 to inject through opening 570 for contact and mixing with the gas 160. (FIG. 13 shows the piston 136 as it is being moved, whereas FIG. 14 shows the piston 136 after it has been fully displaced and the liquid 144 has been fully expelled out of the chamber 140). The liquid 144 injected into the interior of the piston 136 is vaporized and used to inflate the airbag. However, the vaporization operates to cool the gas, as described herein. The gas 160 and the liquid 144 may mix in the mixing chamber 560. It should be noted that, in some embodiments, the gas 160 will push against the piston head 576 of piston 136 as a means of moving the piston 136. As the gas 160 pushes against this piston head 576, particulates and other undesirable byproducts can be deposited onto the piston head 576 and are thus separated from the quantity of gas 160. Further, as noted above, the piston 136 may displace towards the chamber 140. In some embodiments, this movement of the piston 136 may be facilitated by not having the chamber 140 completely full with liquid 144. In other words, there is a space (sometimes called a “head space”) within the chamber 140 into which the piston 136 may displace. In some embodiments, this head space 580 may be filled with a compressible gas 584 that allows the piston 136 to move into the chamber 140 during deployment. This compressible gas 584 may be air, argon, or any other suitable gas. This gas will escape out of the chamber 140 when it is unsealed and may further be used in the inflation process. Referring now to all of the Figures generally, it will be appreciated that the present embodiments provide various advantages. For example, the present embodiments do not require the use of a filter to cool the gas. Rather, the cooling function is replaced by the injection of the liquid 144 into the gas stream during deployment. The particulate removal function of the filter is accomplished by turning the gas flow significantly prior to exiting the inflator. The gas produced by vaporization or decomposition of the liquid contributes to the airbag inflation. The pressure required to inject the liquid into the gas stream may be provided by the combustion pressure of the inflator. Further, in some embodiments, the rate of liquid injection may be proportional to the combustion pressure of the inflator, so the liquid may inject faster at hot temperatures (higher combustion pressures) and slower at cooler temperatures (lower combustion pressures). The present invention may be embodied in other specific forms without departing from its structures, methods, or other essential characteristics as broadly described herein and claimed hereinafter. The described embodiments are to be considered in all respects only as illustrative, and not restrictive. The scope of the invention is, therefore, indicated by the appended claims, rather than by the foregoing description. All changes that come within the meaning and range of equivalency of the claims are to be embraced within their scope. 1. An inflator comprising: a quantity of gas generant housed within a gas generant chamber, the gas generant chamber having a constant volume; an initiator to ignite the gas generant and form gas during deployment; a chamber that houses a piston and a liquid, the piston having a hollow interior wherein the hollow interior of the piston is a mixing chamber sealed from the liquid, wherein during deployment the piston moves un sealing the mixing chamber from the liquid and hydraulically expelling the liquid through at least one opening in the piston such that the liquid contacts and cools the gas formed by combustion of the gas generant. 2. An inflator as in claim 1 wherein the gas formed by combustion of the gas generant and the liquid moving into the hollow interior of the piston passes through an exit opening to allow passage to the exterior of the piston. 3. An inflator as in claim 1 wherein the inflator is filterless. 4. An inflator as in claim 1 further comprising more than one opening, wherein at least one of the openings is off-set from the flow of the gas, thereby allowing the liquid to more readily be hydraulically expelled through the openings. 5. An inflator as in claim 1 wherein the liquid hydraulically expelled through the opening vaporizes and mixes with the gas formed by combustion of the gas generant. 6. An inflator as in claim 5 wherein the liquid is vaporized and mixes with the gas in a mixing chamber. 7. An inflator as in claim 1, wherein the gas formed by combustion of the gas generant flows through a tortuous path prior to exiting the inflator. 8. An inflator as in claim 7 wherein the tortuous path has a right angle turn for capturing particulates entrained in the gas. 9. An inflator as in claim 7 wherein as the gas flows through the tortuous path, at least some particulates entrained in the gas are removed from the gas and deposited within the inflator. 10. An inflator as in claim 9 wherein the particulates are deposited at a capture area. 11. An inflator as in claim 1 further comprising one or more baffles. 12. An inflator as in claim 1 wherein the liquid contains water. 13. An inflator as in claim 1 wherein the liquid is housed in a chamber and sealed by a seal, further comprising a quantity of gas that is positioned proximate the liquid. 14. An inflator as in claim 13 wherein the seal is press fit and the seal is unsealed by movement of the piston. 15. An inflator as in claim 1 wherein the interior of the piston is sealed by a burst disk. 16. An inflator as in claim 1 wherein an impact area of the piston that contacts the liquid is smaller than a drive area that contacts the gas, wherein the pressure of the liquid is amplified and causes injection of the liquid due to the differences between the impact area and the drive area. 17. An inflator as in claim 1, further comprising one or more vent holes that relieve pressure in the inflator. 18. An inflator comprising: a quantity of gas generant housed within a gas generant chamber, the gas generant chamber having a constant volume; an initiator to ignite the gas generant and form gas during deployment; a chamber that houses a piston and a liquid, the piston having a hollow interior wherein the hollow interior of the piston is a mixing chamber sealed from the liquid, wherein during deployment the piston moves and un seals the mixing chamber from the liquid and the movement of the piston hydraulically expels the liquid through at least one opening in the piston such that the liquid contacts, vaporizes, mixes with, and cools the gas formed by combustion of the gas generant, wherein the gas flows through a tortuous path, at least some particulates entrained in the gas are removed from the gas and deposited within the inflator. 19. An inflator as in claim 18 wherein the tortuous path has a right angle turn for capturing particulates entrained in the gas.
Entanglement of superconducting qubits via acceleration radiation We show that simulated relativistic motion can generate entanglement between artificial atoms and protect them from spontaneous emission. We consider a pair of superconducting qubits coupled to a resonator mode, where the modulation of the coupling strength can mimic the harmonic motion of the qubits at relativistic speeds, generating acceleration radiation. We find the optimal feasible conditions for generating a stationary entangled state between the qubits when they are initially prepared in their ground state. Furthermore, we analyse the effects of motion on the probability of spontaneous emission in the standard scenarios of single-atom and two-atom superradiance, where one or two excitations are initially present. Finally, we show that relativistic motion induces sub-radiance and can generate a Zeno-like effect, preserving the excitations from radiative decay. We show that simulated relativistic motion can generate entanglement between artificial atoms and protect them from spontaneous emission.We consider a pair of superconducting qubits coupled to a resonator mode, where the modulation of the coupling strength can mimic the harmonic motion of the qubits at relativistic speeds, generating acceleration radiation.We find the optimal feasible conditions for generating a stationary entangled state between the qubits when they are initially prepared in their ground state.Furthermore, we analyze the effects of motion on the probability of spontaneous emission in the standard scenarios of single-atom and twoatom superradiance, where one or two excitations are initially present.Finally, we show that relativistic motion induces sub-radiance and can generate a Zeno-like effect, preserving the excitations from radiative decay. Circuit Quantum Electrodynamics (cQED) [1][2][3][4] offers both a promising architecture for quantum technologies, such as quantum computers [5,6] and simulators [7][8][9], and a natural arena for the study of quantum field theory and relativistic effects, either in a direct or simulated fashion [10][11][12][13][14].For instance, the dynamical Casimir effect (DCE), produced by the modulation of the boundary conditions of the electromagnetic field at relativistic speeds, has been observed in superconducting devices [15][16][17].Along these lines, it has been shown that DCE radiation possesses several forms of quantum correlations [18][19][20][21][22] that can be transferred to superconducting qubits [23,24].A related phenomenon is the Unruh effect, where an accelerated detector in vacuum should detect thermal radiation [25,26].Recently, some of us have shown how to mimic the generation of acceleration radiation by means of the modulation of the coupling strength of a superconducting qubit [27], a phenomenon resembling the cavity-enhanced Unruh effect [28][29][30].The simulation in a superconducting architecture of both phenomena, DCE and acceleration radiation, relies on the possibility of performing an ultrafast variation of the magnetic flux threading a superconducting quantum interferometric device (SQUID) [31][32][33]. In this paper, we consider a superconducting circuit setup in which two superconducting qubits interact with the same resonator mode and effectively move at relativistic speeds, see Fig. (1).The simulation of the relativistic motion of the qubits comes from the modulation of the coupling strength between the qubits and the resonator, which can be interpreted as the qubits movement and activates the counterrotating terms of the quantum Rabi Hamiltonian.We analyse the role of the generated acceleration radiation in several collective properties of the qubits.First, we consider an initial state with no excitations and find the conditions for an efficient generation of stationary entangled states.We find several optimal scenarios for entanglement production: either both qubits move resonantly with the (or half the) natural frequency of the cavity, or one qubit remains static while the other moves at twice the cavity frequency.Second, we analyse the effect of relativis- tic motion on the spontaneous emission rate when one or two qubit excitations are initially present.Namely, we add the ingredient of relativistic motion to the celebrated Dicke scenario of single-atom and two-atom superradiance [34], which has been recently implemented in a circuit QED architecture [35].We will show that the counterrotating dynamics generated by the motion of the qubits tends to suppress the superradiance.Moreover, we find experimental conditions under which the qubit decay is completely frozen, giving rise to a Zeno-like effect induced by the continuous modulation of the coupling strength [36][37][38].In this second case, the optimal scenario for the appearance of Zeno-like effect correspond to a synchronised motion of the qubits at twice the frequency of the cavity, which generates an anti-Jaynes-Cummings dynamics in both qubits that prevents them from spontaneous emission. Entanglement and acceleration radiation.-TheHamiltonian of the system describes two superconducting qubits of frequency gaps ω q coupled to a single resonator mode of fre-arXiv:1606.06170v2[quant-ph] 7 Apr 2017 Here, σ z are Pauli matrices for the qubits, and a (a † ) is an annihilation (creation) operator for the resonator mode.The interaction Hamiltonian depends on the qubits position as with g the coupling strength and x q the qubit position [39]. In order to simulate the motion of the qubits, which are located in fixed positions, we modulate the coupling with external drivings, such that the interaction Hamiltonian for a qubit reads H I (x q ) = g cos (f 0 + ∆f cos(ω d t))σ x (a † + a), and kx q = f 0 + ∆f cos(ω d t).The velocity of the qubits vary harmonically in time, with the maximum value of ≈ λω d .For λ = 2L c = 1cm and ω d = 10 GHz, we reach values of ≈ 10 8 m/s = c/3. Initially, we consider the system in the ground state for the qubits and the resonator mode |g 1 g 2 0 .In order to determine the degree of entanglement between the qubits after a certain interaction time T , we compute the concurrence, which up to second order in perturbation theory with respect to g /ω reads C = 2 Max{|X| − P e , 0} [40].Here, X is the amplitude for photon exchange between the qubits, X = 0|T (S + 1 S + 2 )|0 , with T the time-ordering operator.P e is the probability of emitting a photon, P e = 0|S − 1 S + 1 |0 , with S + = − i g T 0 e iω q t dt cos(kx q )(e iωt a † + e −iωt a)dt = −(S − ) † .In this configuration, if both qubits are at fixed positions, the emission and photon exchange are counterrotating processes, that is, related to the breakdown of the rotatingwave approximation (RWA).These processes will be significant only for ultrastrong couplings or short interaction times.However, the motion of the qubits can excite these counterrotating terms of the Hamiltonian, giving rise to a sizeable emission of photons by a sort of cavity-enhanced Unruh effect [28]. We will analyse now under which conditions this phenomenon can be exploited to efficiently generate entanglement between the qubits.We can gain first insights by using analytical techniques.For the sake of simplicity, we assume that ω q 1 = ω q 2 = ω q and g 1 = g 2 = g.Moreover, for the sake of simplicity, we consider that the harmonic oscillations of the qubits with frequency ω d preserve its relative distance, , and λ is the wavelength associated with the cavity frequency FIG.2: Concurrence C of two qubits initially located in the centre of the resonator and oscillating from mirror to mirror with frequencies ω d1 and ω d2 , respectively.For coupling constants g1 = g2 = g = 0.02, initial state |g1 g2 0 , qubit decay parameter Γ = 0.002, and T2/T1 = 0.67, we consider two cavity decay rates: a) κ = 0.002, and b) κ = 0.2 (bad-cavity limit), in units of ω.These numerical results correspond to a broader parameter range, not limited by the perturbative approximation gT < 1, which for this case of g = 0.02 breaks for ≈ 8ωt/2π.ω.Then, we have where ∆ is the detuning between the qubits and the cavity, ∆ = ω q − ω, and J n (x) are Bessel functions of the first kind.By inspection of Eq. ( 3), we observe that both X and P e are, in general, oscillating functions.However, under certain resonant conditions, the oscillations are suppressed and these magnitudes grow monotonically in time.For instance, for negligible detuning ∆ = 0, and frequencies ω d = ω = ω q , we find that |X| , by keeping only the non-oscillating terms.Since 1, the entanglement grows quadratically in time as Therefore, we predict an entanglement resonance around ω d = ω.These analytical results are limited by the perturbative approximation employed, which assumes that gT < 1.Even in the weak coupling regime, the perturbative approximation would eventually break down.In Fig. (2), we plot the results of numerical simulations which generalise our analytical insights.The dynamics is governed by a master equation where we introduce a cavity decay rate κ, a decay parameter Γ accounting for dissipative processes, as well as a decay Γ ϕ for the dephasing of the qubits.The energy relaxation time and phase coherence time are denoted with T 1 = 1/Γ and T 2 = 1/Γ φ , respectively.We consider realistic parameters, achievable with present technology [41].This allows us to analyse the long-term dynamics of the system and to consider more general types of motion with ω d1 = ω d2 , in which the relative distance of the qubits is no longer preserved.Numerical simulations confirm the generation of a high degree of entanglement in the case of ω d1 = ω d2 = ω d = ω, as expected for short-time dynamics.In the long-term dynamics, we observe non-trivial entanglement oscillations, where maximum values are achieved at particular times shown in Fig. (2).Another optimal scenario for entanglement generation appears when one qubit is effectively moving with frequency ω d1 = 2ω, and the other remains static, ω d2 = 0.Under these circumstances, the first qubit is ruled by an anti-Jaynes Cummings (anti-JC) dynamics which maximises the counterrotating emission of photons [27].In this case, the concurrence reaches its maximal value, and the amplitude of initially perfect collapse-revival cycles eventually diminish.Asymptotically, entanglement exhibits small fluctuations around a mean value close to one.Moreover, if we also consider the bad-cavity limit, where κ g Γ, entanglement oscillations are smoothed out, and highly entangled stationary states are reached, see Fig. (2b).We extend our analysis of the generation of entanglement between both qubits for the case in which the cavity is out of resonance from both drivings and qubits frequencies.Although we have analytically seen that the concurrence increases when the resonant condition is fulfilled, we expect that considering a detuned cavity will enhance the quantum correlation between the qubits (see Supplementary Information). Single-atom superradiance and Zeno-like effect.-Inhis seminal work, Dicke showed that the decay of atomic emitters is enhanced by the presence of other atomic emitters [34].The simplest case, called single-atom superradiance, consists of a single emitter in an excited state influenced by the proximity of another emitter, even if the latter is in the ground state.This gedanken experiment has been recently realised in a circuit QED architecture in the bad-cavity limit [35].Here, we analyse the effects of relativistic motion in this scenario.To this end, we consider the initial state |g 1 e 2 0 , and discuss, firstly, the effects of the relativistic motion of the second qubit in its decay, and, secondly, the effects of the presence of the first qubit in the decay of the second one.We observe that the relativistic motion of the second qubit, encoded in ω d2 , tends to inhibit its decay leading to a decreased emission FIG.3: Motion effects in single-atom superradiance, observed in the probability P of excitation of the second qubit Q2.We consider the first qubit Q1 decoupled, g1 = 0, and the second moving with frequency ω d2 , for an initial state |g1 e2 0 .We show the behaviour for different velocities of the second qubit, ranging from the static case, ω d2 = 0, to a velocity of ω d2 = 2ω.For a coupling constant g2 = 0.02, a qubit decay parameter Γ = 0.002 and T2/T1 = 0.67, we consider two cavity decay rates: a) κ = 0.002, and b) κ = 0.2 (bad-cavity limit), in units of ω. rate, known as sub-radiance, see Fig. 3. Again, we can get some insight on the system dynamics from first-order analytical computations.Since in this case the qubit is initially excited, the probability of emitting a photon is not given by Eq. (3).In particular, in the resonant case (∆ = 0), we have . Then, considering the trajectories for both qubits such that . This means that for a static qubit -ω d2 close to 0-, the probability of emission grows quadratically in time P e g 2 T 2 , while for frequencies of motion significantly different from 0 the probability oscillates with an amplitude which decreases with ω d2 .Therefore, for large enough ω d2 the probability of emission is suppressed, as can be seen in Fig. (3).Note that the maximum acceleration of the qubit motion is proportional to ω 2 d2 .Thus, the larger the acceleration is, the larger the suppression of the probability of emission.Then, the subradiance can be seen as another relativistic effect, hitherto unexplored.At first glance, this subradiant dynamics might look surprising, since relativistic accelerated motion is typically associated to the emission of photons.However, note that both FIG.4: Motion effects in single-atom superradiance, observed in the probability P of excitation of the second qubit Q2.We consider the influence of the movement of the first qubit Q1 analysing different velocities ω d1 and the decoupled case, g1 = 0.The second qubit is moving with frequency ω d2 , for an initial state |g1 e2 0 .We compute for a coupling constant g1 = g2 = 0.02 in the cases with the first qubit coupled, a qubit decay parameter Γ = 0.002 and T2/T1 = 0.67, and a cavity decay rate κ = 0.2 (bad-cavity limit), in units of ω.We show the modification in the behaviour in the case of the second qubit a) static ω d2 = 0, and b) moving with ω d2 = ω.phenomena, Unruh effect and subrradiance, are activated by the counterrotating terms of the Hamiltonian, which become dominant for large enough ω d2 associated with relativistic motion.While non-RWA dynamics gives rise to emission of photons when the qubit and cavity start in the ground state, in the present case the initial state |e 2 0 would be stationary in the presence of only non-RWA terms.The decay dynamics of the second qubit is effectively frozen, that is, we observe a Zeno-like effect generated by the continuous modulation of the qubit-cavity coupling strength, which has an effect similar to a continuous monitoring of the system [36][37][38]. Not only the relativistic motion of the second qubit, encoded in ω d2 , tends to inhibit its decay, but also the motion of the first qubit, ω d1 , has a significant effect on the emission rate for certain values of ω d1 .In order to analyse this influence, we compare in Fig. 4 the probability of excitation of the second qubit for the case in which the first qubit is decoupled, and the case in which it is coupled moving at different relativistic speeds, ω d1 .Firstly, we consider the extreme case with the second qubit static, ω d2 = 0, and we observe that a first qubit relativistic speed ω d1 = 2ω leads to a decreased emission rate, FIG.5: Zeno-like effect in the probability P of excitation of two qubits moving with the same frequency ω d1 = ω d2 = 2ω for different amplitudes of oscillation for the first qubit Q1.We consider the initial state |g1 e2 0 , a cavity decay rate κ = 0.1, coupling constants g1 = g2 = g = 0.01, and qubit decay parameter Γ = 0.001 and T2/T1 = 0.67, in units of ω, and that the first qubit is initially placed at Lc/4, with Lc the cavity length.We show the excitation probabilities of both qubits, Q1 and Q2, with frequencies ω q 1 = ω q 2 = ω, and adding a detuning ∆ = 0.1ω between the first qubit and the cavity, for an amplitude of motion of the first qubit a) Lc/4, and b) Lc/16.known as sub-radiance, whereas for other combinations of frequencies, the behaviour of the probability of excitation of the second qubit remains unaltered.Secondly, we analyse the case with the second qubit moving with ω d2 = ω, and observe a dramatic change in the decay rate of the second qubit for the same frequency ω d1 = 2ω, as in the previous case.We notice that the emission rate of the second qubit is also slightly modified for a relation of frequencies ω d1 = ω d2 = ω that generates entanglement.However, a further analysis in the relation of the generated entanglement and single-atom superradiance allows us to discard drastic influences of the former in the decay rate (see Supplementary Information).We also interpret the frozen dynamics of the second qubit for ω d2 = ω, and for ω d2 = ω and ω d1 = 2ω, as a Zeno-like effect [36][37][38]. We extend our analysis to other scenarios by considering different frequencies and initial conditions.We confirm the Zeno-like effect when we reduce the oscillation amplitude and the excitation probability of the first qubit, which enhances even more the effect on the second qubit (see Fig. 5).Twoatom superradiance is a collective effect consisting in an enhancement of the decay rate of two emitters with respect to their individual ones [34].We consider the effects of relativistic motion on the decay of two artificial atoms, |e 1 e 2 0 and observe a sub-radiance phenomenon, as in the case of singleatom superradiance (see Supplementary Information). Implementation in superconducting circuits.-Themodel described in Eq. ( 1) can be implemented in a circuit QED architecture [42], using a single-mode transmission line resonator (TLR) interacting with two tunable-coupling superconducting qubits.In order to observe all the phenomenology so far described in a single device, it is required independent tuning of the qubit transition frequencies and of the qubitcavity coupling strengths.Tunable coupling superconducting qubits [41,43,44] coupled to a TLR and tuning of effective couplings over nanosecond time-scale [45] have been proven in circuit QED architectures. Conclusions.-We have proposed a possible realisation in which simulated relativistic motion generates true entanglement between artificial atoms, while protecting them from spontaneous emission in a Zeno-like effect.A natural extension would be to consider the effects of multi-atom relativistic motion, with a study of superradiant phase transition.Both the ability of generating entanglement and state protection may pave the way for new applications in superconducting quantum technologies. RELATION OF CONCURRENCE AND SINGLE-ATOM SUPERRADIANCE We have explored numerically the influence of the movement of the first qubit in the emission rate of the second qubit in Fig. 4 of the main text.In order to better interpret the possible relation between the generated entanglement and the influence of the first qubit, we show in Fig. 7 the concurrence for the relation of frequencies ω d1 and ω d2 analyzed in Fig. 4. We conclude that a combination of frequencies generating entanglement is not relevant for observing significant changes in the probability of the decay of the second qubit, although for ω d1 = ω d2 = ω it results in a slight modification of the decay rate.FIG.7: Concurrence C of two qubits moving with different velocities ω d1 and ω d2 , previously analyzed in the light of single-atom superradiance.We consider a coupling constant g1 = g2 = 0.02, a qubit decay parameter Γ = 0.002 and T2/T1 = 0.67, and a cavity decay rate κ = 0.2 (bad-cavity limit), in units of ω.We show results for different velocities ω d1 of the first qubit, when the second qubit is a) static ω d2 = 0, and b) moving with ω d2 = ω.FIG.8: Probability P of excitation of two qubits, Q1 and Q2, moving with different frequencies ω d1 , ω d2 , and a qubit decay parameter Γ = 0.001 and T2/T1 = 0.67, in units of ω. a), b) Both qubits are initially excited.c), d) The initial state of the qubits is | +1 +2 .a) , c) κ = 0.001, and b), d) κ = 0.1 (bad-cavity limit), in units of ω.All figures show the case of both qubits coupled to the cavity with coupling strengths g1 = g2 = g = 0.01, ω d1 = ω d2 = 2ω, and the case of an uncoupled first qubit g1 = 0, with a static second qubit ω d2 = 0 with g2 = 0.01. TWO-ATOM SUPERRADIANCE AND COLLECTIVE ZENO-LIKE EFFECT We consider the phenomenon of two-atom superradiance in our setup.We compare in Fig. 8 the decay rate of both qubits moving with relativistic velocities, encoded in ω d1 and ω d2 , with respect to their individual decay rates, i.e., with respect to the decay of a single static qubit.We observe that the relativistic motion alters the emission rate and reduces drastically for both qubits the probability of decay, leading to a sub-radiance phenomenon.Both in the good cavity and bad cavity limit, we observe a collective Zeno-like effect, namely the effective freezing of the decay of both qubits.This ties in with the prediction of the expected anti-JC dynamics at this particular driving frequency [27].Finally, if we consider a different initial state where both qubits are in a superposition of their ground and excited states, | + 1 + 2 0 , with |+ = 1/ √ 2(|e + |g ), we observe that the qubits are driven fast to their excited states, and the ensuing dynamics follows then the one already observed for an initial |e 1 e 2 . IMPLEMENTATION DETAILS With the aim of providing a specific implementation proposal, we focus on a three island superconducting qubit [43,44].The qubit scheme consists in two shunted SQUID loops described by a model composed of two interacting anharmonic oscillators, whose dynamics can be effectively restricted to their two lowest eigenstates.Independent control over the corresponding transition frequencies results in a completely tunable qubit.We denote |0 a the ground and |1 a the first excited state of the oscillator a, and we call ω a the corresponding transition frequency.Same notation will apply for the oscillator b.The qubit logical levels are given by the ground |E 0 and first-excited |E 1 states of the collective system, whose structure depends on the ratio between the frequencies ω i .When the two anharmonic oscillators are detuned, the lowest collective eigenstates are given by |E 0 = |0 a |0 b and |E 1 = |1 a |0 b , where we set ω a < ω b .In this configuration, the qubit is strongly coupled to the TLR.On the other hand, when the two anharmonic oscillators are nearly degenerate ω a ≈ ω b , the first collective excited state is given by Such state corresponds to an antiparallel configuration of the dipoles of the SQUID loops, hence creating a quadrupolar moment that does not couple with the resonator.Swapping between the two collective energyconfigurations, the effective coupling can be continuously tuned in real time, without exciting higher energy levels [46], which are detuned by at least 1 GHz from the primary qubit transition Using this scheme, tuning of the qubit-cavity coupling strength has been experimentally achieved in the range from 40 MHz to less than 200 KHz [44]. System initialization is trivial in all the cases considered in this manuscript, as only ground-state cooling and single-qubit gates are required.Fast read-out of the qubit state can be implemented [41] driving a transition to the second-excited collective state |E 1 → |E 2 .In the resonant case, ω a = ω b , such state is given by |E 2 = (|1 a |0 b + |0 a |1 b ) / √ 2 and it is strongly coupled to the resonator.The qubit excitation states can be then individually obtained measuring the state-dependent shift on the TLR resonant frequency. FIG. 1 : FIG.1: Two superconducting qubits strongly coupled to a single resonator mode and driven with frequencies ω d1 and ω d2 simulating harmonic relativistic speeds.The resonator of length Lc is initially in the vacuum while the qubits, which are located in the middle of the cavity x = 0, are initially a) both in the ground state, b) one in the excited state and the other in the ground state, c) both in the excited state.The red wavy arrows indicate emission or absence of emission of photons from the qubits, showing a subradiant behaviour.
/** * Copyright (c) 2014-2015 Spoqa, All Rights Reserved. */ package com.spoqa.battery; import java.lang.annotation.Annotation; import java.lang.reflect.Field; import java.lang.reflect.Method; import java.util.HashMap; import java.util.List; import java.util.Map; public class ReflectionCache { private static class AnnotatedPropertyKey { public static AnnotatedPropertyKey instance = null; static { instance = new AnnotatedPropertyKey(); } public Class<? extends Annotation> annotation; public Class baseClass; public AnnotatedPropertyKey() { } public AnnotatedPropertyKey(Class<? extends Annotation> annotation, Class baseClass) { this.annotation = annotation; this.baseClass = baseClass; } @Override public int hashCode() { return annotation.hashCode() ^ baseClass.hashCode(); } @Override public boolean equals(Object other_) { if (!(other_ instanceof AnnotatedPropertyKey)) return false; AnnotatedPropertyKey other = (AnnotatedPropertyKey) other_; return other.annotation == annotation && other.baseClass == baseClass; } } private static class MemberAnnotationKey<T> { public T member; public Class<? extends Annotation> annotation; public MemberAnnotationKey(T member, Class<? extends Annotation> annotation) { this.member = member; this.annotation = annotation; } @Override public int hashCode() { return member.hashCode() ^ annotation.hashCode(); } @Override public boolean equals(Object other_) { if (!(other_ instanceof MemberAnnotationKey)) return false; MemberAnnotationKey other = (MemberAnnotationKey) other_; return other.member == member && other.annotation == annotation; } } private static class FieldAnnotationKey extends MemberAnnotationKey<Field> { static public FieldAnnotationKey instance = null; static { instance = new FieldAnnotationKey(null, null); } public FieldAnnotationKey(Field field, Class<? extends Annotation> annotation) { super(field, annotation); } } private static class MethodAnnotationKey extends MemberAnnotationKey<Method> { static public MethodAnnotationKey instance = null; static { instance = new MethodAnnotationKey(null, null); } public MethodAnnotationKey(Method method, Class<? extends Annotation> annotation) { super(method, annotation); } } private Map<AnnotatedPropertyKey, List<Field>> mAnnotatedFieldsCache; private Map<AnnotatedPropertyKey, List<Method>> mAnnotatedGetterMethodsCache; private Map<AnnotatedPropertyKey, List<Method>> mAnnotatedSetterMethodsCache; private Map<Class, List<Field>> mFieldsCache; private Map<Class, List<Method>> mGetterMethodsCache; private Map<Class, List<Method>> mSetterMethodsCache; private Map<FieldAnnotationKey, Annotation> mFieldAnnotationCache; private Map<MethodAnnotationKey, Annotation> mMethodAnnotationCache; public ReflectionCache() { mAnnotatedFieldsCache = new HashMap<AnnotatedPropertyKey, List<Field>>(); mAnnotatedGetterMethodsCache = new HashMap<AnnotatedPropertyKey, List<Method>>(); mAnnotatedSetterMethodsCache = new HashMap<AnnotatedPropertyKey, List<Method>>(); mFieldsCache = new HashMap<Class, List<Field>>(); mGetterMethodsCache = new HashMap<Class, List<Method>>(); mSetterMethodsCache = new HashMap<Class, List<Method>>(); mFieldAnnotationCache = new HashMap<FieldAnnotationKey, Annotation>(); mMethodAnnotationCache = new HashMap<MethodAnnotationKey, Annotation>(); } public List<Field> queryCachedAnnotatedFields(Class<? extends Annotation> annotation, Class baseClass) { AnnotatedPropertyKey key = AnnotatedPropertyKey.instance; key.annotation = annotation; key.baseClass = baseClass; if (mAnnotatedFieldsCache.containsKey(key)) return mAnnotatedFieldsCache.get(key); return null; } public List<Method> queryCachedAnnotatedGetterMethods(Class<? extends Annotation> annotation, Class baseClass) { AnnotatedPropertyKey key = AnnotatedPropertyKey.instance; key.annotation = annotation; key.baseClass = baseClass; if (mAnnotatedGetterMethodsCache.containsKey(key)) return mAnnotatedGetterMethodsCache.get(key); return null; } public List<Method> queryCachedAnnotatedSetterMethods(Class<? extends Annotation> annotation, Class baseClass) { AnnotatedPropertyKey key = AnnotatedPropertyKey.instance; key.annotation = annotation; key.baseClass = baseClass; if (mAnnotatedSetterMethodsCache.containsKey(key)) return mAnnotatedSetterMethodsCache.get(key); return null; } public List<Field> queryCachedFields(Class baseClass) { if (mFieldsCache.containsKey(baseClass)) return mFieldsCache.get(baseClass); return null; } public List<Method> queryCachedGetterMethods(Class baseClass) { if (mGetterMethodsCache.containsKey(baseClass)) return mGetterMethodsCache.get(baseClass); return null; } public List<Method> queryCachedSetterMethods(Class baseClass) { if (mSetterMethodsCache.containsKey(baseClass)) return mSetterMethodsCache.get(baseClass); return null; } public Annotation queryFieldAnnotation(Field field, Class<? extends Annotation> annotationClass) { FieldAnnotationKey key = FieldAnnotationKey.instance; key.member = field; key.annotation = annotationClass; if (mFieldAnnotationCache.containsKey(key)) { return mFieldAnnotationCache.get(key); } return null; } public Annotation queryMethodAnnotation(Method method, Class<? extends Annotation> annotationClass) { MethodAnnotationKey key = MethodAnnotationKey.instance; key.member = method; key.annotation = annotationClass; if (mFieldAnnotationCache.containsKey(key)) { return mFieldAnnotationCache.get(key); } return null; } public void cacheAnnotatedFields(Class<? extends Annotation> annotation, Class baseClass, List<Field> fields) { mAnnotatedFieldsCache.put(new AnnotatedPropertyKey(annotation, baseClass), fields); } public void cacheAnnotatedGetterMethods(Class<? extends Annotation> annotation, Class baseClass, List<Method> methods) { mAnnotatedGetterMethodsCache.put(new AnnotatedPropertyKey(annotation, baseClass), methods); } public void cacheAnnotatedSetterMethods(Class<? extends Annotation> annotation, Class baseClass, List<Method> methods) { mAnnotatedSetterMethodsCache.put(new AnnotatedPropertyKey(annotation, baseClass), methods); } public void cacheFields(Class baseClass, List<Field> fields) { mFieldsCache.put(baseClass, fields); } public void cacheGetterMethods(Class baseClass, List<Method> methods) { mGetterMethodsCache.put(baseClass, methods); } public void cacheSetterMethods(Class baseClass, List<Method> methods) { mSetterMethodsCache.put(baseClass, methods); } public void cacheFieldAnnotation(Field field, Class<? extends Annotation> annotationClass, Annotation a) { mFieldAnnotationCache.put(new FieldAnnotationKey(field, annotationClass), a); } public void cacheMethodAnnotation(Method method, Class<? extends Annotation> annotationClass, Annotation a) { mMethodAnnotationCache.put(new MethodAnnotationKey(method, annotationClass), a); } public boolean containsFieldAnnotation(Field field, Class<? extends Annotation> annotationClass) { FieldAnnotationKey key = FieldAnnotationKey.instance; key.member = field; key.annotation = annotationClass; return mFieldAnnotationCache.containsKey(key); } public boolean containsMethodAnnotation(Method method, Class<? extends Annotation> annotationClass) { MethodAnnotationKey key = MethodAnnotationKey.instance; key.member = method; key.annotation = annotationClass; return mFieldAnnotationCache.containsKey(key); } }
PAPILI0N1H.K. Eurycus of Boisduval, and probably Leptocircus of Swainson, but excluding P. Pylades and two other species, of which he formed the genus Zelima. In this he is followed by Latreille and Godart, the latter, however, incorporating the Fabrician genus Zelima, and excluding the P. Curius of Fabricius, now the type of Leptocircus. Three years previous to the appearance of Godart's volume of the Encyelopedie Methodique, Hiibner, in his Verzeichniss belannter Schmetterlinge, had divided the Fabrician genus Papilio into eighteen " Vereine," of which the only one that can be considered generic is Troides ; this name, as has been already remarked, cannot be retained. Swainson, in his Zooloqical Illustrations, next indicated various sections, to which he gave names, unfortunately entirely inadmissible, from his adoption of specific for generic names. Two of his groups correspond to Ornithoptera and Eurycus of Boisduval. The genus is here adopted precisely as limited by Boisduval ; for, though, from its great extent and the variety of forms it comprises, it would be very desirable to subdivide it, " there does not exist a more compact or more natural genus, or one which more entirely resists all attempt at division. There is no middle course, we must either leave it as it is, or divide it into two scores of genera." * In the Papiliones the palpi are triarticulate, generally densely covered with scales and hairs, are closely applied to the forehead, and but little, if at all, visible from above ; the terminal joints are very small. The antennae are more or less elongate; the club gradually enlarged, and curved outwards. The thorax is robust; the prothorax less developed than in Ornithoptera ; the abdomen less elongate, and more oval, than in that genus. The anterior wings are mostly triangular, the costal and outer margins being longer than the inner margin. They are sometimes more rounded, elongate, or falcate, than what may be considered the typical form. They have a distinct baseo-median nervule, and an upper disco-cellular of considerable length ; the subcostal nervure throws off two nervules near together, about the middle of the cellf; the third exactly at the end of the cell; and the fourth about midway between this and the apex. In P. Sarpedon, P. Agapenor, and their allies, and also in the small African group of which P. Leonidas is the type, the first subcostal nervule, instead of running to the costa, below, and parallel to, the costal nervure, runs immediately into this nervule ; a peculiarity which recurs in the Danai with green spotted wings, so closely analogous to the last-named species. The posterior wings vary much in form, even in the same species, as, for instance, in P. Pammon, where the tail is sometimes wanting in the males ; and in P. Memnon, where they are never caudate in the males, but not unfrequently in the females have a spatulate tail. When not tailed, the outer margin is mostly rounded and dentate ; but sometimes, as in P. Sarpedon and its allies, the wings have a triangular outline, sometimes an oval or ovate, as in P. Rhetenor and P. Elephenor. When tailed they vary still more in form : the tails are sometimes short, obtuse, spatulate, or short and pointed, sometimes very long and slender. Occasionally, as in P. Payeni and P. Evan, the whole wing is gradually produced into a tail; in general there is merely a greater or less prolongation of one of the dentations. The group of which P. Grayii and P. Lenaeus are a type have the posterior wings very similar in form to those of some species of Charaxes, and, like these, have the costal margin of the anterior wings serrated ; a correspondence in structure analogous to that already noticed between P. Leonidas and certain Danai. or less curved, generally equal, but in P. Triopas of unequal length. The Larvae differ materially in form, and, if ever we can gain tolerably complete information in regard to them, will probably afford good characters for dividing the species into sections. The little we know of any, except those of European species, is to be learned chiefly from the works of Stoll, Abbot, and Horsfield, and from the drawings of Abbot and Hardwicke now in the British Museum. Those of P. Hector, Polydorus, &c, which, like those of Ornithoptera and Thais, live on Aristolochiaj, are dark coloured, have tubercles on each segment, disposed in rows, as in the larva' of those genera. From those of the former genus they differ solely in not having the external sheath for the tentacula; from those of the latter in not having the tops of the tubercles hairy. Those of P. Polymnestor, P. Pammon, P. Arjuna, P. Erectheus, P. Cresphontes, P. Troilus,
from bangtal import * import threading setGameOption(GameOption.ROOM_TITLE, False) setGameOption(GameOption.INVENTORY_BUTTON, False) setGameOption(GameOption.MESSAGE_BOX_BUTTON, False) #res 1280*720 mainScene = Scene('','images/main.png') inGame = Scene('','images/medi1.1.png') end = Scene('','images/end.png') currentScene = 0 Sound.play(Sound('sounds/maintheme.mp3')) startButton = Object('images/startbutton.png') startButton.locate(mainScene, 1000, 5) startButton.setScale(0.3) startButton.show() def startButton_onMouseAction(x,y,action): global currentScene currentScene = 1 startgame() startButton.onMouseAction = startButton_onMouseAction #=====Player Moves, Act CyanX = 112 CyanY = 288 Cyan = Object('images/Cyan1.png') Cyan.locate(inGame, CyanX, CyanY) Cyan.setScale(4) Cyan.show() key = 0 def keyboard(k, pressed): global dir global pre global CyanMoving global key key = k if pressed: dir = key if (k==1 or k==4 or k==19 or k==23): CyanMoving = True else: CyanMoving = False else: dir = 0 CyanMoving = False inGame.onKeyboard = keyboard CyanMoving = False speed = 8 dir = 0 def CyanMove(): global CyanX global CyanY global dir global speed global wallX global wallY if dir==4: if wallX != 1: CyanX+=speed Cyan.locate(inGame, CyanX, CyanY) elif dir==19: if wallY != -1: CyanY-=speed Cyan.locate(inGame, CyanX, CyanY) elif dir==1: if wallX != -1: CyanX-=speed Cyan.locate(inGame, CyanX, CyanY) elif dir ==23: if wallY != 1: CyanY+=speed Cyan.locate(inGame, CyanX, CyanY) wallX = 0 wallY = 0 def Collision(): global currentScene global CyanX global CyanY global wallX global wallY global door1Open if currentScene ==1: if (CyanX<80): wallX = -1 elif (CyanX<=184 and CyanY>344): wallX = -1 elif (CyanX>750): wallX = 1 else: wallX = 0 if (CyanY<40): wallY = -1 elif (72<=CyanX<184 and CyanY>=344): wallY = 1 elif (CyanY>400): wallY = 1 else: wallY = 0 elif currentScene ==2: if (CyanX<=272): wallX = -1 elif (CyanX>824): wallX = 1 else: wallX = 0 if (CyanY<160): wallY = -1 elif (CyanY>560): wallY = 1 else: wallY = 0 elif currentScene == 3: if (CyanY<200): if (CyanX<=288): wallX = -1 elif (CyanX>=480): wallX = 1 else: wallX = 0 else: if (CyanX<40): wallX = -1 elif (CyanX>950): wallX = 1 else: wallX = 0 if (288<=CyanX<=480): if (CyanY<40): wallY = -1 elif (CyanY>=424): wallY = 1 else: wallY = 0 else: if (CyanY<208): wallY = -1 elif (CyanY>=424): wallY = 1 else: wallY = 0 elif currentScene == 4: if (CyanX>1144): wallX = 1 else: wallX = 0 if (CyanY<=200): wallY = -1 elif (CyanY>=424): wallY = 1 else: wallY = 0 elif currentScene == 5: if (CyanY >= 384): if (CyanX<352): wallX=-1 elif (CyanX>=960): wallX=1 else: wallX=0 else: if (CyanX<=888): wallX=-1 elif (CyanX>=960): wallX=1 else: wallX=0 if (CyanX >= 888): if (CyanY<=16): wallY=-1 elif (CyanY>=496): wallY=1 else: wallY=0 else: if (CyanY<=388): wallY=-1 elif (CyanY>=496): wallY=1 else: wallY=0 elif currentScene == 6: if (552<=CyanX<=704): if CyanY>480: wallY=1 elif CyanY<=16: wallY=-1 else: wallY=0 else: if CyanY>480: wallY=1 elif CyanY<=360: wallY=-1 else: wallY=0 if (CyanY>=360): if (CyanX<=144): wallX=-1 elif (CyanX>=960): wallX=1 else: wallX=0 else: if (CyanX<=552): wallX=-1 elif (CyanX>=704): wallX=1 else: wallX=0 elif currentScene == 7: if (CyanY<=360): if (CyanX<=280): wallX=-1 elif (CyanX>=992): wallX=1 else: wallX=0 else: if (CyanX<=552): wallX=-1 elif (CyanX>=704): wallX=1 else: wallX=0 if (552<=CyanX<=704): if CyanY>576: wallY=1 elif CyanY<=192: wallY=-1 else: wallY=0 else: if CyanY>=360: wallY=1 elif CyanY<=192: wallY=-1 else: wallY=0 CyanImageToggle = 1 CyanImageRight = 1 def CyanAni(): global CyanMoving global CyanImageToggle global CyanImageRight global isWall global dir if CyanMoving: if dir == 4: CyanImageRight = True elif dir == 1: CyanImageRight = False if CyanImageRight==1: if CyanImageToggle==1: Cyan.setImage('images/Cyan2.png') CyanImageToggle = 2 elif CyanImageToggle==2: Cyan.setImage('images/Cyan1.png') CyanImageToggle = 1 elif CyanImageRight==0: if CyanImageToggle==1: Cyan.setImage('images/Cyan2.1.png') CyanImageToggle = 2 elif CyanImageToggle==2: Cyan.setImage('images/Cyan1.1.png') CyanImageToggle = 1 else: if CyanImageRight==1: Cyan.setImage('images/Cyan1.png') elif CyanImageRight==0: Cyan.setImage('images/Cyan1.1.png') def walkSound(): global CyanMoving if CyanMoving: Sound.play(Sound('sounds/walk.mp3')) else: Sound.stop(Sound('sounds/walk.mp3')) def showXY(): global CyanX global CyanY print(CyanX, CyanY) time = 0.0 def tick(): global time timer = threading.Timer(0.01,tick) #0.1초 주기 time += 0.01 timer.start() Collision() CyanMove() sceneChange() if (int(time*100)%5==0): #0.5초 주기 CyanAni() walkSound() #showXY() #=====Scene & Room def sceneChange(): global currentScene global CyanX global CyanY global isLightOn global door1Open global door2Open global door3Open if currentScene==1: if (264<=CyanX<=552 and CyanY<=32): currentScene = 2 CyanY = 496 Cyan.locate(inGame, CyanX, CyanY) inGame.setImage('images/medi2.1.png') roomChanged(2) elif (264<=CyanX<=552 and CyanY>=400): if door1Open: currentScene = 3 CyanY = 56 Cyan.locate(inGame, CyanX, CyanY) inGame.setImage('images/hall1.jpg') roomChanged(3) else: CyanY=392 Cyan.locate(inGame, CyanX, CyanY) showMessage("열리지 않는다...") elif currentScene==2: if (264<=CyanX<=552 and CyanY>=512): currentScene = 1 CyanY = 56 Cyan.locate(inGame, CyanX, CyanY) if door1Open: inGame.setImage('images/medi1.2.png') else: inGame.setImage('images/medi1.1.png') roomChanged(1) elif currentScene==3: if (280<=CyanX<=488 and CyanY<=32): currentScene = 1 CyanY = 392 Cyan.locate(inGame, CyanX, CyanY) inGame.setImage('images/medi1.2.png') roomChanged(1) elif (CyanX<=48): currentScene = 4 CyanX = 1072 Cyan.locate(inGame, CyanX, CyanY) if isLightOn: inGame.setImage('images/hall2.1.jpg') else: inGame.setImage('images/hall2.2.jpg') roomChanged(4) elif (CyanX>=952): CyanX = 944 Cyan.locate(inGame, CyanX, CyanY) showMessage("열리지 않는다...") elif currentScene==4: if (CyanX>1128): currentScene = 3 CyanX =56 Cyan.locate(inGame, CyanX, CyanY) inGame.setImage('images/hall1.jpg') roomChanged(3) elif (CyanX<=8): if door2Open: currentScene = 5 CyanX = 936 Cyan.locate(inGame, CyanX, CyanY) inGame.setImage('images/upperengine1.png') roomChanged(5) else: CyanX = 16 Cyan.locate(inGame, CyanX, CyanY) showMessage("열리지 않는다...") elif currentScene==5: if (200<=CyanY<=424): if CyanX >= 960: currentScene = 4 CyanX = 88 Cyan.locate(inGame, CyanX, CyanY) if isLightOn: inGame.setImage('images/hall2.1.jpg') else: inGame.setImage('images/hall2.2.jpg') roomChanged(4) elif (CyanY<=16): currentScene = 6 CyanY = 480 Cyan.locate(inGame, CyanX, CyanY) inGame.setImage('images/upperengine2.png') roomChanged(6) elif currentScene==6: if (888<=CyanX and CyanY>=480): currentScene = 5 CyanY = 24 Cyan.locate(inGame, CyanX, CyanY) inGame.setImage('images/upperengine1.png') roomChanged(5) elif CyanY<=16: currentScene = 7 CyanY = 576 Cyan.locate(inGame, CyanX, CyanY) if door3Open: inGame.setImage('images/hall3.2.png') else: inGame.setImage('images/hall3.1.png') roomChanged(7) elif currentScene == 7: if CyanY>576: currentScene = 6 CyanY = 24 Cyan.locate(inGame, CyanX, CyanY) inGame.setImage('images/upperengine2.png') roomChanged(6) if (CyanX>=992): CyanX = 984 Cyan.locate(inGame, CyanX, CyanY) showMessage("열리지 않는다...") if (CyanX<=280): if not door3Open: CyanX = 288 Cyan.locate(inGame, CyanX, CyanY) showMessage("열리지 않는다...") else: finishGame() def roomChanged(room): global vent1Removed if room==1: lightButton1.show() medicom.hide() vent1.hide() cctv1.hide() elif room==2: lightButton1.hide() medicom.show() if vent1Removed: vent1.hide() else: vent1.show() elif room==3: bulb1.hide() bulb2.hide() bulb3.hide() bulb4.hide() lightButton1.hide() cctv1.show() elif room==4: bulb1.show() bulb2.show() bulb3.show() bulb4.show() cctv1.hide() elif room==5: bulb1.hide() bulb2.hide() bulb3.hide() bulb4.hide() elif room==6: cctv2.hide() lightButton2.hide() elif room==7: cctv2.show() lightButton2.show() Cyan.show() #=====Doors, Objects door1Open = False door2Open = False door3Open = False medicom = Object('images/medicom.png') medicom.locate(inGame, 940, 314) medicom.setScale(1.1) def medicom_onMouseAction(x,y,action): global CyanX global CyanY global door1Open global isLightOn if (CyanX>=824 and 272<CyanY<440): if (not door1Open): if isLightOn: Sound.play(Sound('sounds/dooropen.mp3')) door1Open=True showMessage("문을 열었다...") else: showMessage("화면이 잘 안보인다...") medicom.onMouseAction = medicom_onMouseAction lightButton1 = Object('images/lightbutton1.png') lightButton1.locate(inGame, 90, 475) isLightOn = False def lightButton1_onMouseAction(x,y,action): global CyanX global CyanY global isLightOn if (CyanX<=184 and CyanY>=344): if isLightOn: isLightOn = False inGame.setLight(0.25) else: isLightOn = True inGame.setLight(1) lightButton1.onMouseAction = lightButton1_onMouseAction lightButton2 = Object('images/lightbutton2.png') lightButton2.locate(inGame, 940, 512) def lightButton2_onMouseAction(x,y,action): global CyanX global CyanY global isLightOn if (CyanX>=840 and CyanY>=360): if isLightOn: isLightOn = False inGame.setLight(0.25) else: isLightOn = True inGame.setLight(1) lightButton2.onMouseAction = lightButton2_onMouseAction vent1Removed = False vent1 = Object('images/vent.png') vent1.locate(inGame, 150, 440) vent1.setScale(1.4) def vent1_onMouseAction(x,y,action): global CyanX global CyanY global vent1Removed if (CyanX<=280 and 360<=CyanY<=480): if isLightOn: vent1Removed = True vent1.hide() Sound.play(Sound('sounds/vent.mp3')) showMessage('벤트를 치웠다...') else: showMessage('보이지 않아 치우기 어렵다...') vent1.onMouseAction = vent1_onMouseAction bulb1On = True bulb1 = Object('images/bulbon.png') bulb1.locate(inGame, 0, 570) def bulb1_onMouseAction(x,y,action): global bulb1On if bulb1On: bulb1.setImage('images/bulboff.png') bulb1On=False else: bulb1.setImage('images/bulbon.png') bulb1On=True bulbCheck() bulb1.onMouseAction = bulb1_onMouseAction bulb2On = True bulb2 = Object('images/bulbon.png') bulb2.locate(inGame, 532, 570) def bulb2_onMouseAction(x,y,action): global bulb2On if bulb2On: bulb2.setImage('images/bulboff.png') bulb2On=False else: bulb2.setImage('images/bulbon.png') bulb2On=True bulbCheck() bulb2.onMouseAction = bulb2_onMouseAction bulb3On = True bulb3 = Object('images/bulbon.png') bulb3.locate(inGame, 656, 570) def bulb3_onMouseAction(x,y,action): global bulb3On if bulb3On: bulb3.setImage('images/bulboff.png') bulb3On=False else: bulb3.setImage('images/bulbon.png') bulb3On=True bulbCheck() bulb3.onMouseAction = bulb3_onMouseAction bulb4On = True bulb4 = Object('images/bulbon.png') bulb4.locate(inGame, 1196, 570) def bulb4_onMouseAction(x,y,action): global bulb4On if bulb4On: bulb4.setImage('images/bulboff.png') bulb4On=False else: bulb4.setImage('images/bulbon.png') bulb4On=True bulbCheck() bulb4.onMouseAction = bulb4_onMouseAction def bulbCheck(): global bulb1On global bulb2On global bulb3On global bulb4On global door2Open if not door2Open: if bulb1On and bulb4On: if not bulb2On: if not bulb3On: showMessage("문이 열렸다...") Sound.play(Sound('sounds/dooropen.mp3')) door2Open=True cctv1On = False cctv1 = Object('images/cctv1off.png') cctv1.locate(inGame, 872, 602) def cctv1_onMouseAction(x,y,action): global CyanX global CyanY global isLightOn global cctv1On global cctv2On global door3Open if (808<=CyanX<=936 and CyanY>=424): if (not cctv1On): if isLightOn: cctv1On = True cctv1.setImage('images/cctv1on.png') if cctv2On: showMessage("문이 열렸다...") Sound.play(Sound('dooropen.mp3')) door3Open=True else: showMessage("CCTV를 켰다...") else: showMessage("어두워서 CCTV를 켤수가 없다...") cctv1.onMouseAction = cctv1_onMouseAction cctv2On = False cctv2 = Object('images/cctv2off.png') cctv2.locate(inGame, 482, 510) def cctv2_onMouseAction(x,y,action): global CyanX global CyanY global isLightOn global cctv1On global cctv2On global door3Open if (400<=CyanX<=504 and 360<=CyanY): if (not cctv2On): if isLightOn: cctv2On = True cctv2.setImage('images/cctv2on.png') if cctv1On: inGame.setImage('images/hall3.2.png') Sound.play(Sound('sounds/dooropen.mp3')) showMessage("문이 열렸다...") door3Open=True else: showMessage("CCTV를 켰다...") else: showMessage("어두워서 CCTV를 켤수가 없다...") cctv2.onMouseAction = cctv2_onMouseAction quit = Object('images/quit.png') quit.locate(end, 580, 20) def quit_onMouseAction(x,y,aciton): endGame() quit.onMouseAction = quit_onMouseAction #===== def startgame(): tick() inGame.enter() roomChanged(1) inGame.setLight(0.25) def finishGame(): global CyanMoving CyanMoving = False end.enter() quit.show() Sound.stop(Sound('sounds/maintheme.mp3')) Sound.play(Sound('sounds/victory.mp3')) startGame(mainScene)
Board Thread:General Discussion/@comment-35148884-20180502225645/@comment-33577126-20180503214651 Oh and top tip! Spend the rubies to double your crafting stations.
#!/usr/bin/env ruby # frozen_string_literal: true # Released under the MIT License. # Copyright, 2023, by Samuel Williams. require_relative '../../lib/async/redis' class Subscription def initialize(topic, endpoint = Async::Redis.local_endpoint) @topic = topic @endpoint = endpoint @client = nil end def client @client ||= Async::Redis::Client.new(@endpoint) end def subscribe client.subscribe(@topic) do |context| while event = context.listen yield event end end end def publish(message) client.publish @topic, message end end Sync do |task| subscription = Subscription.new("my-topic") subscriber = task.async do subscription.subscribe do |message| pp message end end 10.times do |i| subscription.publish("Hello World #{i}") end subscriber.stop end
easily observed, and have been described in detail by Huber (1810) and Forel (1874). According to Forel, 'They use their mandi bles in two ways. When closed these organs form a kind of trowel, convex in front and above, concave beneath and behind, and pointed at the tip. This trowel is used for raking up the soft earth and also for moulding and compressing their constructions and thus rendering them more solid and continuous. This is accomplished by pushing the an terior portion of the closed mandibles forward or upward. In the second place, the mandibles, when open, constitute a veritable pair of tongues with toothed edges, at least in all of the workers of our native ants that do any excavating. They thus serve not only for transporting but also for moulding or comminuting the earth." The forelegs are used for scratching up the soil, in moulding pellets and patting them down after they have been placed in position by the mandibles, and are of so much assistance in this work that when they are cut off the insects are unable to excavate or build without great difficulty and soon abandon their work altogether. Ants dislike to excavate in soil that is too dry and friable. When compelled to do this in artificial nests they will sometimes moisten it with water brought from a distance, as Miss Fielde ( 1901 ) has ob served. She says that the workers of Aphcenogaster picca. " like the Termites, are able to carry water for domestic uses. They probably lap the water into the pouch above the lower lip | the hypopharyngeal pocket] and eject it at its destination. A hundred or two of ants that I brought in and left in a heap of dry earth upon a Lubbock nest, dur ing the ensuing night took water from the surrounding moat, moistened a full pint of earth, built therein a proper nest, and were busy deposit ing their larva? in its recesses when I saw them on the following morning. As even the most extensively excavated nests represent little labor compared with the nests of social wasps and bees, ants are able to leave their homes and make new ones without serious inconvenience. Such changes are often necessitated by the habit of nesting in situa tions exposed to great and sudden changes in temperature and mois ture or to the inroads of more aggressive ants and larger terrestrial animals. Barring the intervention of such unusual conditions, how ever, most ants cling to their nests tenaciously and with every evidence of a keen sense of proprietorship, although there are a few species, besides the nomadic Dorylinse, that seem to delight in an occasional change of residence. Wasmann has shown that Formica san</iiiiica often has summer and winter residences analogous to the city and country homes of wealthy people. The ants migrate from one to the