id
int32 0
252k
| repo
stringlengths 7
55
| path
stringlengths 4
127
| func_name
stringlengths 1
88
| original_string
stringlengths 75
19.8k
| language
stringclasses 1
value | code
stringlengths 51
19.8k
| code_tokens
sequence | docstring
stringlengths 3
17.3k
| docstring_tokens
sequence | sha
stringlengths 40
40
| url
stringlengths 87
242
|
---|---|---|---|---|---|---|---|---|---|---|---|
250,900 | lawsie/guizero | guizero/base.py | Container.enable | def enable(self):
"""
Enable all the widgets in this container
"""
self._enabled = True
for child in self.children:
if isinstance(child, (Container, Widget)):
child.enable() | python | def enable(self):
self._enabled = True
for child in self.children:
if isinstance(child, (Container, Widget)):
child.enable() | [
"def",
"enable",
"(",
"self",
")",
":",
"self",
".",
"_enabled",
"=",
"True",
"for",
"child",
"in",
"self",
".",
"children",
":",
"if",
"isinstance",
"(",
"child",
",",
"(",
"Container",
",",
"Widget",
")",
")",
":",
"child",
".",
"enable",
"(",
")"
] | Enable all the widgets in this container | [
"Enable",
"all",
"the",
"widgets",
"in",
"this",
"container"
] | 84c7f0b314fa86f9fc88eb11c9a0f6c4b57155e2 | https://github.com/lawsie/guizero/blob/84c7f0b314fa86f9fc88eb11c9a0f6c4b57155e2/guizero/base.py#L451-L458 |
250,901 | lawsie/guizero | guizero/base.py | BaseWindow.exit_full_screen | def exit_full_screen(self):
"""Change from full screen to windowed mode and remove key binding"""
self.tk.attributes("-fullscreen", False)
self._full_screen = False
self.events.remove_event("<FullScreen.Escape>") | python | def exit_full_screen(self):
self.tk.attributes("-fullscreen", False)
self._full_screen = False
self.events.remove_event("<FullScreen.Escape>") | [
"def",
"exit_full_screen",
"(",
"self",
")",
":",
"self",
".",
"tk",
".",
"attributes",
"(",
"\"-fullscreen\"",
",",
"False",
")",
"self",
".",
"_full_screen",
"=",
"False",
"self",
".",
"events",
".",
"remove_event",
"(",
"\"<FullScreen.Escape>\"",
")"
] | Change from full screen to windowed mode and remove key binding | [
"Change",
"from",
"full",
"screen",
"to",
"windowed",
"mode",
"and",
"remove",
"key",
"binding"
] | 84c7f0b314fa86f9fc88eb11c9a0f6c4b57155e2 | https://github.com/lawsie/guizero/blob/84c7f0b314fa86f9fc88eb11c9a0f6c4b57155e2/guizero/base.py#L585-L589 |
250,902 | lawsie/guizero | guizero/base.py | ContainerWidget._set_propagation | def _set_propagation(self, width, height):
"""
Set the propagation value of the tk widget dependent on the width and height
:param int width:
The width of the widget.
:param int height:
The height of the widget.
"""
if width is None:
width = 0
if height is None:
height = 0
# set the propagate value
propagate_function = self.tk.pack_propagate
if self.layout == "grid":
propagate_function = self.tk.grid_propagate
propagate_value = True
# if height or width > 0 need to stop propagation
if isinstance(width, int):
if width > 0:
propagate_value = False
if isinstance(height, int):
if height > 0:
propagate_value = False
# if you specify a height or width you must specify they other
# (unless its a fill)
if isinstance(width, int) and isinstance(height, int):
if (width == 0 and height > 0) or (height == 0 and width > 0):
utils.error_format("You must specify a width and a height for {}".format(self.description))
propagate_function(propagate_value) | python | def _set_propagation(self, width, height):
if width is None:
width = 0
if height is None:
height = 0
# set the propagate value
propagate_function = self.tk.pack_propagate
if self.layout == "grid":
propagate_function = self.tk.grid_propagate
propagate_value = True
# if height or width > 0 need to stop propagation
if isinstance(width, int):
if width > 0:
propagate_value = False
if isinstance(height, int):
if height > 0:
propagate_value = False
# if you specify a height or width you must specify they other
# (unless its a fill)
if isinstance(width, int) and isinstance(height, int):
if (width == 0 and height > 0) or (height == 0 and width > 0):
utils.error_format("You must specify a width and a height for {}".format(self.description))
propagate_function(propagate_value) | [
"def",
"_set_propagation",
"(",
"self",
",",
"width",
",",
"height",
")",
":",
"if",
"width",
"is",
"None",
":",
"width",
"=",
"0",
"if",
"height",
"is",
"None",
":",
"height",
"=",
"0",
"# set the propagate value",
"propagate_function",
"=",
"self",
".",
"tk",
".",
"pack_propagate",
"if",
"self",
".",
"layout",
"==",
"\"grid\"",
":",
"propagate_function",
"=",
"self",
".",
"tk",
".",
"grid_propagate",
"propagate_value",
"=",
"True",
"# if height or width > 0 need to stop propagation",
"if",
"isinstance",
"(",
"width",
",",
"int",
")",
":",
"if",
"width",
">",
"0",
":",
"propagate_value",
"=",
"False",
"if",
"isinstance",
"(",
"height",
",",
"int",
")",
":",
"if",
"height",
">",
"0",
":",
"propagate_value",
"=",
"False",
"# if you specify a height or width you must specify they other",
"# (unless its a fill)",
"if",
"isinstance",
"(",
"width",
",",
"int",
")",
"and",
"isinstance",
"(",
"height",
",",
"int",
")",
":",
"if",
"(",
"width",
"==",
"0",
"and",
"height",
">",
"0",
")",
"or",
"(",
"height",
"==",
"0",
"and",
"width",
">",
"0",
")",
":",
"utils",
".",
"error_format",
"(",
"\"You must specify a width and a height for {}\"",
".",
"format",
"(",
"self",
".",
"description",
")",
")",
"propagate_function",
"(",
"propagate_value",
")"
] | Set the propagation value of the tk widget dependent on the width and height
:param int width:
The width of the widget.
:param int height:
The height of the widget. | [
"Set",
"the",
"propagation",
"value",
"of",
"the",
"tk",
"widget",
"dependent",
"on",
"the",
"width",
"and",
"height"
] | 84c7f0b314fa86f9fc88eb11c9a0f6c4b57155e2 | https://github.com/lawsie/guizero/blob/84c7f0b314fa86f9fc88eb11c9a0f6c4b57155e2/guizero/base.py#L661-L698 |
250,903 | angr/pyvex | pyvex/lifting/util/instr_helper.py | Instruction.load | def load(self, addr, ty):
"""
Load a value from memory into a VEX temporary register.
:param addr: The VexValue containing the addr to load from.
:param ty: The Type of the resulting data
:return: a VexValue
"""
rdt = self.irsb_c.load(addr.rdt, ty)
return VexValue(self.irsb_c, rdt) | python | def load(self, addr, ty):
rdt = self.irsb_c.load(addr.rdt, ty)
return VexValue(self.irsb_c, rdt) | [
"def",
"load",
"(",
"self",
",",
"addr",
",",
"ty",
")",
":",
"rdt",
"=",
"self",
".",
"irsb_c",
".",
"load",
"(",
"addr",
".",
"rdt",
",",
"ty",
")",
"return",
"VexValue",
"(",
"self",
".",
"irsb_c",
",",
"rdt",
")"
] | Load a value from memory into a VEX temporary register.
:param addr: The VexValue containing the addr to load from.
:param ty: The Type of the resulting data
:return: a VexValue | [
"Load",
"a",
"value",
"from",
"memory",
"into",
"a",
"VEX",
"temporary",
"register",
"."
] | c418edc1146982b2a0579bf56e5993c1c7046b19 | https://github.com/angr/pyvex/blob/c418edc1146982b2a0579bf56e5993c1c7046b19/pyvex/lifting/util/instr_helper.py#L207-L216 |
250,904 | angr/pyvex | pyvex/lifting/util/instr_helper.py | Instruction.constant | def constant(self, val, ty):
"""
Creates a constant as a VexValue
:param val: The value, as an integer
:param ty: The type of the resulting VexValue
:return: a VexValue
"""
if isinstance(val, VexValue) and not isinstance(val, IRExpr):
raise Exception('Constant cannot be made from VexValue or IRExpr')
rdt = self.irsb_c.mkconst(val, ty)
return VexValue(self.irsb_c, rdt) | python | def constant(self, val, ty):
if isinstance(val, VexValue) and not isinstance(val, IRExpr):
raise Exception('Constant cannot be made from VexValue or IRExpr')
rdt = self.irsb_c.mkconst(val, ty)
return VexValue(self.irsb_c, rdt) | [
"def",
"constant",
"(",
"self",
",",
"val",
",",
"ty",
")",
":",
"if",
"isinstance",
"(",
"val",
",",
"VexValue",
")",
"and",
"not",
"isinstance",
"(",
"val",
",",
"IRExpr",
")",
":",
"raise",
"Exception",
"(",
"'Constant cannot be made from VexValue or IRExpr'",
")",
"rdt",
"=",
"self",
".",
"irsb_c",
".",
"mkconst",
"(",
"val",
",",
"ty",
")",
"return",
"VexValue",
"(",
"self",
".",
"irsb_c",
",",
"rdt",
")"
] | Creates a constant as a VexValue
:param val: The value, as an integer
:param ty: The type of the resulting VexValue
:return: a VexValue | [
"Creates",
"a",
"constant",
"as",
"a",
"VexValue"
] | c418edc1146982b2a0579bf56e5993c1c7046b19 | https://github.com/angr/pyvex/blob/c418edc1146982b2a0579bf56e5993c1c7046b19/pyvex/lifting/util/instr_helper.py#L218-L229 |
250,905 | angr/pyvex | pyvex/lifting/util/instr_helper.py | Instruction.put | def put(self, val, reg):
"""
Puts a value from a VEX temporary register into a machine register.
This is how the results of operations done to registers get committed to the machine's state.
:param val: The VexValue to store (Want to store a constant? See Constant() first)
:param reg: The integer register number to store into, or register name
:return: None
"""
offset = self.lookup_register(self.irsb_c.irsb.arch, reg)
self.irsb_c.put(val.rdt, offset) | python | def put(self, val, reg):
offset = self.lookup_register(self.irsb_c.irsb.arch, reg)
self.irsb_c.put(val.rdt, offset) | [
"def",
"put",
"(",
"self",
",",
"val",
",",
"reg",
")",
":",
"offset",
"=",
"self",
".",
"lookup_register",
"(",
"self",
".",
"irsb_c",
".",
"irsb",
".",
"arch",
",",
"reg",
")",
"self",
".",
"irsb_c",
".",
"put",
"(",
"val",
".",
"rdt",
",",
"offset",
")"
] | Puts a value from a VEX temporary register into a machine register.
This is how the results of operations done to registers get committed to the machine's state.
:param val: The VexValue to store (Want to store a constant? See Constant() first)
:param reg: The integer register number to store into, or register name
:return: None | [
"Puts",
"a",
"value",
"from",
"a",
"VEX",
"temporary",
"register",
"into",
"a",
"machine",
"register",
".",
"This",
"is",
"how",
"the",
"results",
"of",
"operations",
"done",
"to",
"registers",
"get",
"committed",
"to",
"the",
"machine",
"s",
"state",
"."
] | c418edc1146982b2a0579bf56e5993c1c7046b19 | https://github.com/angr/pyvex/blob/c418edc1146982b2a0579bf56e5993c1c7046b19/pyvex/lifting/util/instr_helper.py#L255-L265 |
250,906 | angr/pyvex | pyvex/lifting/util/instr_helper.py | Instruction.put_conditional | def put_conditional(self, cond, valiftrue, valiffalse, reg):
"""
Like put, except it checks a condition
to decide what to put in the destination register.
:param cond: The VexValue representing the logical expression for the condition
(if your expression only has constants, don't use this method!)
:param valiftrue: the VexValue to put in reg if cond evals as true
:param validfalse: the VexValue to put in reg if cond evals as false
:param reg: The integer register number to store into, or register name
:return: None
"""
val = self.irsb_c.ite(cond.rdt , valiftrue.rdt, valiffalse.rdt)
offset = self.lookup_register(self.irsb_c.irsb.arch, reg)
self.irsb_c.put(val, offset) | python | def put_conditional(self, cond, valiftrue, valiffalse, reg):
val = self.irsb_c.ite(cond.rdt , valiftrue.rdt, valiffalse.rdt)
offset = self.lookup_register(self.irsb_c.irsb.arch, reg)
self.irsb_c.put(val, offset) | [
"def",
"put_conditional",
"(",
"self",
",",
"cond",
",",
"valiftrue",
",",
"valiffalse",
",",
"reg",
")",
":",
"val",
"=",
"self",
".",
"irsb_c",
".",
"ite",
"(",
"cond",
".",
"rdt",
",",
"valiftrue",
".",
"rdt",
",",
"valiffalse",
".",
"rdt",
")",
"offset",
"=",
"self",
".",
"lookup_register",
"(",
"self",
".",
"irsb_c",
".",
"irsb",
".",
"arch",
",",
"reg",
")",
"self",
".",
"irsb_c",
".",
"put",
"(",
"val",
",",
"offset",
")"
] | Like put, except it checks a condition
to decide what to put in the destination register.
:param cond: The VexValue representing the logical expression for the condition
(if your expression only has constants, don't use this method!)
:param valiftrue: the VexValue to put in reg if cond evals as true
:param validfalse: the VexValue to put in reg if cond evals as false
:param reg: The integer register number to store into, or register name
:return: None | [
"Like",
"put",
"except",
"it",
"checks",
"a",
"condition",
"to",
"decide",
"what",
"to",
"put",
"in",
"the",
"destination",
"register",
"."
] | c418edc1146982b2a0579bf56e5993c1c7046b19 | https://github.com/angr/pyvex/blob/c418edc1146982b2a0579bf56e5993c1c7046b19/pyvex/lifting/util/instr_helper.py#L268-L283 |
250,907 | angr/pyvex | pyvex/lifting/util/instr_helper.py | Instruction.store | def store(self, val, addr):
"""
Store a VexValue in memory at the specified loaction.
:param val: The VexValue of the value to store
:param addr: The VexValue of the address to store into
:return: None
"""
self.irsb_c.store(addr.rdt, val.rdt) | python | def store(self, val, addr):
self.irsb_c.store(addr.rdt, val.rdt) | [
"def",
"store",
"(",
"self",
",",
"val",
",",
"addr",
")",
":",
"self",
".",
"irsb_c",
".",
"store",
"(",
"addr",
".",
"rdt",
",",
"val",
".",
"rdt",
")"
] | Store a VexValue in memory at the specified loaction.
:param val: The VexValue of the value to store
:param addr: The VexValue of the address to store into
:return: None | [
"Store",
"a",
"VexValue",
"in",
"memory",
"at",
"the",
"specified",
"loaction",
"."
] | c418edc1146982b2a0579bf56e5993c1c7046b19 | https://github.com/angr/pyvex/blob/c418edc1146982b2a0579bf56e5993c1c7046b19/pyvex/lifting/util/instr_helper.py#L285-L293 |
250,908 | angr/pyvex | pyvex/lifting/util/instr_helper.py | Instruction.jump | def jump(self, condition, to_addr, jumpkind=JumpKind.Boring, ip_offset=None):
"""
Jump to a specified destination, under the specified condition.
Used for branches, jumps, calls, returns, etc.
:param condition: The VexValue representing the expression for the guard, or None for an unconditional jump
:param to_addr: The address to jump to.
:param jumpkind: The JumpKind to use. See the VEX docs for what these are; you only need them for things
aren't normal jumps (e.g., calls, interrupts, program exits, etc etc)
:return: None
"""
to_addr_ty = None
if isinstance(to_addr, VexValue):
# Unpack a VV
to_addr_rdt = to_addr.rdt
to_addr_ty = to_addr.ty
elif isinstance(to_addr, int):
# Direct jump to an int, make an RdT and Ty
to_addr_ty = vex_int_class(self.irsb_c.irsb.arch.bits).type
to_addr = self.constant(to_addr, to_addr_ty) # TODO archinfo may be changing
to_addr_rdt = to_addr.rdt
elif isinstance(to_addr, RdTmp):
# An RdT; just get the Ty of the arch's pointer type
to_addr_ty = vex_int_class(self.irsb_c.irsb.arch.bits).type
to_addr_rdt = to_addr
else:
raise ValueError("Jump destination has unknown type: " + repr(type(to_addr)))
if not condition:
# This is the default exit.
self.irsb_c.irsb.jumpkind = jumpkind
self.irsb_c.irsb.next = to_addr_rdt
else:
# add another exit
# EDG says: We should make sure folks set ArchXYZ.ip_offset like they're supposed to
if ip_offset is None:
ip_offset = self.arch.ip_offset
assert ip_offset is not None
negated_condition_rdt = self.ite(condition, self.constant(0, condition.ty), self.constant(1, condition.ty))
direct_exit_target = self.constant(self.addr + (self.bitwidth // 8), to_addr_ty)
self.irsb_c.add_exit(negated_condition_rdt, direct_exit_target.rdt, jumpkind, ip_offset)
self.irsb_c.irsb.jumpkind = jumpkind
self.irsb_c.irsb.next = to_addr_rdt | python | def jump(self, condition, to_addr, jumpkind=JumpKind.Boring, ip_offset=None):
to_addr_ty = None
if isinstance(to_addr, VexValue):
# Unpack a VV
to_addr_rdt = to_addr.rdt
to_addr_ty = to_addr.ty
elif isinstance(to_addr, int):
# Direct jump to an int, make an RdT and Ty
to_addr_ty = vex_int_class(self.irsb_c.irsb.arch.bits).type
to_addr = self.constant(to_addr, to_addr_ty) # TODO archinfo may be changing
to_addr_rdt = to_addr.rdt
elif isinstance(to_addr, RdTmp):
# An RdT; just get the Ty of the arch's pointer type
to_addr_ty = vex_int_class(self.irsb_c.irsb.arch.bits).type
to_addr_rdt = to_addr
else:
raise ValueError("Jump destination has unknown type: " + repr(type(to_addr)))
if not condition:
# This is the default exit.
self.irsb_c.irsb.jumpkind = jumpkind
self.irsb_c.irsb.next = to_addr_rdt
else:
# add another exit
# EDG says: We should make sure folks set ArchXYZ.ip_offset like they're supposed to
if ip_offset is None:
ip_offset = self.arch.ip_offset
assert ip_offset is not None
negated_condition_rdt = self.ite(condition, self.constant(0, condition.ty), self.constant(1, condition.ty))
direct_exit_target = self.constant(self.addr + (self.bitwidth // 8), to_addr_ty)
self.irsb_c.add_exit(negated_condition_rdt, direct_exit_target.rdt, jumpkind, ip_offset)
self.irsb_c.irsb.jumpkind = jumpkind
self.irsb_c.irsb.next = to_addr_rdt | [
"def",
"jump",
"(",
"self",
",",
"condition",
",",
"to_addr",
",",
"jumpkind",
"=",
"JumpKind",
".",
"Boring",
",",
"ip_offset",
"=",
"None",
")",
":",
"to_addr_ty",
"=",
"None",
"if",
"isinstance",
"(",
"to_addr",
",",
"VexValue",
")",
":",
"# Unpack a VV",
"to_addr_rdt",
"=",
"to_addr",
".",
"rdt",
"to_addr_ty",
"=",
"to_addr",
".",
"ty",
"elif",
"isinstance",
"(",
"to_addr",
",",
"int",
")",
":",
"# Direct jump to an int, make an RdT and Ty",
"to_addr_ty",
"=",
"vex_int_class",
"(",
"self",
".",
"irsb_c",
".",
"irsb",
".",
"arch",
".",
"bits",
")",
".",
"type",
"to_addr",
"=",
"self",
".",
"constant",
"(",
"to_addr",
",",
"to_addr_ty",
")",
"# TODO archinfo may be changing",
"to_addr_rdt",
"=",
"to_addr",
".",
"rdt",
"elif",
"isinstance",
"(",
"to_addr",
",",
"RdTmp",
")",
":",
"# An RdT; just get the Ty of the arch's pointer type",
"to_addr_ty",
"=",
"vex_int_class",
"(",
"self",
".",
"irsb_c",
".",
"irsb",
".",
"arch",
".",
"bits",
")",
".",
"type",
"to_addr_rdt",
"=",
"to_addr",
"else",
":",
"raise",
"ValueError",
"(",
"\"Jump destination has unknown type: \"",
"+",
"repr",
"(",
"type",
"(",
"to_addr",
")",
")",
")",
"if",
"not",
"condition",
":",
"# This is the default exit.",
"self",
".",
"irsb_c",
".",
"irsb",
".",
"jumpkind",
"=",
"jumpkind",
"self",
".",
"irsb_c",
".",
"irsb",
".",
"next",
"=",
"to_addr_rdt",
"else",
":",
"# add another exit",
"# EDG says: We should make sure folks set ArchXYZ.ip_offset like they're supposed to",
"if",
"ip_offset",
"is",
"None",
":",
"ip_offset",
"=",
"self",
".",
"arch",
".",
"ip_offset",
"assert",
"ip_offset",
"is",
"not",
"None",
"negated_condition_rdt",
"=",
"self",
".",
"ite",
"(",
"condition",
",",
"self",
".",
"constant",
"(",
"0",
",",
"condition",
".",
"ty",
")",
",",
"self",
".",
"constant",
"(",
"1",
",",
"condition",
".",
"ty",
")",
")",
"direct_exit_target",
"=",
"self",
".",
"constant",
"(",
"self",
".",
"addr",
"+",
"(",
"self",
".",
"bitwidth",
"//",
"8",
")",
",",
"to_addr_ty",
")",
"self",
".",
"irsb_c",
".",
"add_exit",
"(",
"negated_condition_rdt",
",",
"direct_exit_target",
".",
"rdt",
",",
"jumpkind",
",",
"ip_offset",
")",
"self",
".",
"irsb_c",
".",
"irsb",
".",
"jumpkind",
"=",
"jumpkind",
"self",
".",
"irsb_c",
".",
"irsb",
".",
"next",
"=",
"to_addr_rdt"
] | Jump to a specified destination, under the specified condition.
Used for branches, jumps, calls, returns, etc.
:param condition: The VexValue representing the expression for the guard, or None for an unconditional jump
:param to_addr: The address to jump to.
:param jumpkind: The JumpKind to use. See the VEX docs for what these are; you only need them for things
aren't normal jumps (e.g., calls, interrupts, program exits, etc etc)
:return: None | [
"Jump",
"to",
"a",
"specified",
"destination",
"under",
"the",
"specified",
"condition",
".",
"Used",
"for",
"branches",
"jumps",
"calls",
"returns",
"etc",
"."
] | c418edc1146982b2a0579bf56e5993c1c7046b19 | https://github.com/angr/pyvex/blob/c418edc1146982b2a0579bf56e5993c1c7046b19/pyvex/lifting/util/instr_helper.py#L295-L337 |
250,909 | angr/pyvex | pyvex/lifting/__init__.py | register | def register(lifter, arch_name):
"""
Registers a Lifter or Postprocessor to be used by pyvex. Lifters are are given priority based on the order
in which they are registered. Postprocessors will be run in registration order.
:param lifter: The Lifter or Postprocessor to register
:vartype lifter: :class:`Lifter` or :class:`Postprocessor`
"""
if issubclass(lifter, Lifter):
l.debug("Registering lifter %s for architecture %s.", lifter.__name__, arch_name)
lifters[arch_name].append(lifter)
if issubclass(lifter, Postprocessor):
l.debug("Registering postprocessor %s for architecture %s.", lifter.__name__, arch_name)
postprocessors[arch_name].append(lifter) | python | def register(lifter, arch_name):
if issubclass(lifter, Lifter):
l.debug("Registering lifter %s for architecture %s.", lifter.__name__, arch_name)
lifters[arch_name].append(lifter)
if issubclass(lifter, Postprocessor):
l.debug("Registering postprocessor %s for architecture %s.", lifter.__name__, arch_name)
postprocessors[arch_name].append(lifter) | [
"def",
"register",
"(",
"lifter",
",",
"arch_name",
")",
":",
"if",
"issubclass",
"(",
"lifter",
",",
"Lifter",
")",
":",
"l",
".",
"debug",
"(",
"\"Registering lifter %s for architecture %s.\"",
",",
"lifter",
".",
"__name__",
",",
"arch_name",
")",
"lifters",
"[",
"arch_name",
"]",
".",
"append",
"(",
"lifter",
")",
"if",
"issubclass",
"(",
"lifter",
",",
"Postprocessor",
")",
":",
"l",
".",
"debug",
"(",
"\"Registering postprocessor %s for architecture %s.\"",
",",
"lifter",
".",
"__name__",
",",
"arch_name",
")",
"postprocessors",
"[",
"arch_name",
"]",
".",
"append",
"(",
"lifter",
")"
] | Registers a Lifter or Postprocessor to be used by pyvex. Lifters are are given priority based on the order
in which they are registered. Postprocessors will be run in registration order.
:param lifter: The Lifter or Postprocessor to register
:vartype lifter: :class:`Lifter` or :class:`Postprocessor` | [
"Registers",
"a",
"Lifter",
"or",
"Postprocessor",
"to",
"be",
"used",
"by",
"pyvex",
".",
"Lifters",
"are",
"are",
"given",
"priority",
"based",
"on",
"the",
"order",
"in",
"which",
"they",
"are",
"registered",
".",
"Postprocessors",
"will",
"be",
"run",
"in",
"registration",
"order",
"."
] | c418edc1146982b2a0579bf56e5993c1c7046b19 | https://github.com/angr/pyvex/blob/c418edc1146982b2a0579bf56e5993c1c7046b19/pyvex/lifting/__init__.py#L216-L229 |
250,910 | angr/pyvex | pyvex/expr.py | IRExpr.child_expressions | def child_expressions(self):
"""
A list of all of the expressions that this expression ends up evaluating.
"""
expressions = [ ]
for k in self.__slots__:
v = getattr(self, k)
if isinstance(v, IRExpr):
expressions.append(v)
expressions.extend(v.child_expressions)
return expressions | python | def child_expressions(self):
expressions = [ ]
for k in self.__slots__:
v = getattr(self, k)
if isinstance(v, IRExpr):
expressions.append(v)
expressions.extend(v.child_expressions)
return expressions | [
"def",
"child_expressions",
"(",
"self",
")",
":",
"expressions",
"=",
"[",
"]",
"for",
"k",
"in",
"self",
".",
"__slots__",
":",
"v",
"=",
"getattr",
"(",
"self",
",",
"k",
")",
"if",
"isinstance",
"(",
"v",
",",
"IRExpr",
")",
":",
"expressions",
".",
"append",
"(",
"v",
")",
"expressions",
".",
"extend",
"(",
"v",
".",
"child_expressions",
")",
"return",
"expressions"
] | A list of all of the expressions that this expression ends up evaluating. | [
"A",
"list",
"of",
"all",
"of",
"the",
"expressions",
"that",
"this",
"expression",
"ends",
"up",
"evaluating",
"."
] | c418edc1146982b2a0579bf56e5993c1c7046b19 | https://github.com/angr/pyvex/blob/c418edc1146982b2a0579bf56e5993c1c7046b19/pyvex/expr.py#L26-L36 |
250,911 | angr/pyvex | pyvex/expr.py | IRExpr.constants | def constants(self):
"""
A list of all of the constants that this expression ends up using.
"""
constants = [ ]
for k in self.__slots__:
v = getattr(self, k)
if isinstance(v, IRExpr):
constants.extend(v.constants)
elif isinstance(v, IRConst):
constants.append(v)
return constants | python | def constants(self):
constants = [ ]
for k in self.__slots__:
v = getattr(self, k)
if isinstance(v, IRExpr):
constants.extend(v.constants)
elif isinstance(v, IRConst):
constants.append(v)
return constants | [
"def",
"constants",
"(",
"self",
")",
":",
"constants",
"=",
"[",
"]",
"for",
"k",
"in",
"self",
".",
"__slots__",
":",
"v",
"=",
"getattr",
"(",
"self",
",",
"k",
")",
"if",
"isinstance",
"(",
"v",
",",
"IRExpr",
")",
":",
"constants",
".",
"extend",
"(",
"v",
".",
"constants",
")",
"elif",
"isinstance",
"(",
"v",
",",
"IRConst",
")",
":",
"constants",
".",
"append",
"(",
"v",
")",
"return",
"constants"
] | A list of all of the constants that this expression ends up using. | [
"A",
"list",
"of",
"all",
"of",
"the",
"constants",
"that",
"this",
"expression",
"ends",
"up",
"using",
"."
] | c418edc1146982b2a0579bf56e5993c1c7046b19 | https://github.com/angr/pyvex/blob/c418edc1146982b2a0579bf56e5993c1c7046b19/pyvex/expr.py#L39-L50 |
250,912 | angr/pyvex | pyvex/block.py | IRSB.expressions | def expressions(self):
"""
Return an iterator of all expressions contained in the IRSB.
"""
for s in self.statements:
for expr_ in s.expressions:
yield expr_
yield self.next | python | def expressions(self):
for s in self.statements:
for expr_ in s.expressions:
yield expr_
yield self.next | [
"def",
"expressions",
"(",
"self",
")",
":",
"for",
"s",
"in",
"self",
".",
"statements",
":",
"for",
"expr_",
"in",
"s",
".",
"expressions",
":",
"yield",
"expr_",
"yield",
"self",
".",
"next"
] | Return an iterator of all expressions contained in the IRSB. | [
"Return",
"an",
"iterator",
"of",
"all",
"expressions",
"contained",
"in",
"the",
"IRSB",
"."
] | c418edc1146982b2a0579bf56e5993c1c7046b19 | https://github.com/angr/pyvex/blob/c418edc1146982b2a0579bf56e5993c1c7046b19/pyvex/block.py#L338-L345 |
250,913 | angr/pyvex | pyvex/block.py | IRSB.instructions | def instructions(self):
"""
The number of instructions in this block
"""
if self._instructions is None:
if self.statements is None:
self._instructions = 0
else:
self._instructions = len([s for s in self.statements if type(s) is stmt.IMark])
return self._instructions | python | def instructions(self):
if self._instructions is None:
if self.statements is None:
self._instructions = 0
else:
self._instructions = len([s for s in self.statements if type(s) is stmt.IMark])
return self._instructions | [
"def",
"instructions",
"(",
"self",
")",
":",
"if",
"self",
".",
"_instructions",
"is",
"None",
":",
"if",
"self",
".",
"statements",
"is",
"None",
":",
"self",
".",
"_instructions",
"=",
"0",
"else",
":",
"self",
".",
"_instructions",
"=",
"len",
"(",
"[",
"s",
"for",
"s",
"in",
"self",
".",
"statements",
"if",
"type",
"(",
"s",
")",
"is",
"stmt",
".",
"IMark",
"]",
")",
"return",
"self",
".",
"_instructions"
] | The number of instructions in this block | [
"The",
"number",
"of",
"instructions",
"in",
"this",
"block"
] | c418edc1146982b2a0579bf56e5993c1c7046b19 | https://github.com/angr/pyvex/blob/c418edc1146982b2a0579bf56e5993c1c7046b19/pyvex/block.py#L348-L357 |
250,914 | angr/pyvex | pyvex/block.py | IRSB.instruction_addresses | def instruction_addresses(self):
"""
Addresses of instructions in this block.
"""
if self._instruction_addresses is None:
if self.statements is None:
self._instruction_addresses = [ ]
else:
self._instruction_addresses = [ (s.addr + s.delta) for s in self.statements if type(s) is stmt.IMark ]
return self._instruction_addresses | python | def instruction_addresses(self):
if self._instruction_addresses is None:
if self.statements is None:
self._instruction_addresses = [ ]
else:
self._instruction_addresses = [ (s.addr + s.delta) for s in self.statements if type(s) is stmt.IMark ]
return self._instruction_addresses | [
"def",
"instruction_addresses",
"(",
"self",
")",
":",
"if",
"self",
".",
"_instruction_addresses",
"is",
"None",
":",
"if",
"self",
".",
"statements",
"is",
"None",
":",
"self",
".",
"_instruction_addresses",
"=",
"[",
"]",
"else",
":",
"self",
".",
"_instruction_addresses",
"=",
"[",
"(",
"s",
".",
"addr",
"+",
"s",
".",
"delta",
")",
"for",
"s",
"in",
"self",
".",
"statements",
"if",
"type",
"(",
"s",
")",
"is",
"stmt",
".",
"IMark",
"]",
"return",
"self",
".",
"_instruction_addresses"
] | Addresses of instructions in this block. | [
"Addresses",
"of",
"instructions",
"in",
"this",
"block",
"."
] | c418edc1146982b2a0579bf56e5993c1c7046b19 | https://github.com/angr/pyvex/blob/c418edc1146982b2a0579bf56e5993c1c7046b19/pyvex/block.py#L360-L369 |
250,915 | angr/pyvex | pyvex/block.py | IRSB.size | def size(self):
"""
The size of this block, in bytes
"""
if self._size is None:
self._size = sum(s.len for s in self.statements if type(s) is stmt.IMark)
return self._size | python | def size(self):
if self._size is None:
self._size = sum(s.len for s in self.statements if type(s) is stmt.IMark)
return self._size | [
"def",
"size",
"(",
"self",
")",
":",
"if",
"self",
".",
"_size",
"is",
"None",
":",
"self",
".",
"_size",
"=",
"sum",
"(",
"s",
".",
"len",
"for",
"s",
"in",
"self",
".",
"statements",
"if",
"type",
"(",
"s",
")",
"is",
"stmt",
".",
"IMark",
")",
"return",
"self",
".",
"_size"
] | The size of this block, in bytes | [
"The",
"size",
"of",
"this",
"block",
"in",
"bytes"
] | c418edc1146982b2a0579bf56e5993c1c7046b19 | https://github.com/angr/pyvex/blob/c418edc1146982b2a0579bf56e5993c1c7046b19/pyvex/block.py#L372-L378 |
250,916 | angr/pyvex | pyvex/block.py | IRSB.operations | def operations(self):
"""
A list of all operations done by the IRSB, as libVEX enum names
"""
ops = []
for e in self.expressions:
if hasattr(e, 'op'):
ops.append(e.op)
return ops | python | def operations(self):
ops = []
for e in self.expressions:
if hasattr(e, 'op'):
ops.append(e.op)
return ops | [
"def",
"operations",
"(",
"self",
")",
":",
"ops",
"=",
"[",
"]",
"for",
"e",
"in",
"self",
".",
"expressions",
":",
"if",
"hasattr",
"(",
"e",
",",
"'op'",
")",
":",
"ops",
".",
"append",
"(",
"e",
".",
"op",
")",
"return",
"ops"
] | A list of all operations done by the IRSB, as libVEX enum names | [
"A",
"list",
"of",
"all",
"operations",
"done",
"by",
"the",
"IRSB",
"as",
"libVEX",
"enum",
"names"
] | c418edc1146982b2a0579bf56e5993c1c7046b19 | https://github.com/angr/pyvex/blob/c418edc1146982b2a0579bf56e5993c1c7046b19/pyvex/block.py#L381-L389 |
250,917 | angr/pyvex | pyvex/block.py | IRSB.constant_jump_targets | def constant_jump_targets(self):
"""
A set of the static jump targets of the basic block.
"""
exits = set()
if self.exit_statements:
for _, _, stmt_ in self.exit_statements:
exits.add(stmt_.dst.value)
default_target = self.default_exit_target
if default_target is not None:
exits.add(default_target)
return exits | python | def constant_jump_targets(self):
exits = set()
if self.exit_statements:
for _, _, stmt_ in self.exit_statements:
exits.add(stmt_.dst.value)
default_target = self.default_exit_target
if default_target is not None:
exits.add(default_target)
return exits | [
"def",
"constant_jump_targets",
"(",
"self",
")",
":",
"exits",
"=",
"set",
"(",
")",
"if",
"self",
".",
"exit_statements",
":",
"for",
"_",
",",
"_",
",",
"stmt_",
"in",
"self",
".",
"exit_statements",
":",
"exits",
".",
"add",
"(",
"stmt_",
".",
"dst",
".",
"value",
")",
"default_target",
"=",
"self",
".",
"default_exit_target",
"if",
"default_target",
"is",
"not",
"None",
":",
"exits",
".",
"add",
"(",
"default_target",
")",
"return",
"exits"
] | A set of the static jump targets of the basic block. | [
"A",
"set",
"of",
"the",
"static",
"jump",
"targets",
"of",
"the",
"basic",
"block",
"."
] | c418edc1146982b2a0579bf56e5993c1c7046b19 | https://github.com/angr/pyvex/blob/c418edc1146982b2a0579bf56e5993c1c7046b19/pyvex/block.py#L407-L421 |
250,918 | angr/pyvex | pyvex/block.py | IRSB.constant_jump_targets_and_jumpkinds | def constant_jump_targets_and_jumpkinds(self):
"""
A dict of the static jump targets of the basic block to their jumpkind.
"""
exits = dict()
if self.exit_statements:
for _, _, stmt_ in self.exit_statements:
exits[stmt_.dst.value] = stmt_.jumpkind
default_target = self.default_exit_target
if default_target is not None:
exits[default_target] = self.jumpkind
return exits | python | def constant_jump_targets_and_jumpkinds(self):
exits = dict()
if self.exit_statements:
for _, _, stmt_ in self.exit_statements:
exits[stmt_.dst.value] = stmt_.jumpkind
default_target = self.default_exit_target
if default_target is not None:
exits[default_target] = self.jumpkind
return exits | [
"def",
"constant_jump_targets_and_jumpkinds",
"(",
"self",
")",
":",
"exits",
"=",
"dict",
"(",
")",
"if",
"self",
".",
"exit_statements",
":",
"for",
"_",
",",
"_",
",",
"stmt_",
"in",
"self",
".",
"exit_statements",
":",
"exits",
"[",
"stmt_",
".",
"dst",
".",
"value",
"]",
"=",
"stmt_",
".",
"jumpkind",
"default_target",
"=",
"self",
".",
"default_exit_target",
"if",
"default_target",
"is",
"not",
"None",
":",
"exits",
"[",
"default_target",
"]",
"=",
"self",
".",
"jumpkind",
"return",
"exits"
] | A dict of the static jump targets of the basic block to their jumpkind. | [
"A",
"dict",
"of",
"the",
"static",
"jump",
"targets",
"of",
"the",
"basic",
"block",
"to",
"their",
"jumpkind",
"."
] | c418edc1146982b2a0579bf56e5993c1c7046b19 | https://github.com/angr/pyvex/blob/c418edc1146982b2a0579bf56e5993c1c7046b19/pyvex/block.py#L424-L438 |
250,919 | angr/pyvex | pyvex/block.py | IRSB._pp_str | def _pp_str(self):
"""
Return the pretty-printed IRSB.
:rtype: str
"""
sa = []
sa.append("IRSB {")
if self.statements is not None:
sa.append(" %s" % self.tyenv)
sa.append("")
if self.statements is not None:
for i, s in enumerate(self.statements):
if isinstance(s, stmt.Put):
stmt_str = s.__str__(reg_name=self.arch.translate_register_name(s.offset, s.data.result_size(self.tyenv) // 8))
elif isinstance(s, stmt.WrTmp) and isinstance(s.data, expr.Get):
stmt_str = s.__str__(reg_name=self.arch.translate_register_name(s.data.offset, s.data.result_size(self.tyenv) // 8))
elif isinstance(s, stmt.Exit):
stmt_str = s.__str__(reg_name=self.arch.translate_register_name(s.offsIP, self.arch.bits // 8))
else:
stmt_str = s.__str__()
sa.append(" %02d | %s" % (i, stmt_str))
else:
sa.append(" Statements are omitted.")
sa.append(
" NEXT: PUT(%s) = %s; %s" % (self.arch.translate_register_name(self.offsIP), self.next, self.jumpkind))
sa.append("}")
return '\n'.join(sa) | python | def _pp_str(self):
sa = []
sa.append("IRSB {")
if self.statements is not None:
sa.append(" %s" % self.tyenv)
sa.append("")
if self.statements is not None:
for i, s in enumerate(self.statements):
if isinstance(s, stmt.Put):
stmt_str = s.__str__(reg_name=self.arch.translate_register_name(s.offset, s.data.result_size(self.tyenv) // 8))
elif isinstance(s, stmt.WrTmp) and isinstance(s.data, expr.Get):
stmt_str = s.__str__(reg_name=self.arch.translate_register_name(s.data.offset, s.data.result_size(self.tyenv) // 8))
elif isinstance(s, stmt.Exit):
stmt_str = s.__str__(reg_name=self.arch.translate_register_name(s.offsIP, self.arch.bits // 8))
else:
stmt_str = s.__str__()
sa.append(" %02d | %s" % (i, stmt_str))
else:
sa.append(" Statements are omitted.")
sa.append(
" NEXT: PUT(%s) = %s; %s" % (self.arch.translate_register_name(self.offsIP), self.next, self.jumpkind))
sa.append("}")
return '\n'.join(sa) | [
"def",
"_pp_str",
"(",
"self",
")",
":",
"sa",
"=",
"[",
"]",
"sa",
".",
"append",
"(",
"\"IRSB {\"",
")",
"if",
"self",
".",
"statements",
"is",
"not",
"None",
":",
"sa",
".",
"append",
"(",
"\" %s\"",
"%",
"self",
".",
"tyenv",
")",
"sa",
".",
"append",
"(",
"\"\"",
")",
"if",
"self",
".",
"statements",
"is",
"not",
"None",
":",
"for",
"i",
",",
"s",
"in",
"enumerate",
"(",
"self",
".",
"statements",
")",
":",
"if",
"isinstance",
"(",
"s",
",",
"stmt",
".",
"Put",
")",
":",
"stmt_str",
"=",
"s",
".",
"__str__",
"(",
"reg_name",
"=",
"self",
".",
"arch",
".",
"translate_register_name",
"(",
"s",
".",
"offset",
",",
"s",
".",
"data",
".",
"result_size",
"(",
"self",
".",
"tyenv",
")",
"//",
"8",
")",
")",
"elif",
"isinstance",
"(",
"s",
",",
"stmt",
".",
"WrTmp",
")",
"and",
"isinstance",
"(",
"s",
".",
"data",
",",
"expr",
".",
"Get",
")",
":",
"stmt_str",
"=",
"s",
".",
"__str__",
"(",
"reg_name",
"=",
"self",
".",
"arch",
".",
"translate_register_name",
"(",
"s",
".",
"data",
".",
"offset",
",",
"s",
".",
"data",
".",
"result_size",
"(",
"self",
".",
"tyenv",
")",
"//",
"8",
")",
")",
"elif",
"isinstance",
"(",
"s",
",",
"stmt",
".",
"Exit",
")",
":",
"stmt_str",
"=",
"s",
".",
"__str__",
"(",
"reg_name",
"=",
"self",
".",
"arch",
".",
"translate_register_name",
"(",
"s",
".",
"offsIP",
",",
"self",
".",
"arch",
".",
"bits",
"//",
"8",
")",
")",
"else",
":",
"stmt_str",
"=",
"s",
".",
"__str__",
"(",
")",
"sa",
".",
"append",
"(",
"\" %02d | %s\"",
"%",
"(",
"i",
",",
"stmt_str",
")",
")",
"else",
":",
"sa",
".",
"append",
"(",
"\" Statements are omitted.\"",
")",
"sa",
".",
"append",
"(",
"\" NEXT: PUT(%s) = %s; %s\"",
"%",
"(",
"self",
".",
"arch",
".",
"translate_register_name",
"(",
"self",
".",
"offsIP",
")",
",",
"self",
".",
"next",
",",
"self",
".",
"jumpkind",
")",
")",
"sa",
".",
"append",
"(",
"\"}\"",
")",
"return",
"'\\n'",
".",
"join",
"(",
"sa",
")"
] | Return the pretty-printed IRSB.
:rtype: str | [
"Return",
"the",
"pretty",
"-",
"printed",
"IRSB",
"."
] | c418edc1146982b2a0579bf56e5993c1c7046b19 | https://github.com/angr/pyvex/blob/c418edc1146982b2a0579bf56e5993c1c7046b19/pyvex/block.py#L444-L471 |
250,920 | angr/pyvex | pyvex/block.py | IRSB._is_defaultexit_direct_jump | def _is_defaultexit_direct_jump(self):
"""
Checks if the default of this IRSB a direct jump or not.
"""
if not (self.jumpkind == 'Ijk_InvalICache' or self.jumpkind == 'Ijk_Boring' or self.jumpkind == 'Ijk_Call'):
return False
target = self.default_exit_target
return target is not None | python | def _is_defaultexit_direct_jump(self):
if not (self.jumpkind == 'Ijk_InvalICache' or self.jumpkind == 'Ijk_Boring' or self.jumpkind == 'Ijk_Call'):
return False
target = self.default_exit_target
return target is not None | [
"def",
"_is_defaultexit_direct_jump",
"(",
"self",
")",
":",
"if",
"not",
"(",
"self",
".",
"jumpkind",
"==",
"'Ijk_InvalICache'",
"or",
"self",
".",
"jumpkind",
"==",
"'Ijk_Boring'",
"or",
"self",
".",
"jumpkind",
"==",
"'Ijk_Call'",
")",
":",
"return",
"False",
"target",
"=",
"self",
".",
"default_exit_target",
"return",
"target",
"is",
"not",
"None"
] | Checks if the default of this IRSB a direct jump or not. | [
"Checks",
"if",
"the",
"default",
"of",
"this",
"IRSB",
"a",
"direct",
"jump",
"or",
"not",
"."
] | c418edc1146982b2a0579bf56e5993c1c7046b19 | https://github.com/angr/pyvex/blob/c418edc1146982b2a0579bf56e5993c1c7046b19/pyvex/block.py#L473-L481 |
250,921 | angr/pyvex | pyvex/block.py | IRTypeEnv.lookup | def lookup(self, tmp):
"""
Return the type of temporary variable `tmp` as an enum string
"""
if tmp < 0 or tmp > self.types_used:
l.debug("Invalid temporary number %d", tmp)
raise IndexError(tmp)
return self.types[tmp] | python | def lookup(self, tmp):
if tmp < 0 or tmp > self.types_used:
l.debug("Invalid temporary number %d", tmp)
raise IndexError(tmp)
return self.types[tmp] | [
"def",
"lookup",
"(",
"self",
",",
"tmp",
")",
":",
"if",
"tmp",
"<",
"0",
"or",
"tmp",
">",
"self",
".",
"types_used",
":",
"l",
".",
"debug",
"(",
"\"Invalid temporary number %d\"",
",",
"tmp",
")",
"raise",
"IndexError",
"(",
"tmp",
")",
"return",
"self",
".",
"types",
"[",
"tmp",
"]"
] | Return the type of temporary variable `tmp` as an enum string | [
"Return",
"the",
"type",
"of",
"temporary",
"variable",
"tmp",
"as",
"an",
"enum",
"string"
] | c418edc1146982b2a0579bf56e5993c1c7046b19 | https://github.com/angr/pyvex/blob/c418edc1146982b2a0579bf56e5993c1c7046b19/pyvex/block.py#L568-L575 |
250,922 | angr/pyvex | pyvex/lifting/lifter.py | Lifter._lift | def _lift(self,
data,
bytes_offset=None,
max_bytes=None,
max_inst=None,
opt_level=1,
traceflags=None,
allow_arch_optimizations=None,
strict_block_end=None,
skip_stmts=False,
collect_data_refs=False):
"""
Wrapper around the `lift` method on Lifters. Should not be overridden in child classes.
:param data: The bytes to lift as either a python string of bytes or a cffi buffer object.
:param bytes_offset: The offset into `data` to start lifting at.
:param max_bytes: The maximum number of bytes to lift. If set to None, no byte limit is used.
:param max_inst: The maximum number of instructions to lift. If set to None, no instruction limit is used.
:param opt_level: The level of optimization to apply to the IR, 0-2. Most likely will be ignored in any lifter
other then LibVEX.
:param traceflags: The libVEX traceflags, controlling VEX debug prints. Most likely will be ignored in any
lifter other than LibVEX.
:param allow_arch_optimizations: Should the LibVEX lifter be allowed to perform lift-time preprocessing optimizations
(e.g., lookback ITSTATE optimization on THUMB)
Most likely will be ignored in any lifter other than LibVEX.
:param strict_block_end: Should the LibVEX arm-thumb split block at some instructions, for example CB{N}Z.
:param skip_stmts: Should the lifter skip transferring IRStmts from C to Python.
:param collect_data_refs: Should the LibVEX lifter collect data references in C.
"""
irsb = IRSB.empty_block(self.arch, self.addr)
self.data = data
self.bytes_offset = bytes_offset
self.opt_level = opt_level
self.traceflags = traceflags
self.allow_arch_optimizations = allow_arch_optimizations
self.strict_block_end = strict_block_end
self.collect_data_refs = collect_data_refs
self.max_inst = max_inst
self.max_bytes = max_bytes
self.skip_stmts = skip_stmts
self.irsb = irsb
self.lift()
return self.irsb | python | def _lift(self,
data,
bytes_offset=None,
max_bytes=None,
max_inst=None,
opt_level=1,
traceflags=None,
allow_arch_optimizations=None,
strict_block_end=None,
skip_stmts=False,
collect_data_refs=False):
irsb = IRSB.empty_block(self.arch, self.addr)
self.data = data
self.bytes_offset = bytes_offset
self.opt_level = opt_level
self.traceflags = traceflags
self.allow_arch_optimizations = allow_arch_optimizations
self.strict_block_end = strict_block_end
self.collect_data_refs = collect_data_refs
self.max_inst = max_inst
self.max_bytes = max_bytes
self.skip_stmts = skip_stmts
self.irsb = irsb
self.lift()
return self.irsb | [
"def",
"_lift",
"(",
"self",
",",
"data",
",",
"bytes_offset",
"=",
"None",
",",
"max_bytes",
"=",
"None",
",",
"max_inst",
"=",
"None",
",",
"opt_level",
"=",
"1",
",",
"traceflags",
"=",
"None",
",",
"allow_arch_optimizations",
"=",
"None",
",",
"strict_block_end",
"=",
"None",
",",
"skip_stmts",
"=",
"False",
",",
"collect_data_refs",
"=",
"False",
")",
":",
"irsb",
"=",
"IRSB",
".",
"empty_block",
"(",
"self",
".",
"arch",
",",
"self",
".",
"addr",
")",
"self",
".",
"data",
"=",
"data",
"self",
".",
"bytes_offset",
"=",
"bytes_offset",
"self",
".",
"opt_level",
"=",
"opt_level",
"self",
".",
"traceflags",
"=",
"traceflags",
"self",
".",
"allow_arch_optimizations",
"=",
"allow_arch_optimizations",
"self",
".",
"strict_block_end",
"=",
"strict_block_end",
"self",
".",
"collect_data_refs",
"=",
"collect_data_refs",
"self",
".",
"max_inst",
"=",
"max_inst",
"self",
".",
"max_bytes",
"=",
"max_bytes",
"self",
".",
"skip_stmts",
"=",
"skip_stmts",
"self",
".",
"irsb",
"=",
"irsb",
"self",
".",
"lift",
"(",
")",
"return",
"self",
".",
"irsb"
] | Wrapper around the `lift` method on Lifters. Should not be overridden in child classes.
:param data: The bytes to lift as either a python string of bytes or a cffi buffer object.
:param bytes_offset: The offset into `data` to start lifting at.
:param max_bytes: The maximum number of bytes to lift. If set to None, no byte limit is used.
:param max_inst: The maximum number of instructions to lift. If set to None, no instruction limit is used.
:param opt_level: The level of optimization to apply to the IR, 0-2. Most likely will be ignored in any lifter
other then LibVEX.
:param traceflags: The libVEX traceflags, controlling VEX debug prints. Most likely will be ignored in any
lifter other than LibVEX.
:param allow_arch_optimizations: Should the LibVEX lifter be allowed to perform lift-time preprocessing optimizations
(e.g., lookback ITSTATE optimization on THUMB)
Most likely will be ignored in any lifter other than LibVEX.
:param strict_block_end: Should the LibVEX arm-thumb split block at some instructions, for example CB{N}Z.
:param skip_stmts: Should the lifter skip transferring IRStmts from C to Python.
:param collect_data_refs: Should the LibVEX lifter collect data references in C. | [
"Wrapper",
"around",
"the",
"lift",
"method",
"on",
"Lifters",
".",
"Should",
"not",
"be",
"overridden",
"in",
"child",
"classes",
"."
] | c418edc1146982b2a0579bf56e5993c1c7046b19 | https://github.com/angr/pyvex/blob/c418edc1146982b2a0579bf56e5993c1c7046b19/pyvex/lifting/lifter.py#L36-L78 |
250,923 | TeamHG-Memex/scrapy-rotating-proxies | rotating_proxies/expire.py | exp_backoff | def exp_backoff(attempt, cap=3600, base=300):
""" Exponential backoff time """
# this is a numerically stable version of
# min(cap, base * 2 ** attempt)
max_attempts = math.log(cap / base, 2)
if attempt <= max_attempts:
return base * 2 ** attempt
return cap | python | def exp_backoff(attempt, cap=3600, base=300):
# this is a numerically stable version of
# min(cap, base * 2 ** attempt)
max_attempts = math.log(cap / base, 2)
if attempt <= max_attempts:
return base * 2 ** attempt
return cap | [
"def",
"exp_backoff",
"(",
"attempt",
",",
"cap",
"=",
"3600",
",",
"base",
"=",
"300",
")",
":",
"# this is a numerically stable version of",
"# min(cap, base * 2 ** attempt)",
"max_attempts",
"=",
"math",
".",
"log",
"(",
"cap",
"/",
"base",
",",
"2",
")",
"if",
"attempt",
"<=",
"max_attempts",
":",
"return",
"base",
"*",
"2",
"**",
"attempt",
"return",
"cap"
] | Exponential backoff time | [
"Exponential",
"backoff",
"time"
] | 89bb677fea6285a5e02e0a5c7dfb1c40330b17f0 | https://github.com/TeamHG-Memex/scrapy-rotating-proxies/blob/89bb677fea6285a5e02e0a5c7dfb1c40330b17f0/rotating_proxies/expire.py#L149-L156 |
250,924 | TeamHG-Memex/scrapy-rotating-proxies | rotating_proxies/expire.py | Proxies.get_proxy | def get_proxy(self, proxy_address):
"""
Return complete proxy name associated with a hostport of a given
``proxy_address``. If ``proxy_address`` is unkonwn or empty,
return None.
"""
if not proxy_address:
return None
hostport = extract_proxy_hostport(proxy_address)
return self.proxies_by_hostport.get(hostport, None) | python | def get_proxy(self, proxy_address):
if not proxy_address:
return None
hostport = extract_proxy_hostport(proxy_address)
return self.proxies_by_hostport.get(hostport, None) | [
"def",
"get_proxy",
"(",
"self",
",",
"proxy_address",
")",
":",
"if",
"not",
"proxy_address",
":",
"return",
"None",
"hostport",
"=",
"extract_proxy_hostport",
"(",
"proxy_address",
")",
"return",
"self",
".",
"proxies_by_hostport",
".",
"get",
"(",
"hostport",
",",
"None",
")"
] | Return complete proxy name associated with a hostport of a given
``proxy_address``. If ``proxy_address`` is unkonwn or empty,
return None. | [
"Return",
"complete",
"proxy",
"name",
"associated",
"with",
"a",
"hostport",
"of",
"a",
"given",
"proxy_address",
".",
"If",
"proxy_address",
"is",
"unkonwn",
"or",
"empty",
"return",
"None",
"."
] | 89bb677fea6285a5e02e0a5c7dfb1c40330b17f0 | https://github.com/TeamHG-Memex/scrapy-rotating-proxies/blob/89bb677fea6285a5e02e0a5c7dfb1c40330b17f0/rotating_proxies/expire.py#L56-L65 |
250,925 | TeamHG-Memex/scrapy-rotating-proxies | rotating_proxies/expire.py | Proxies.mark_dead | def mark_dead(self, proxy, _time=None):
""" Mark a proxy as dead """
if proxy not in self.proxies:
logger.warn("Proxy <%s> was not found in proxies list" % proxy)
return
if proxy in self.good:
logger.debug("GOOD proxy became DEAD: <%s>" % proxy)
else:
logger.debug("Proxy <%s> is DEAD" % proxy)
self.unchecked.discard(proxy)
self.good.discard(proxy)
self.dead.add(proxy)
now = _time or time.time()
state = self.proxies[proxy]
state.backoff_time = self.backoff(state.failed_attempts)
state.next_check = now + state.backoff_time
state.failed_attempts += 1 | python | def mark_dead(self, proxy, _time=None):
if proxy not in self.proxies:
logger.warn("Proxy <%s> was not found in proxies list" % proxy)
return
if proxy in self.good:
logger.debug("GOOD proxy became DEAD: <%s>" % proxy)
else:
logger.debug("Proxy <%s> is DEAD" % proxy)
self.unchecked.discard(proxy)
self.good.discard(proxy)
self.dead.add(proxy)
now = _time or time.time()
state = self.proxies[proxy]
state.backoff_time = self.backoff(state.failed_attempts)
state.next_check = now + state.backoff_time
state.failed_attempts += 1 | [
"def",
"mark_dead",
"(",
"self",
",",
"proxy",
",",
"_time",
"=",
"None",
")",
":",
"if",
"proxy",
"not",
"in",
"self",
".",
"proxies",
":",
"logger",
".",
"warn",
"(",
"\"Proxy <%s> was not found in proxies list\"",
"%",
"proxy",
")",
"return",
"if",
"proxy",
"in",
"self",
".",
"good",
":",
"logger",
".",
"debug",
"(",
"\"GOOD proxy became DEAD: <%s>\"",
"%",
"proxy",
")",
"else",
":",
"logger",
".",
"debug",
"(",
"\"Proxy <%s> is DEAD\"",
"%",
"proxy",
")",
"self",
".",
"unchecked",
".",
"discard",
"(",
"proxy",
")",
"self",
".",
"good",
".",
"discard",
"(",
"proxy",
")",
"self",
".",
"dead",
".",
"add",
"(",
"proxy",
")",
"now",
"=",
"_time",
"or",
"time",
".",
"time",
"(",
")",
"state",
"=",
"self",
".",
"proxies",
"[",
"proxy",
"]",
"state",
".",
"backoff_time",
"=",
"self",
".",
"backoff",
"(",
"state",
".",
"failed_attempts",
")",
"state",
".",
"next_check",
"=",
"now",
"+",
"state",
".",
"backoff_time",
"state",
".",
"failed_attempts",
"+=",
"1"
] | Mark a proxy as dead | [
"Mark",
"a",
"proxy",
"as",
"dead"
] | 89bb677fea6285a5e02e0a5c7dfb1c40330b17f0 | https://github.com/TeamHG-Memex/scrapy-rotating-proxies/blob/89bb677fea6285a5e02e0a5c7dfb1c40330b17f0/rotating_proxies/expire.py#L67-L86 |
250,926 | TeamHG-Memex/scrapy-rotating-proxies | rotating_proxies/expire.py | Proxies.mark_good | def mark_good(self, proxy):
""" Mark a proxy as good """
if proxy not in self.proxies:
logger.warn("Proxy <%s> was not found in proxies list" % proxy)
return
if proxy not in self.good:
logger.debug("Proxy <%s> is GOOD" % proxy)
self.unchecked.discard(proxy)
self.dead.discard(proxy)
self.good.add(proxy)
self.proxies[proxy].failed_attempts = 0 | python | def mark_good(self, proxy):
if proxy not in self.proxies:
logger.warn("Proxy <%s> was not found in proxies list" % proxy)
return
if proxy not in self.good:
logger.debug("Proxy <%s> is GOOD" % proxy)
self.unchecked.discard(proxy)
self.dead.discard(proxy)
self.good.add(proxy)
self.proxies[proxy].failed_attempts = 0 | [
"def",
"mark_good",
"(",
"self",
",",
"proxy",
")",
":",
"if",
"proxy",
"not",
"in",
"self",
".",
"proxies",
":",
"logger",
".",
"warn",
"(",
"\"Proxy <%s> was not found in proxies list\"",
"%",
"proxy",
")",
"return",
"if",
"proxy",
"not",
"in",
"self",
".",
"good",
":",
"logger",
".",
"debug",
"(",
"\"Proxy <%s> is GOOD\"",
"%",
"proxy",
")",
"self",
".",
"unchecked",
".",
"discard",
"(",
"proxy",
")",
"self",
".",
"dead",
".",
"discard",
"(",
"proxy",
")",
"self",
".",
"good",
".",
"add",
"(",
"proxy",
")",
"self",
".",
"proxies",
"[",
"proxy",
"]",
".",
"failed_attempts",
"=",
"0"
] | Mark a proxy as good | [
"Mark",
"a",
"proxy",
"as",
"good"
] | 89bb677fea6285a5e02e0a5c7dfb1c40330b17f0 | https://github.com/TeamHG-Memex/scrapy-rotating-proxies/blob/89bb677fea6285a5e02e0a5c7dfb1c40330b17f0/rotating_proxies/expire.py#L88-L100 |
250,927 | TeamHG-Memex/scrapy-rotating-proxies | rotating_proxies/expire.py | Proxies.reanimate | def reanimate(self, _time=None):
""" Move dead proxies to unchecked if a backoff timeout passes """
n_reanimated = 0
now = _time or time.time()
for proxy in list(self.dead):
state = self.proxies[proxy]
assert state.next_check is not None
if state.next_check <= now:
self.dead.remove(proxy)
self.unchecked.add(proxy)
n_reanimated += 1
return n_reanimated | python | def reanimate(self, _time=None):
n_reanimated = 0
now = _time or time.time()
for proxy in list(self.dead):
state = self.proxies[proxy]
assert state.next_check is not None
if state.next_check <= now:
self.dead.remove(proxy)
self.unchecked.add(proxy)
n_reanimated += 1
return n_reanimated | [
"def",
"reanimate",
"(",
"self",
",",
"_time",
"=",
"None",
")",
":",
"n_reanimated",
"=",
"0",
"now",
"=",
"_time",
"or",
"time",
".",
"time",
"(",
")",
"for",
"proxy",
"in",
"list",
"(",
"self",
".",
"dead",
")",
":",
"state",
"=",
"self",
".",
"proxies",
"[",
"proxy",
"]",
"assert",
"state",
".",
"next_check",
"is",
"not",
"None",
"if",
"state",
".",
"next_check",
"<=",
"now",
":",
"self",
".",
"dead",
".",
"remove",
"(",
"proxy",
")",
"self",
".",
"unchecked",
".",
"add",
"(",
"proxy",
")",
"n_reanimated",
"+=",
"1",
"return",
"n_reanimated"
] | Move dead proxies to unchecked if a backoff timeout passes | [
"Move",
"dead",
"proxies",
"to",
"unchecked",
"if",
"a",
"backoff",
"timeout",
"passes"
] | 89bb677fea6285a5e02e0a5c7dfb1c40330b17f0 | https://github.com/TeamHG-Memex/scrapy-rotating-proxies/blob/89bb677fea6285a5e02e0a5c7dfb1c40330b17f0/rotating_proxies/expire.py#L102-L113 |
250,928 | TeamHG-Memex/scrapy-rotating-proxies | rotating_proxies/expire.py | Proxies.reset | def reset(self):
""" Mark all dead proxies as unchecked """
for proxy in list(self.dead):
self.dead.remove(proxy)
self.unchecked.add(proxy) | python | def reset(self):
for proxy in list(self.dead):
self.dead.remove(proxy)
self.unchecked.add(proxy) | [
"def",
"reset",
"(",
"self",
")",
":",
"for",
"proxy",
"in",
"list",
"(",
"self",
".",
"dead",
")",
":",
"self",
".",
"dead",
".",
"remove",
"(",
"proxy",
")",
"self",
".",
"unchecked",
".",
"add",
"(",
"proxy",
")"
] | Mark all dead proxies as unchecked | [
"Mark",
"all",
"dead",
"proxies",
"as",
"unchecked"
] | 89bb677fea6285a5e02e0a5c7dfb1c40330b17f0 | https://github.com/TeamHG-Memex/scrapy-rotating-proxies/blob/89bb677fea6285a5e02e0a5c7dfb1c40330b17f0/rotating_proxies/expire.py#L115-L119 |
250,929 | boakley/robotframework-hub | rfhub/kwdb.py | KeywordTable.on_change | def on_change(self, path, event_type):
"""Respond to changes in the file system
This method will be given the path to a file that
has changed on disk. We need to reload the keywords
from that file
"""
# I can do all this work in a sql statement, but
# for debugging it's easier to do it in stages.
sql = """SELECT collection_id
FROM collection_table
WHERE path == ?
"""
cursor = self._execute(sql, (path,))
results = cursor.fetchall()
# there should always be exactly one result, but
# there's no harm in using a loop to process the
# single result
for result in results:
collection_id = result[0]
# remove all keywords in this collection
sql = """DELETE from keyword_table
WHERE collection_id == ?
"""
cursor = self._execute(sql, (collection_id,))
self._load_keywords(collection_id, path=path) | python | def on_change(self, path, event_type):
# I can do all this work in a sql statement, but
# for debugging it's easier to do it in stages.
sql = """SELECT collection_id
FROM collection_table
WHERE path == ?
"""
cursor = self._execute(sql, (path,))
results = cursor.fetchall()
# there should always be exactly one result, but
# there's no harm in using a loop to process the
# single result
for result in results:
collection_id = result[0]
# remove all keywords in this collection
sql = """DELETE from keyword_table
WHERE collection_id == ?
"""
cursor = self._execute(sql, (collection_id,))
self._load_keywords(collection_id, path=path) | [
"def",
"on_change",
"(",
"self",
",",
"path",
",",
"event_type",
")",
":",
"# I can do all this work in a sql statement, but",
"# for debugging it's easier to do it in stages.",
"sql",
"=",
"\"\"\"SELECT collection_id\n FROM collection_table\n WHERE path == ?\n \"\"\"",
"cursor",
"=",
"self",
".",
"_execute",
"(",
"sql",
",",
"(",
"path",
",",
")",
")",
"results",
"=",
"cursor",
".",
"fetchall",
"(",
")",
"# there should always be exactly one result, but",
"# there's no harm in using a loop to process the",
"# single result",
"for",
"result",
"in",
"results",
":",
"collection_id",
"=",
"result",
"[",
"0",
"]",
"# remove all keywords in this collection",
"sql",
"=",
"\"\"\"DELETE from keyword_table\n WHERE collection_id == ?\n \"\"\"",
"cursor",
"=",
"self",
".",
"_execute",
"(",
"sql",
",",
"(",
"collection_id",
",",
")",
")",
"self",
".",
"_load_keywords",
"(",
"collection_id",
",",
"path",
"=",
"path",
")"
] | Respond to changes in the file system
This method will be given the path to a file that
has changed on disk. We need to reload the keywords
from that file | [
"Respond",
"to",
"changes",
"in",
"the",
"file",
"system"
] | f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c | https://github.com/boakley/robotframework-hub/blob/f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c/rfhub/kwdb.py#L86-L111 |
250,930 | boakley/robotframework-hub | rfhub/kwdb.py | KeywordTable._load_keywords | def _load_keywords(self, collection_id, path=None, libdoc=None):
"""Load a collection of keywords
One of path or libdoc needs to be passed in...
"""
if libdoc is None and path is None:
raise(Exception("You must provide either a path or libdoc argument"))
if libdoc is None:
libdoc = LibraryDocumentation(path)
if len(libdoc.keywords) > 0:
for keyword in libdoc.keywords:
self._add_keyword(collection_id, keyword.name, keyword.doc, keyword.args) | python | def _load_keywords(self, collection_id, path=None, libdoc=None):
if libdoc is None and path is None:
raise(Exception("You must provide either a path or libdoc argument"))
if libdoc is None:
libdoc = LibraryDocumentation(path)
if len(libdoc.keywords) > 0:
for keyword in libdoc.keywords:
self._add_keyword(collection_id, keyword.name, keyword.doc, keyword.args) | [
"def",
"_load_keywords",
"(",
"self",
",",
"collection_id",
",",
"path",
"=",
"None",
",",
"libdoc",
"=",
"None",
")",
":",
"if",
"libdoc",
"is",
"None",
"and",
"path",
"is",
"None",
":",
"raise",
"(",
"Exception",
"(",
"\"You must provide either a path or libdoc argument\"",
")",
")",
"if",
"libdoc",
"is",
"None",
":",
"libdoc",
"=",
"LibraryDocumentation",
"(",
"path",
")",
"if",
"len",
"(",
"libdoc",
".",
"keywords",
")",
">",
"0",
":",
"for",
"keyword",
"in",
"libdoc",
".",
"keywords",
":",
"self",
".",
"_add_keyword",
"(",
"collection_id",
",",
"keyword",
".",
"name",
",",
"keyword",
".",
"doc",
",",
"keyword",
".",
"args",
")"
] | Load a collection of keywords
One of path or libdoc needs to be passed in... | [
"Load",
"a",
"collection",
"of",
"keywords"
] | f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c | https://github.com/boakley/robotframework-hub/blob/f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c/rfhub/kwdb.py#L113-L126 |
250,931 | boakley/robotframework-hub | rfhub/kwdb.py | KeywordTable.add_file | def add_file(self, path):
"""Add a resource file or library file to the database"""
libdoc = LibraryDocumentation(path)
if len(libdoc.keywords) > 0:
if libdoc.doc.startswith("Documentation for resource file"):
# bah! The file doesn't have an file-level documentation
# and libdoc substitutes some placeholder text.
libdoc.doc = ""
collection_id = self.add_collection(path, libdoc.name, libdoc.type,
libdoc.doc, libdoc.version,
libdoc.scope, libdoc.named_args,
libdoc.doc_format)
self._load_keywords(collection_id, libdoc=libdoc) | python | def add_file(self, path):
libdoc = LibraryDocumentation(path)
if len(libdoc.keywords) > 0:
if libdoc.doc.startswith("Documentation for resource file"):
# bah! The file doesn't have an file-level documentation
# and libdoc substitutes some placeholder text.
libdoc.doc = ""
collection_id = self.add_collection(path, libdoc.name, libdoc.type,
libdoc.doc, libdoc.version,
libdoc.scope, libdoc.named_args,
libdoc.doc_format)
self._load_keywords(collection_id, libdoc=libdoc) | [
"def",
"add_file",
"(",
"self",
",",
"path",
")",
":",
"libdoc",
"=",
"LibraryDocumentation",
"(",
"path",
")",
"if",
"len",
"(",
"libdoc",
".",
"keywords",
")",
">",
"0",
":",
"if",
"libdoc",
".",
"doc",
".",
"startswith",
"(",
"\"Documentation for resource file\"",
")",
":",
"# bah! The file doesn't have an file-level documentation",
"# and libdoc substitutes some placeholder text.",
"libdoc",
".",
"doc",
"=",
"\"\"",
"collection_id",
"=",
"self",
".",
"add_collection",
"(",
"path",
",",
"libdoc",
".",
"name",
",",
"libdoc",
".",
"type",
",",
"libdoc",
".",
"doc",
",",
"libdoc",
".",
"version",
",",
"libdoc",
".",
"scope",
",",
"libdoc",
".",
"named_args",
",",
"libdoc",
".",
"doc_format",
")",
"self",
".",
"_load_keywords",
"(",
"collection_id",
",",
"libdoc",
"=",
"libdoc",
")"
] | Add a resource file or library file to the database | [
"Add",
"a",
"resource",
"file",
"or",
"library",
"file",
"to",
"the",
"database"
] | f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c | https://github.com/boakley/robotframework-hub/blob/f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c/rfhub/kwdb.py#L128-L141 |
250,932 | boakley/robotframework-hub | rfhub/kwdb.py | KeywordTable.add_library | def add_library(self, name):
"""Add a library to the database
This method is for adding a library by name (eg: "BuiltIn")
rather than by a file.
"""
libdoc = LibraryDocumentation(name)
if len(libdoc.keywords) > 0:
# FIXME: figure out the path to the library file
collection_id = self.add_collection(None, libdoc.name, libdoc.type,
libdoc.doc, libdoc.version,
libdoc.scope, libdoc.named_args,
libdoc.doc_format)
self._load_keywords(collection_id, libdoc=libdoc) | python | def add_library(self, name):
libdoc = LibraryDocumentation(name)
if len(libdoc.keywords) > 0:
# FIXME: figure out the path to the library file
collection_id = self.add_collection(None, libdoc.name, libdoc.type,
libdoc.doc, libdoc.version,
libdoc.scope, libdoc.named_args,
libdoc.doc_format)
self._load_keywords(collection_id, libdoc=libdoc) | [
"def",
"add_library",
"(",
"self",
",",
"name",
")",
":",
"libdoc",
"=",
"LibraryDocumentation",
"(",
"name",
")",
"if",
"len",
"(",
"libdoc",
".",
"keywords",
")",
">",
"0",
":",
"# FIXME: figure out the path to the library file",
"collection_id",
"=",
"self",
".",
"add_collection",
"(",
"None",
",",
"libdoc",
".",
"name",
",",
"libdoc",
".",
"type",
",",
"libdoc",
".",
"doc",
",",
"libdoc",
".",
"version",
",",
"libdoc",
".",
"scope",
",",
"libdoc",
".",
"named_args",
",",
"libdoc",
".",
"doc_format",
")",
"self",
".",
"_load_keywords",
"(",
"collection_id",
",",
"libdoc",
"=",
"libdoc",
")"
] | Add a library to the database
This method is for adding a library by name (eg: "BuiltIn")
rather than by a file. | [
"Add",
"a",
"library",
"to",
"the",
"database"
] | f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c | https://github.com/boakley/robotframework-hub/blob/f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c/rfhub/kwdb.py#L143-L156 |
250,933 | boakley/robotframework-hub | rfhub/kwdb.py | KeywordTable.add_folder | def add_folder(self, dirname, watch=True):
"""Recursively add all files in a folder to the database
By "all files" I mean, "all files that are resource files
or library files". It will silently ignore files that don't
look like they belong in the database. Pity the fool who
uses non-standard suffixes.
N.B. folders with names that begin with '." will be skipped
"""
ignore_file = os.path.join(dirname, ".rfhubignore")
exclude_patterns = []
try:
with open(ignore_file, "r") as f:
exclude_patterns = []
for line in f.readlines():
line = line.strip()
if (re.match(r'^\s*#', line)): continue
if len(line.strip()) > 0:
exclude_patterns.append(line)
except:
# should probably warn the user?
pass
for filename in os.listdir(dirname):
path = os.path.join(dirname, filename)
(basename, ext) = os.path.splitext(filename.lower())
try:
if (os.path.isdir(path)):
if (not basename.startswith(".")):
if os.access(path, os.R_OK):
self.add_folder(path, watch=False)
else:
if (ext in (".xml", ".robot", ".txt", ".py", ".tsv")):
if os.access(path, os.R_OK):
self.add(path)
except Exception as e:
# I really need to get the logging situation figured out.
print("bummer:", str(e))
# FIXME:
# instead of passing a flag around, I should just keep track
# of which folders we're watching, and don't add wathers for
# any subfolders. That will work better in the case where
# the user accidentally starts up the hub giving the same
# folder, or a folder and it's children, on the command line...
if watch:
# add watcher on normalized path
dirname = os.path.abspath(dirname)
event_handler = WatchdogHandler(self, dirname)
self.observer.schedule(event_handler, dirname, recursive=True) | python | def add_folder(self, dirname, watch=True):
ignore_file = os.path.join(dirname, ".rfhubignore")
exclude_patterns = []
try:
with open(ignore_file, "r") as f:
exclude_patterns = []
for line in f.readlines():
line = line.strip()
if (re.match(r'^\s*#', line)): continue
if len(line.strip()) > 0:
exclude_patterns.append(line)
except:
# should probably warn the user?
pass
for filename in os.listdir(dirname):
path = os.path.join(dirname, filename)
(basename, ext) = os.path.splitext(filename.lower())
try:
if (os.path.isdir(path)):
if (not basename.startswith(".")):
if os.access(path, os.R_OK):
self.add_folder(path, watch=False)
else:
if (ext in (".xml", ".robot", ".txt", ".py", ".tsv")):
if os.access(path, os.R_OK):
self.add(path)
except Exception as e:
# I really need to get the logging situation figured out.
print("bummer:", str(e))
# FIXME:
# instead of passing a flag around, I should just keep track
# of which folders we're watching, and don't add wathers for
# any subfolders. That will work better in the case where
# the user accidentally starts up the hub giving the same
# folder, or a folder and it's children, on the command line...
if watch:
# add watcher on normalized path
dirname = os.path.abspath(dirname)
event_handler = WatchdogHandler(self, dirname)
self.observer.schedule(event_handler, dirname, recursive=True) | [
"def",
"add_folder",
"(",
"self",
",",
"dirname",
",",
"watch",
"=",
"True",
")",
":",
"ignore_file",
"=",
"os",
".",
"path",
".",
"join",
"(",
"dirname",
",",
"\".rfhubignore\"",
")",
"exclude_patterns",
"=",
"[",
"]",
"try",
":",
"with",
"open",
"(",
"ignore_file",
",",
"\"r\"",
")",
"as",
"f",
":",
"exclude_patterns",
"=",
"[",
"]",
"for",
"line",
"in",
"f",
".",
"readlines",
"(",
")",
":",
"line",
"=",
"line",
".",
"strip",
"(",
")",
"if",
"(",
"re",
".",
"match",
"(",
"r'^\\s*#'",
",",
"line",
")",
")",
":",
"continue",
"if",
"len",
"(",
"line",
".",
"strip",
"(",
")",
")",
">",
"0",
":",
"exclude_patterns",
".",
"append",
"(",
"line",
")",
"except",
":",
"# should probably warn the user?",
"pass",
"for",
"filename",
"in",
"os",
".",
"listdir",
"(",
"dirname",
")",
":",
"path",
"=",
"os",
".",
"path",
".",
"join",
"(",
"dirname",
",",
"filename",
")",
"(",
"basename",
",",
"ext",
")",
"=",
"os",
".",
"path",
".",
"splitext",
"(",
"filename",
".",
"lower",
"(",
")",
")",
"try",
":",
"if",
"(",
"os",
".",
"path",
".",
"isdir",
"(",
"path",
")",
")",
":",
"if",
"(",
"not",
"basename",
".",
"startswith",
"(",
"\".\"",
")",
")",
":",
"if",
"os",
".",
"access",
"(",
"path",
",",
"os",
".",
"R_OK",
")",
":",
"self",
".",
"add_folder",
"(",
"path",
",",
"watch",
"=",
"False",
")",
"else",
":",
"if",
"(",
"ext",
"in",
"(",
"\".xml\"",
",",
"\".robot\"",
",",
"\".txt\"",
",",
"\".py\"",
",",
"\".tsv\"",
")",
")",
":",
"if",
"os",
".",
"access",
"(",
"path",
",",
"os",
".",
"R_OK",
")",
":",
"self",
".",
"add",
"(",
"path",
")",
"except",
"Exception",
"as",
"e",
":",
"# I really need to get the logging situation figured out.",
"print",
"(",
"\"bummer:\"",
",",
"str",
"(",
"e",
")",
")",
"# FIXME:",
"# instead of passing a flag around, I should just keep track",
"# of which folders we're watching, and don't add wathers for",
"# any subfolders. That will work better in the case where",
"# the user accidentally starts up the hub giving the same",
"# folder, or a folder and it's children, on the command line...",
"if",
"watch",
":",
"# add watcher on normalized path",
"dirname",
"=",
"os",
".",
"path",
".",
"abspath",
"(",
"dirname",
")",
"event_handler",
"=",
"WatchdogHandler",
"(",
"self",
",",
"dirname",
")",
"self",
".",
"observer",
".",
"schedule",
"(",
"event_handler",
",",
"dirname",
",",
"recursive",
"=",
"True",
")"
] | Recursively add all files in a folder to the database
By "all files" I mean, "all files that are resource files
or library files". It will silently ignore files that don't
look like they belong in the database. Pity the fool who
uses non-standard suffixes.
N.B. folders with names that begin with '." will be skipped | [
"Recursively",
"add",
"all",
"files",
"in",
"a",
"folder",
"to",
"the",
"database"
] | f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c | https://github.com/boakley/robotframework-hub/blob/f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c/rfhub/kwdb.py#L158-L210 |
250,934 | boakley/robotframework-hub | rfhub/kwdb.py | KeywordTable.add_installed_libraries | def add_installed_libraries(self, extra_libs = ["Selenium2Library",
"SudsLibrary",
"RequestsLibrary"]):
"""Add any installed libraries that we can find
We do this by looking in the `libraries` folder where
robot is installed. If you have libraries installed
in a non-standard place, this won't pick them up.
"""
libdir = os.path.dirname(robot.libraries.__file__)
loaded = []
for filename in os.listdir(libdir):
if filename.endswith(".py") or filename.endswith(".pyc"):
libname, ext = os.path.splitext(filename)
if (libname.lower() not in loaded and
not self._should_ignore(libname)):
try:
self.add(libname)
loaded.append(libname.lower())
except Exception as e:
# need a better way to log this...
self.log.debug("unable to add library: " + str(e))
# I hate how I implemented this, but I don't think there's
# any way to find out which installed python packages are
# robot libraries.
for library in extra_libs:
if (library.lower() not in loaded and
not self._should_ignore(library)):
try:
self.add(library)
loaded.append(library.lower())
except Exception as e:
self.log.debug("unable to add external library %s: %s" % \
(library, str(e))) | python | def add_installed_libraries(self, extra_libs = ["Selenium2Library",
"SudsLibrary",
"RequestsLibrary"]):
libdir = os.path.dirname(robot.libraries.__file__)
loaded = []
for filename in os.listdir(libdir):
if filename.endswith(".py") or filename.endswith(".pyc"):
libname, ext = os.path.splitext(filename)
if (libname.lower() not in loaded and
not self._should_ignore(libname)):
try:
self.add(libname)
loaded.append(libname.lower())
except Exception as e:
# need a better way to log this...
self.log.debug("unable to add library: " + str(e))
# I hate how I implemented this, but I don't think there's
# any way to find out which installed python packages are
# robot libraries.
for library in extra_libs:
if (library.lower() not in loaded and
not self._should_ignore(library)):
try:
self.add(library)
loaded.append(library.lower())
except Exception as e:
self.log.debug("unable to add external library %s: %s" % \
(library, str(e))) | [
"def",
"add_installed_libraries",
"(",
"self",
",",
"extra_libs",
"=",
"[",
"\"Selenium2Library\"",
",",
"\"SudsLibrary\"",
",",
"\"RequestsLibrary\"",
"]",
")",
":",
"libdir",
"=",
"os",
".",
"path",
".",
"dirname",
"(",
"robot",
".",
"libraries",
".",
"__file__",
")",
"loaded",
"=",
"[",
"]",
"for",
"filename",
"in",
"os",
".",
"listdir",
"(",
"libdir",
")",
":",
"if",
"filename",
".",
"endswith",
"(",
"\".py\"",
")",
"or",
"filename",
".",
"endswith",
"(",
"\".pyc\"",
")",
":",
"libname",
",",
"ext",
"=",
"os",
".",
"path",
".",
"splitext",
"(",
"filename",
")",
"if",
"(",
"libname",
".",
"lower",
"(",
")",
"not",
"in",
"loaded",
"and",
"not",
"self",
".",
"_should_ignore",
"(",
"libname",
")",
")",
":",
"try",
":",
"self",
".",
"add",
"(",
"libname",
")",
"loaded",
".",
"append",
"(",
"libname",
".",
"lower",
"(",
")",
")",
"except",
"Exception",
"as",
"e",
":",
"# need a better way to log this...",
"self",
".",
"log",
".",
"debug",
"(",
"\"unable to add library: \"",
"+",
"str",
"(",
"e",
")",
")",
"# I hate how I implemented this, but I don't think there's",
"# any way to find out which installed python packages are",
"# robot libraries.",
"for",
"library",
"in",
"extra_libs",
":",
"if",
"(",
"library",
".",
"lower",
"(",
")",
"not",
"in",
"loaded",
"and",
"not",
"self",
".",
"_should_ignore",
"(",
"library",
")",
")",
":",
"try",
":",
"self",
".",
"add",
"(",
"library",
")",
"loaded",
".",
"append",
"(",
"library",
".",
"lower",
"(",
")",
")",
"except",
"Exception",
"as",
"e",
":",
"self",
".",
"log",
".",
"debug",
"(",
"\"unable to add external library %s: %s\"",
"%",
"(",
"library",
",",
"str",
"(",
"e",
")",
")",
")"
] | Add any installed libraries that we can find
We do this by looking in the `libraries` folder where
robot is installed. If you have libraries installed
in a non-standard place, this won't pick them up. | [
"Add",
"any",
"installed",
"libraries",
"that",
"we",
"can",
"find"
] | f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c | https://github.com/boakley/robotframework-hub/blob/f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c/rfhub/kwdb.py#L230-L266 |
250,935 | boakley/robotframework-hub | rfhub/kwdb.py | KeywordTable.get_collection | def get_collection(self, collection_id):
"""Get a specific collection"""
sql = """SELECT collection.collection_id, collection.type,
collection.name, collection.path,
collection.doc,
collection.version, collection.scope,
collection.namedargs,
collection.doc_format
FROM collection_table as collection
WHERE collection_id == ? OR collection.name like ?
"""
cursor = self._execute(sql, (collection_id, collection_id))
# need to handle the case where we get more than one result...
sql_result = cursor.fetchone()
return {
"collection_id": sql_result[0],
"type": sql_result[1],
"name": sql_result[2],
"path": sql_result[3],
"doc": sql_result[4],
"version": sql_result[5],
"scope": sql_result[6],
"namedargs": sql_result[7],
"doc_format": sql_result[8]
}
return sql_result | python | def get_collection(self, collection_id):
sql = """SELECT collection.collection_id, collection.type,
collection.name, collection.path,
collection.doc,
collection.version, collection.scope,
collection.namedargs,
collection.doc_format
FROM collection_table as collection
WHERE collection_id == ? OR collection.name like ?
"""
cursor = self._execute(sql, (collection_id, collection_id))
# need to handle the case where we get more than one result...
sql_result = cursor.fetchone()
return {
"collection_id": sql_result[0],
"type": sql_result[1],
"name": sql_result[2],
"path": sql_result[3],
"doc": sql_result[4],
"version": sql_result[5],
"scope": sql_result[6],
"namedargs": sql_result[7],
"doc_format": sql_result[8]
}
return sql_result | [
"def",
"get_collection",
"(",
"self",
",",
"collection_id",
")",
":",
"sql",
"=",
"\"\"\"SELECT collection.collection_id, collection.type,\n collection.name, collection.path,\n collection.doc,\n collection.version, collection.scope,\n collection.namedargs,\n collection.doc_format\n FROM collection_table as collection\n WHERE collection_id == ? OR collection.name like ?\n \"\"\"",
"cursor",
"=",
"self",
".",
"_execute",
"(",
"sql",
",",
"(",
"collection_id",
",",
"collection_id",
")",
")",
"# need to handle the case where we get more than one result...",
"sql_result",
"=",
"cursor",
".",
"fetchone",
"(",
")",
"return",
"{",
"\"collection_id\"",
":",
"sql_result",
"[",
"0",
"]",
",",
"\"type\"",
":",
"sql_result",
"[",
"1",
"]",
",",
"\"name\"",
":",
"sql_result",
"[",
"2",
"]",
",",
"\"path\"",
":",
"sql_result",
"[",
"3",
"]",
",",
"\"doc\"",
":",
"sql_result",
"[",
"4",
"]",
",",
"\"version\"",
":",
"sql_result",
"[",
"5",
"]",
",",
"\"scope\"",
":",
"sql_result",
"[",
"6",
"]",
",",
"\"namedargs\"",
":",
"sql_result",
"[",
"7",
"]",
",",
"\"doc_format\"",
":",
"sql_result",
"[",
"8",
"]",
"}",
"return",
"sql_result"
] | Get a specific collection | [
"Get",
"a",
"specific",
"collection"
] | f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c | https://github.com/boakley/robotframework-hub/blob/f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c/rfhub/kwdb.py#L268-L293 |
250,936 | boakley/robotframework-hub | rfhub/kwdb.py | KeywordTable.get_keyword | def get_keyword(self, collection_id, name):
"""Get a specific keyword from a library"""
sql = """SELECT keyword.name, keyword.args, keyword.doc
FROM keyword_table as keyword
WHERE keyword.collection_id == ?
AND keyword.name like ?
"""
cursor = self._execute(sql, (collection_id,name))
# We're going to assume no library has duplicate keywords
# While that in theory _could_ happen, it never _should_,
# and you get what you deserve if it does.
row = cursor.fetchone()
if row is not None:
return {"name": row[0],
"args": json.loads(row[1]),
"doc": row[2],
"collection_id": collection_id
}
return {} | python | def get_keyword(self, collection_id, name):
sql = """SELECT keyword.name, keyword.args, keyword.doc
FROM keyword_table as keyword
WHERE keyword.collection_id == ?
AND keyword.name like ?
"""
cursor = self._execute(sql, (collection_id,name))
# We're going to assume no library has duplicate keywords
# While that in theory _could_ happen, it never _should_,
# and you get what you deserve if it does.
row = cursor.fetchone()
if row is not None:
return {"name": row[0],
"args": json.loads(row[1]),
"doc": row[2],
"collection_id": collection_id
}
return {} | [
"def",
"get_keyword",
"(",
"self",
",",
"collection_id",
",",
"name",
")",
":",
"sql",
"=",
"\"\"\"SELECT keyword.name, keyword.args, keyword.doc\n FROM keyword_table as keyword\n WHERE keyword.collection_id == ?\n AND keyword.name like ?\n \"\"\"",
"cursor",
"=",
"self",
".",
"_execute",
"(",
"sql",
",",
"(",
"collection_id",
",",
"name",
")",
")",
"# We're going to assume no library has duplicate keywords",
"# While that in theory _could_ happen, it never _should_,",
"# and you get what you deserve if it does.",
"row",
"=",
"cursor",
".",
"fetchone",
"(",
")",
"if",
"row",
"is",
"not",
"None",
":",
"return",
"{",
"\"name\"",
":",
"row",
"[",
"0",
"]",
",",
"\"args\"",
":",
"json",
".",
"loads",
"(",
"row",
"[",
"1",
"]",
")",
",",
"\"doc\"",
":",
"row",
"[",
"2",
"]",
",",
"\"collection_id\"",
":",
"collection_id",
"}",
"return",
"{",
"}"
] | Get a specific keyword from a library | [
"Get",
"a",
"specific",
"keyword",
"from",
"a",
"library"
] | f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c | https://github.com/boakley/robotframework-hub/blob/f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c/rfhub/kwdb.py#L326-L344 |
250,937 | boakley/robotframework-hub | rfhub/kwdb.py | KeywordTable._looks_like_libdoc_file | def _looks_like_libdoc_file(self, name):
"""Return true if an xml file looks like a libdoc file"""
# inefficient since we end up reading the file twice,
# but it's fast enough for our purposes, and prevents
# us from doing a full parse of files that are obviously
# not libdoc files
if name.lower().endswith(".xml"):
with open(name, "r") as f:
# read the first few lines; if we don't see
# what looks like libdoc data, return false
data = f.read(200)
index = data.lower().find("<keywordspec ")
if index > 0:
return True
return False | python | def _looks_like_libdoc_file(self, name):
# inefficient since we end up reading the file twice,
# but it's fast enough for our purposes, and prevents
# us from doing a full parse of files that are obviously
# not libdoc files
if name.lower().endswith(".xml"):
with open(name, "r") as f:
# read the first few lines; if we don't see
# what looks like libdoc data, return false
data = f.read(200)
index = data.lower().find("<keywordspec ")
if index > 0:
return True
return False | [
"def",
"_looks_like_libdoc_file",
"(",
"self",
",",
"name",
")",
":",
"# inefficient since we end up reading the file twice,",
"# but it's fast enough for our purposes, and prevents",
"# us from doing a full parse of files that are obviously",
"# not libdoc files",
"if",
"name",
".",
"lower",
"(",
")",
".",
"endswith",
"(",
"\".xml\"",
")",
":",
"with",
"open",
"(",
"name",
",",
"\"r\"",
")",
"as",
"f",
":",
"# read the first few lines; if we don't see",
"# what looks like libdoc data, return false",
"data",
"=",
"f",
".",
"read",
"(",
"200",
")",
"index",
"=",
"data",
".",
"lower",
"(",
")",
".",
"find",
"(",
"\"<keywordspec \"",
")",
"if",
"index",
">",
"0",
":",
"return",
"True",
"return",
"False"
] | Return true if an xml file looks like a libdoc file | [
"Return",
"true",
"if",
"an",
"xml",
"file",
"looks",
"like",
"a",
"libdoc",
"file"
] | f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c | https://github.com/boakley/robotframework-hub/blob/f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c/rfhub/kwdb.py#L447-L461 |
250,938 | boakley/robotframework-hub | rfhub/kwdb.py | KeywordTable._looks_like_resource_file | def _looks_like_resource_file(self, name):
"""Return true if the file has a keyword table but not a testcase table"""
# inefficient since we end up reading the file twice,
# but it's fast enough for our purposes, and prevents
# us from doing a full parse of files that are obviously
# not robot files
if (re.search(r'__init__.(txt|robot|html|tsv)$', name)):
# These are initialize files, not resource files
return False
found_keyword_table = False
if (name.lower().endswith(".robot") or
name.lower().endswith(".txt") or
name.lower().endswith(".tsv")):
with open(name, "r") as f:
data = f.read()
for match in re.finditer(r'^\*+\s*(Test Cases?|(?:User )?Keywords?)',
data, re.MULTILINE|re.IGNORECASE):
if (re.match(r'Test Cases?', match.group(1), re.IGNORECASE)):
# if there's a test case table, it's not a keyword file
return False
if (not found_keyword_table and
re.match(r'(User )?Keywords?', match.group(1), re.IGNORECASE)):
found_keyword_table = True
return found_keyword_table | python | def _looks_like_resource_file(self, name):
# inefficient since we end up reading the file twice,
# but it's fast enough for our purposes, and prevents
# us from doing a full parse of files that are obviously
# not robot files
if (re.search(r'__init__.(txt|robot|html|tsv)$', name)):
# These are initialize files, not resource files
return False
found_keyword_table = False
if (name.lower().endswith(".robot") or
name.lower().endswith(".txt") or
name.lower().endswith(".tsv")):
with open(name, "r") as f:
data = f.read()
for match in re.finditer(r'^\*+\s*(Test Cases?|(?:User )?Keywords?)',
data, re.MULTILINE|re.IGNORECASE):
if (re.match(r'Test Cases?', match.group(1), re.IGNORECASE)):
# if there's a test case table, it's not a keyword file
return False
if (not found_keyword_table and
re.match(r'(User )?Keywords?', match.group(1), re.IGNORECASE)):
found_keyword_table = True
return found_keyword_table | [
"def",
"_looks_like_resource_file",
"(",
"self",
",",
"name",
")",
":",
"# inefficient since we end up reading the file twice,",
"# but it's fast enough for our purposes, and prevents",
"# us from doing a full parse of files that are obviously",
"# not robot files",
"if",
"(",
"re",
".",
"search",
"(",
"r'__init__.(txt|robot|html|tsv)$'",
",",
"name",
")",
")",
":",
"# These are initialize files, not resource files",
"return",
"False",
"found_keyword_table",
"=",
"False",
"if",
"(",
"name",
".",
"lower",
"(",
")",
".",
"endswith",
"(",
"\".robot\"",
")",
"or",
"name",
".",
"lower",
"(",
")",
".",
"endswith",
"(",
"\".txt\"",
")",
"or",
"name",
".",
"lower",
"(",
")",
".",
"endswith",
"(",
"\".tsv\"",
")",
")",
":",
"with",
"open",
"(",
"name",
",",
"\"r\"",
")",
"as",
"f",
":",
"data",
"=",
"f",
".",
"read",
"(",
")",
"for",
"match",
"in",
"re",
".",
"finditer",
"(",
"r'^\\*+\\s*(Test Cases?|(?:User )?Keywords?)'",
",",
"data",
",",
"re",
".",
"MULTILINE",
"|",
"re",
".",
"IGNORECASE",
")",
":",
"if",
"(",
"re",
".",
"match",
"(",
"r'Test Cases?'",
",",
"match",
".",
"group",
"(",
"1",
")",
",",
"re",
".",
"IGNORECASE",
")",
")",
":",
"# if there's a test case table, it's not a keyword file",
"return",
"False",
"if",
"(",
"not",
"found_keyword_table",
"and",
"re",
".",
"match",
"(",
"r'(User )?Keywords?'",
",",
"match",
".",
"group",
"(",
"1",
")",
",",
"re",
".",
"IGNORECASE",
")",
")",
":",
"found_keyword_table",
"=",
"True",
"return",
"found_keyword_table"
] | Return true if the file has a keyword table but not a testcase table | [
"Return",
"true",
"if",
"the",
"file",
"has",
"a",
"keyword",
"table",
"but",
"not",
"a",
"testcase",
"table"
] | f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c | https://github.com/boakley/robotframework-hub/blob/f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c/rfhub/kwdb.py#L463-L490 |
250,939 | boakley/robotframework-hub | rfhub/kwdb.py | KeywordTable._should_ignore | def _should_ignore(self, name):
"""Return True if a given library name should be ignored
This is necessary because not all files we find in the library
folder are libraries. I wish there was a public robot API
for "give me a list of installed libraries"...
"""
_name = name.lower()
return (_name.startswith("deprecated") or
_name.startswith("_") or
_name in ("remote", "reserved",
"dialogs_py", "dialogs_ipy", "dialogs_jy")) | python | def _should_ignore(self, name):
_name = name.lower()
return (_name.startswith("deprecated") or
_name.startswith("_") or
_name in ("remote", "reserved",
"dialogs_py", "dialogs_ipy", "dialogs_jy")) | [
"def",
"_should_ignore",
"(",
"self",
",",
"name",
")",
":",
"_name",
"=",
"name",
".",
"lower",
"(",
")",
"return",
"(",
"_name",
".",
"startswith",
"(",
"\"deprecated\"",
")",
"or",
"_name",
".",
"startswith",
"(",
"\"_\"",
")",
"or",
"_name",
"in",
"(",
"\"remote\"",
",",
"\"reserved\"",
",",
"\"dialogs_py\"",
",",
"\"dialogs_ipy\"",
",",
"\"dialogs_jy\"",
")",
")"
] | Return True if a given library name should be ignored
This is necessary because not all files we find in the library
folder are libraries. I wish there was a public robot API
for "give me a list of installed libraries"... | [
"Return",
"True",
"if",
"a",
"given",
"library",
"name",
"should",
"be",
"ignored"
] | f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c | https://github.com/boakley/robotframework-hub/blob/f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c/rfhub/kwdb.py#L492-L503 |
250,940 | boakley/robotframework-hub | rfhub/kwdb.py | KeywordTable._execute | def _execute(self, *args):
"""Execute an SQL query
This exists because I think it's tedious to get a cursor and
then use a cursor.
"""
cursor = self.db.cursor()
cursor.execute(*args)
return cursor | python | def _execute(self, *args):
cursor = self.db.cursor()
cursor.execute(*args)
return cursor | [
"def",
"_execute",
"(",
"self",
",",
"*",
"args",
")",
":",
"cursor",
"=",
"self",
".",
"db",
".",
"cursor",
"(",
")",
"cursor",
".",
"execute",
"(",
"*",
"args",
")",
"return",
"cursor"
] | Execute an SQL query
This exists because I think it's tedious to get a cursor and
then use a cursor. | [
"Execute",
"an",
"SQL",
"query"
] | f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c | https://github.com/boakley/robotframework-hub/blob/f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c/rfhub/kwdb.py#L505-L513 |
250,941 | boakley/robotframework-hub | rfhub/kwdb.py | KeywordTable._glob_to_sql | def _glob_to_sql(self, string):
"""Convert glob-like wildcards to SQL wildcards
* becomes %
? becomes _
% becomes \%
\\ remains \\
\* remains \*
\? remains \?
This also adds a leading and trailing %, unless the pattern begins with
^ or ends with $
"""
# What's with the chr(1) and chr(2) nonsense? It's a trick to
# hide \* and \? from the * and ? substitutions. This trick
# depends on the substitutiones being done in order. chr(1)
# and chr(2) were picked because I know those characters
# almost certainly won't be in the input string
table = ((r'\\', chr(1)), (r'\*', chr(2)), (r'\?', chr(3)),
(r'%', r'\%'), (r'?', '_'), (r'*', '%'),
(chr(1), r'\\'), (chr(2), r'\*'), (chr(3), r'\?'))
for (a, b) in table:
string = string.replace(a,b)
string = string[1:] if string.startswith("^") else "%" + string
string = string[:-1] if string.endswith("$") else string + "%"
return string | python | def _glob_to_sql(self, string):
# What's with the chr(1) and chr(2) nonsense? It's a trick to
# hide \* and \? from the * and ? substitutions. This trick
# depends on the substitutiones being done in order. chr(1)
# and chr(2) were picked because I know those characters
# almost certainly won't be in the input string
table = ((r'\\', chr(1)), (r'\*', chr(2)), (r'\?', chr(3)),
(r'%', r'\%'), (r'?', '_'), (r'*', '%'),
(chr(1), r'\\'), (chr(2), r'\*'), (chr(3), r'\?'))
for (a, b) in table:
string = string.replace(a,b)
string = string[1:] if string.startswith("^") else "%" + string
string = string[:-1] if string.endswith("$") else string + "%"
return string | [
"def",
"_glob_to_sql",
"(",
"self",
",",
"string",
")",
":",
"# What's with the chr(1) and chr(2) nonsense? It's a trick to",
"# hide \\* and \\? from the * and ? substitutions. This trick",
"# depends on the substitutiones being done in order. chr(1)",
"# and chr(2) were picked because I know those characters",
"# almost certainly won't be in the input string",
"table",
"=",
"(",
"(",
"r'\\\\'",
",",
"chr",
"(",
"1",
")",
")",
",",
"(",
"r'\\*'",
",",
"chr",
"(",
"2",
")",
")",
",",
"(",
"r'\\?'",
",",
"chr",
"(",
"3",
")",
")",
",",
"(",
"r'%'",
",",
"r'\\%'",
")",
",",
"(",
"r'?'",
",",
"'_'",
")",
",",
"(",
"r'*'",
",",
"'%'",
")",
",",
"(",
"chr",
"(",
"1",
")",
",",
"r'\\\\'",
")",
",",
"(",
"chr",
"(",
"2",
")",
",",
"r'\\*'",
")",
",",
"(",
"chr",
"(",
"3",
")",
",",
"r'\\?'",
")",
")",
"for",
"(",
"a",
",",
"b",
")",
"in",
"table",
":",
"string",
"=",
"string",
".",
"replace",
"(",
"a",
",",
"b",
")",
"string",
"=",
"string",
"[",
"1",
":",
"]",
"if",
"string",
".",
"startswith",
"(",
"\"^\"",
")",
"else",
"\"%\"",
"+",
"string",
"string",
"=",
"string",
"[",
":",
"-",
"1",
"]",
"if",
"string",
".",
"endswith",
"(",
"\"$\"",
")",
"else",
"string",
"+",
"\"%\"",
"return",
"string"
] | Convert glob-like wildcards to SQL wildcards
* becomes %
? becomes _
% becomes \%
\\ remains \\
\* remains \*
\? remains \?
This also adds a leading and trailing %, unless the pattern begins with
^ or ends with $ | [
"Convert",
"glob",
"-",
"like",
"wildcards",
"to",
"SQL",
"wildcards"
] | f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c | https://github.com/boakley/robotframework-hub/blob/f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c/rfhub/kwdb.py#L565-L594 |
250,942 | boakley/robotframework-hub | rfhub/blueprints/doc/__init__.py | doc | def doc():
"""Show a list of libraries, along with the nav panel on the left"""
kwdb = current_app.kwdb
libraries = get_collections(kwdb, libtype="library")
resource_files = get_collections(kwdb, libtype="resource")
hierarchy = get_navpanel_data(kwdb)
return flask.render_template("home.html",
data={"libraries": libraries,
"version": __version__,
"libdoc": None,
"hierarchy": hierarchy,
"resource_files": resource_files
}) | python | def doc():
kwdb = current_app.kwdb
libraries = get_collections(kwdb, libtype="library")
resource_files = get_collections(kwdb, libtype="resource")
hierarchy = get_navpanel_data(kwdb)
return flask.render_template("home.html",
data={"libraries": libraries,
"version": __version__,
"libdoc": None,
"hierarchy": hierarchy,
"resource_files": resource_files
}) | [
"def",
"doc",
"(",
")",
":",
"kwdb",
"=",
"current_app",
".",
"kwdb",
"libraries",
"=",
"get_collections",
"(",
"kwdb",
",",
"libtype",
"=",
"\"library\"",
")",
"resource_files",
"=",
"get_collections",
"(",
"kwdb",
",",
"libtype",
"=",
"\"resource\"",
")",
"hierarchy",
"=",
"get_navpanel_data",
"(",
"kwdb",
")",
"return",
"flask",
".",
"render_template",
"(",
"\"home.html\"",
",",
"data",
"=",
"{",
"\"libraries\"",
":",
"libraries",
",",
"\"version\"",
":",
"__version__",
",",
"\"libdoc\"",
":",
"None",
",",
"\"hierarchy\"",
":",
"hierarchy",
",",
"\"resource_files\"",
":",
"resource_files",
"}",
")"
] | Show a list of libraries, along with the nav panel on the left | [
"Show",
"a",
"list",
"of",
"libraries",
"along",
"with",
"the",
"nav",
"panel",
"on",
"the",
"left"
] | f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c | https://github.com/boakley/robotframework-hub/blob/f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c/rfhub/blueprints/doc/__init__.py#L14-L28 |
250,943 | boakley/robotframework-hub | rfhub/blueprints/doc/__init__.py | index | def index():
"""Show a list of available libraries, and resource files"""
kwdb = current_app.kwdb
libraries = get_collections(kwdb, libtype="library")
resource_files = get_collections(kwdb, libtype="resource")
return flask.render_template("libraryNames.html",
data={"libraries": libraries,
"version": __version__,
"resource_files": resource_files
}) | python | def index():
kwdb = current_app.kwdb
libraries = get_collections(kwdb, libtype="library")
resource_files = get_collections(kwdb, libtype="resource")
return flask.render_template("libraryNames.html",
data={"libraries": libraries,
"version": __version__,
"resource_files": resource_files
}) | [
"def",
"index",
"(",
")",
":",
"kwdb",
"=",
"current_app",
".",
"kwdb",
"libraries",
"=",
"get_collections",
"(",
"kwdb",
",",
"libtype",
"=",
"\"library\"",
")",
"resource_files",
"=",
"get_collections",
"(",
"kwdb",
",",
"libtype",
"=",
"\"resource\"",
")",
"return",
"flask",
".",
"render_template",
"(",
"\"libraryNames.html\"",
",",
"data",
"=",
"{",
"\"libraries\"",
":",
"libraries",
",",
"\"version\"",
":",
"__version__",
",",
"\"resource_files\"",
":",
"resource_files",
"}",
")"
] | Show a list of available libraries, and resource files | [
"Show",
"a",
"list",
"of",
"available",
"libraries",
"and",
"resource",
"files"
] | f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c | https://github.com/boakley/robotframework-hub/blob/f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c/rfhub/blueprints/doc/__init__.py#L32-L43 |
250,944 | boakley/robotframework-hub | rfhub/blueprints/doc/__init__.py | search | def search():
"""Show all keywords that match a pattern"""
pattern = flask.request.args.get('pattern', "*").strip().lower()
# if the pattern contains "in:<collection>" (eg: in:builtin),
# filter results to only that (or those) collections
# This was kind-of hacked together, but seems to work well enough
collections = [c["name"].lower() for c in current_app.kwdb.get_collections()]
words = []
filters = []
if pattern.startswith("name:"):
pattern = pattern[5:].strip()
mode = "name"
else:
mode="both"
for word in pattern.split(" "):
if word.lower().startswith("in:"):
filters.extend([name for name in collections if name.startswith(word[3:])])
else:
words.append(word)
pattern = " ".join(words)
keywords = []
for keyword in current_app.kwdb.search(pattern, mode):
kw = list(keyword)
collection_id = kw[0]
collection_name = kw[1].lower()
if len(filters) == 0 or collection_name in filters:
url = flask.url_for(".doc_for_library", collection_id=kw[0], keyword=kw[2])
row_id = "row-%s.%s" % (keyword[1].lower(), keyword[2].lower().replace(" ","-"))
keywords.append({"collection_id": keyword[0],
"collection_name": keyword[1],
"name": keyword[2],
"synopsis": keyword[3],
"version": __version__,
"url": url,
"row_id": row_id
})
keywords.sort(key=lambda kw: kw["name"])
return flask.render_template("search.html",
data={"keywords": keywords,
"version": __version__,
"pattern": pattern
}) | python | def search():
pattern = flask.request.args.get('pattern', "*").strip().lower()
# if the pattern contains "in:<collection>" (eg: in:builtin),
# filter results to only that (or those) collections
# This was kind-of hacked together, but seems to work well enough
collections = [c["name"].lower() for c in current_app.kwdb.get_collections()]
words = []
filters = []
if pattern.startswith("name:"):
pattern = pattern[5:].strip()
mode = "name"
else:
mode="both"
for word in pattern.split(" "):
if word.lower().startswith("in:"):
filters.extend([name for name in collections if name.startswith(word[3:])])
else:
words.append(word)
pattern = " ".join(words)
keywords = []
for keyword in current_app.kwdb.search(pattern, mode):
kw = list(keyword)
collection_id = kw[0]
collection_name = kw[1].lower()
if len(filters) == 0 or collection_name in filters:
url = flask.url_for(".doc_for_library", collection_id=kw[0], keyword=kw[2])
row_id = "row-%s.%s" % (keyword[1].lower(), keyword[2].lower().replace(" ","-"))
keywords.append({"collection_id": keyword[0],
"collection_name": keyword[1],
"name": keyword[2],
"synopsis": keyword[3],
"version": __version__,
"url": url,
"row_id": row_id
})
keywords.sort(key=lambda kw: kw["name"])
return flask.render_template("search.html",
data={"keywords": keywords,
"version": __version__,
"pattern": pattern
}) | [
"def",
"search",
"(",
")",
":",
"pattern",
"=",
"flask",
".",
"request",
".",
"args",
".",
"get",
"(",
"'pattern'",
",",
"\"*\"",
")",
".",
"strip",
"(",
")",
".",
"lower",
"(",
")",
"# if the pattern contains \"in:<collection>\" (eg: in:builtin),",
"# filter results to only that (or those) collections",
"# This was kind-of hacked together, but seems to work well enough",
"collections",
"=",
"[",
"c",
"[",
"\"name\"",
"]",
".",
"lower",
"(",
")",
"for",
"c",
"in",
"current_app",
".",
"kwdb",
".",
"get_collections",
"(",
")",
"]",
"words",
"=",
"[",
"]",
"filters",
"=",
"[",
"]",
"if",
"pattern",
".",
"startswith",
"(",
"\"name:\"",
")",
":",
"pattern",
"=",
"pattern",
"[",
"5",
":",
"]",
".",
"strip",
"(",
")",
"mode",
"=",
"\"name\"",
"else",
":",
"mode",
"=",
"\"both\"",
"for",
"word",
"in",
"pattern",
".",
"split",
"(",
"\" \"",
")",
":",
"if",
"word",
".",
"lower",
"(",
")",
".",
"startswith",
"(",
"\"in:\"",
")",
":",
"filters",
".",
"extend",
"(",
"[",
"name",
"for",
"name",
"in",
"collections",
"if",
"name",
".",
"startswith",
"(",
"word",
"[",
"3",
":",
"]",
")",
"]",
")",
"else",
":",
"words",
".",
"append",
"(",
"word",
")",
"pattern",
"=",
"\" \"",
".",
"join",
"(",
"words",
")",
"keywords",
"=",
"[",
"]",
"for",
"keyword",
"in",
"current_app",
".",
"kwdb",
".",
"search",
"(",
"pattern",
",",
"mode",
")",
":",
"kw",
"=",
"list",
"(",
"keyword",
")",
"collection_id",
"=",
"kw",
"[",
"0",
"]",
"collection_name",
"=",
"kw",
"[",
"1",
"]",
".",
"lower",
"(",
")",
"if",
"len",
"(",
"filters",
")",
"==",
"0",
"or",
"collection_name",
"in",
"filters",
":",
"url",
"=",
"flask",
".",
"url_for",
"(",
"\".doc_for_library\"",
",",
"collection_id",
"=",
"kw",
"[",
"0",
"]",
",",
"keyword",
"=",
"kw",
"[",
"2",
"]",
")",
"row_id",
"=",
"\"row-%s.%s\"",
"%",
"(",
"keyword",
"[",
"1",
"]",
".",
"lower",
"(",
")",
",",
"keyword",
"[",
"2",
"]",
".",
"lower",
"(",
")",
".",
"replace",
"(",
"\" \"",
",",
"\"-\"",
")",
")",
"keywords",
".",
"append",
"(",
"{",
"\"collection_id\"",
":",
"keyword",
"[",
"0",
"]",
",",
"\"collection_name\"",
":",
"keyword",
"[",
"1",
"]",
",",
"\"name\"",
":",
"keyword",
"[",
"2",
"]",
",",
"\"synopsis\"",
":",
"keyword",
"[",
"3",
"]",
",",
"\"version\"",
":",
"__version__",
",",
"\"url\"",
":",
"url",
",",
"\"row_id\"",
":",
"row_id",
"}",
")",
"keywords",
".",
"sort",
"(",
"key",
"=",
"lambda",
"kw",
":",
"kw",
"[",
"\"name\"",
"]",
")",
"return",
"flask",
".",
"render_template",
"(",
"\"search.html\"",
",",
"data",
"=",
"{",
"\"keywords\"",
":",
"keywords",
",",
"\"version\"",
":",
"__version__",
",",
"\"pattern\"",
":",
"pattern",
"}",
")"
] | Show all keywords that match a pattern | [
"Show",
"all",
"keywords",
"that",
"match",
"a",
"pattern"
] | f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c | https://github.com/boakley/robotframework-hub/blob/f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c/rfhub/blueprints/doc/__init__.py#L47-L92 |
250,945 | boakley/robotframework-hub | rfhub/blueprints/doc/__init__.py | get_collections | def get_collections(kwdb, libtype="*"):
"""Get list of collections from kwdb, then add urls necessary for hyperlinks"""
collections = kwdb.get_collections(libtype=libtype)
for result in collections:
url = flask.url_for(".doc_for_library", collection_id=result["collection_id"])
result["url"] = url
return collections | python | def get_collections(kwdb, libtype="*"):
collections = kwdb.get_collections(libtype=libtype)
for result in collections:
url = flask.url_for(".doc_for_library", collection_id=result["collection_id"])
result["url"] = url
return collections | [
"def",
"get_collections",
"(",
"kwdb",
",",
"libtype",
"=",
"\"*\"",
")",
":",
"collections",
"=",
"kwdb",
".",
"get_collections",
"(",
"libtype",
"=",
"libtype",
")",
"for",
"result",
"in",
"collections",
":",
"url",
"=",
"flask",
".",
"url_for",
"(",
"\".doc_for_library\"",
",",
"collection_id",
"=",
"result",
"[",
"\"collection_id\"",
"]",
")",
"result",
"[",
"\"url\"",
"]",
"=",
"url",
"return",
"collections"
] | Get list of collections from kwdb, then add urls necessary for hyperlinks | [
"Get",
"list",
"of",
"collections",
"from",
"kwdb",
"then",
"add",
"urls",
"necessary",
"for",
"hyperlinks"
] | f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c | https://github.com/boakley/robotframework-hub/blob/f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c/rfhub/blueprints/doc/__init__.py#L129-L136 |
250,946 | boakley/robotframework-hub | rfhub/blueprints/doc/__init__.py | get_navpanel_data | def get_navpanel_data(kwdb):
"""Get navpanel data from kwdb, and add urls necessary for hyperlinks"""
data = kwdb.get_keyword_hierarchy()
for library in data:
library["url"] = flask.url_for(".doc_for_library", collection_id=library["collection_id"])
for keyword in library["keywords"]:
url = flask.url_for(".doc_for_library",
collection_id=library["collection_id"],
keyword=keyword["name"])
keyword["url"] = url
return data | python | def get_navpanel_data(kwdb):
data = kwdb.get_keyword_hierarchy()
for library in data:
library["url"] = flask.url_for(".doc_for_library", collection_id=library["collection_id"])
for keyword in library["keywords"]:
url = flask.url_for(".doc_for_library",
collection_id=library["collection_id"],
keyword=keyword["name"])
keyword["url"] = url
return data | [
"def",
"get_navpanel_data",
"(",
"kwdb",
")",
":",
"data",
"=",
"kwdb",
".",
"get_keyword_hierarchy",
"(",
")",
"for",
"library",
"in",
"data",
":",
"library",
"[",
"\"url\"",
"]",
"=",
"flask",
".",
"url_for",
"(",
"\".doc_for_library\"",
",",
"collection_id",
"=",
"library",
"[",
"\"collection_id\"",
"]",
")",
"for",
"keyword",
"in",
"library",
"[",
"\"keywords\"",
"]",
":",
"url",
"=",
"flask",
".",
"url_for",
"(",
"\".doc_for_library\"",
",",
"collection_id",
"=",
"library",
"[",
"\"collection_id\"",
"]",
",",
"keyword",
"=",
"keyword",
"[",
"\"name\"",
"]",
")",
"keyword",
"[",
"\"url\"",
"]",
"=",
"url",
"return",
"data"
] | Get navpanel data from kwdb, and add urls necessary for hyperlinks | [
"Get",
"navpanel",
"data",
"from",
"kwdb",
"and",
"add",
"urls",
"necessary",
"for",
"hyperlinks"
] | f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c | https://github.com/boakley/robotframework-hub/blob/f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c/rfhub/blueprints/doc/__init__.py#L138-L149 |
250,947 | boakley/robotframework-hub | rfhub/blueprints/doc/__init__.py | doc_to_html | def doc_to_html(doc, doc_format="ROBOT"):
"""Convert documentation to HTML"""
from robot.libdocpkg.htmlwriter import DocToHtml
return DocToHtml(doc_format)(doc) | python | def doc_to_html(doc, doc_format="ROBOT"):
from robot.libdocpkg.htmlwriter import DocToHtml
return DocToHtml(doc_format)(doc) | [
"def",
"doc_to_html",
"(",
"doc",
",",
"doc_format",
"=",
"\"ROBOT\"",
")",
":",
"from",
"robot",
".",
"libdocpkg",
".",
"htmlwriter",
"import",
"DocToHtml",
"return",
"DocToHtml",
"(",
"doc_format",
")",
"(",
"doc",
")"
] | Convert documentation to HTML | [
"Convert",
"documentation",
"to",
"HTML"
] | f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c | https://github.com/boakley/robotframework-hub/blob/f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c/rfhub/blueprints/doc/__init__.py#L152-L155 |
250,948 | boakley/robotframework-hub | rfhub/app.py | RobotHub.start | def start(self):
"""Start the app"""
if self.args.debug:
self.app.run(port=self.args.port, debug=self.args.debug, host=self.args.interface)
else:
root = "http://%s:%s" % (self.args.interface, self.args.port)
print("tornado web server running on " + root)
self.shutdown_requested = False
http_server = HTTPServer(WSGIContainer(self.app))
http_server.listen(port=self.args.port, address=self.args.interface)
signal.signal(signal.SIGINT, self.signal_handler)
tornado.ioloop.PeriodicCallback(self.check_shutdown_flag, 500).start()
tornado.ioloop.IOLoop.instance().start() | python | def start(self):
if self.args.debug:
self.app.run(port=self.args.port, debug=self.args.debug, host=self.args.interface)
else:
root = "http://%s:%s" % (self.args.interface, self.args.port)
print("tornado web server running on " + root)
self.shutdown_requested = False
http_server = HTTPServer(WSGIContainer(self.app))
http_server.listen(port=self.args.port, address=self.args.interface)
signal.signal(signal.SIGINT, self.signal_handler)
tornado.ioloop.PeriodicCallback(self.check_shutdown_flag, 500).start()
tornado.ioloop.IOLoop.instance().start() | [
"def",
"start",
"(",
"self",
")",
":",
"if",
"self",
".",
"args",
".",
"debug",
":",
"self",
".",
"app",
".",
"run",
"(",
"port",
"=",
"self",
".",
"args",
".",
"port",
",",
"debug",
"=",
"self",
".",
"args",
".",
"debug",
",",
"host",
"=",
"self",
".",
"args",
".",
"interface",
")",
"else",
":",
"root",
"=",
"\"http://%s:%s\"",
"%",
"(",
"self",
".",
"args",
".",
"interface",
",",
"self",
".",
"args",
".",
"port",
")",
"print",
"(",
"\"tornado web server running on \"",
"+",
"root",
")",
"self",
".",
"shutdown_requested",
"=",
"False",
"http_server",
"=",
"HTTPServer",
"(",
"WSGIContainer",
"(",
"self",
".",
"app",
")",
")",
"http_server",
".",
"listen",
"(",
"port",
"=",
"self",
".",
"args",
".",
"port",
",",
"address",
"=",
"self",
".",
"args",
".",
"interface",
")",
"signal",
".",
"signal",
"(",
"signal",
".",
"SIGINT",
",",
"self",
".",
"signal_handler",
")",
"tornado",
".",
"ioloop",
".",
"PeriodicCallback",
"(",
"self",
".",
"check_shutdown_flag",
",",
"500",
")",
".",
"start",
"(",
")",
"tornado",
".",
"ioloop",
".",
"IOLoop",
".",
"instance",
"(",
")",
".",
"start",
"(",
")"
] | Start the app | [
"Start",
"the",
"app"
] | f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c | https://github.com/boakley/robotframework-hub/blob/f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c/rfhub/app.py#L50-L63 |
250,949 | boakley/robotframework-hub | rfhub/app.py | RobotHub.check_shutdown_flag | def check_shutdown_flag(self):
"""Shutdown the server if the flag has been set"""
if self.shutdown_requested:
tornado.ioloop.IOLoop.instance().stop()
print("web server stopped.") | python | def check_shutdown_flag(self):
if self.shutdown_requested:
tornado.ioloop.IOLoop.instance().stop()
print("web server stopped.") | [
"def",
"check_shutdown_flag",
"(",
"self",
")",
":",
"if",
"self",
".",
"shutdown_requested",
":",
"tornado",
".",
"ioloop",
".",
"IOLoop",
".",
"instance",
"(",
")",
".",
"stop",
"(",
")",
"print",
"(",
"\"web server stopped.\"",
")"
] | Shutdown the server if the flag has been set | [
"Shutdown",
"the",
"server",
"if",
"the",
"flag",
"has",
"been",
"set"
] | f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c | https://github.com/boakley/robotframework-hub/blob/f3dc7562fe6218a7b8d7aac7b9ef234e1a573f7c/rfhub/app.py#L69-L73 |
250,950 | jazzband/python-geojson | geojson/utils.py | coords | def coords(obj):
"""
Yields the coordinates from a Feature or Geometry.
:param obj: A geometry or feature to extract the coordinates from.
:type obj: Feature, Geometry
:return: A generator with coordinate tuples from the geometry or feature.
:rtype: generator
"""
# Handle recursive case first
if 'features' in obj:
for f in obj['features']:
# For Python 2 compatibility
# See https://www.reddit.com/r/learnpython/comments/4rc15s/yield_from_and_python_27/ # noqa: E501
for c in coords(f):
yield c
else:
if isinstance(obj, (tuple, list)):
coordinates = obj
elif 'geometry' in obj:
coordinates = obj['geometry']['coordinates']
else:
coordinates = obj.get('coordinates', obj)
for e in coordinates:
if isinstance(e, (float, int)):
yield tuple(coordinates)
break
for f in coords(e):
yield f | python | def coords(obj):
# Handle recursive case first
if 'features' in obj:
for f in obj['features']:
# For Python 2 compatibility
# See https://www.reddit.com/r/learnpython/comments/4rc15s/yield_from_and_python_27/ # noqa: E501
for c in coords(f):
yield c
else:
if isinstance(obj, (tuple, list)):
coordinates = obj
elif 'geometry' in obj:
coordinates = obj['geometry']['coordinates']
else:
coordinates = obj.get('coordinates', obj)
for e in coordinates:
if isinstance(e, (float, int)):
yield tuple(coordinates)
break
for f in coords(e):
yield f | [
"def",
"coords",
"(",
"obj",
")",
":",
"# Handle recursive case first",
"if",
"'features'",
"in",
"obj",
":",
"for",
"f",
"in",
"obj",
"[",
"'features'",
"]",
":",
"# For Python 2 compatibility",
"# See https://www.reddit.com/r/learnpython/comments/4rc15s/yield_from_and_python_27/ # noqa: E501",
"for",
"c",
"in",
"coords",
"(",
"f",
")",
":",
"yield",
"c",
"else",
":",
"if",
"isinstance",
"(",
"obj",
",",
"(",
"tuple",
",",
"list",
")",
")",
":",
"coordinates",
"=",
"obj",
"elif",
"'geometry'",
"in",
"obj",
":",
"coordinates",
"=",
"obj",
"[",
"'geometry'",
"]",
"[",
"'coordinates'",
"]",
"else",
":",
"coordinates",
"=",
"obj",
".",
"get",
"(",
"'coordinates'",
",",
"obj",
")",
"for",
"e",
"in",
"coordinates",
":",
"if",
"isinstance",
"(",
"e",
",",
"(",
"float",
",",
"int",
")",
")",
":",
"yield",
"tuple",
"(",
"coordinates",
")",
"break",
"for",
"f",
"in",
"coords",
"(",
"e",
")",
":",
"yield",
"f"
] | Yields the coordinates from a Feature or Geometry.
:param obj: A geometry or feature to extract the coordinates from.
:type obj: Feature, Geometry
:return: A generator with coordinate tuples from the geometry or feature.
:rtype: generator | [
"Yields",
"the",
"coordinates",
"from",
"a",
"Feature",
"or",
"Geometry",
"."
] | 14abb31ba73a9f3cdb81d81c56656ea736f3c865 | https://github.com/jazzband/python-geojson/blob/14abb31ba73a9f3cdb81d81c56656ea736f3c865/geojson/utils.py#L4-L32 |
250,951 | jazzband/python-geojson | geojson/utils.py | map_tuples | def map_tuples(func, obj):
"""
Returns the mapped coordinates from a Geometry after applying the provided
function to each coordinate.
:param func: Function to apply to tuples
:type func: function
:param obj: A geometry or feature to extract the coordinates from.
:type obj: Point, LineString, MultiPoint, MultiLineString, Polygon,
MultiPolygon
:return: The result of applying the function to each dimension in the
array.
:rtype: list
:raises ValueError: if the provided object is not GeoJSON.
"""
if obj['type'] == 'Point':
coordinates = tuple(func(obj['coordinates']))
elif obj['type'] in ['LineString', 'MultiPoint']:
coordinates = [tuple(func(c)) for c in obj['coordinates']]
elif obj['type'] in ['MultiLineString', 'Polygon']:
coordinates = [[
tuple(func(c)) for c in curve]
for curve in obj['coordinates']]
elif obj['type'] == 'MultiPolygon':
coordinates = [[[
tuple(func(c)) for c in curve]
for curve in part]
for part in obj['coordinates']]
elif obj['type'] in ['Feature', 'FeatureCollection', 'GeometryCollection']:
return map_geometries(lambda g: map_tuples(func, g), obj)
else:
raise ValueError("Invalid geometry object %s" % repr(obj))
return {'type': obj['type'], 'coordinates': coordinates} | python | def map_tuples(func, obj):
if obj['type'] == 'Point':
coordinates = tuple(func(obj['coordinates']))
elif obj['type'] in ['LineString', 'MultiPoint']:
coordinates = [tuple(func(c)) for c in obj['coordinates']]
elif obj['type'] in ['MultiLineString', 'Polygon']:
coordinates = [[
tuple(func(c)) for c in curve]
for curve in obj['coordinates']]
elif obj['type'] == 'MultiPolygon':
coordinates = [[[
tuple(func(c)) for c in curve]
for curve in part]
for part in obj['coordinates']]
elif obj['type'] in ['Feature', 'FeatureCollection', 'GeometryCollection']:
return map_geometries(lambda g: map_tuples(func, g), obj)
else:
raise ValueError("Invalid geometry object %s" % repr(obj))
return {'type': obj['type'], 'coordinates': coordinates} | [
"def",
"map_tuples",
"(",
"func",
",",
"obj",
")",
":",
"if",
"obj",
"[",
"'type'",
"]",
"==",
"'Point'",
":",
"coordinates",
"=",
"tuple",
"(",
"func",
"(",
"obj",
"[",
"'coordinates'",
"]",
")",
")",
"elif",
"obj",
"[",
"'type'",
"]",
"in",
"[",
"'LineString'",
",",
"'MultiPoint'",
"]",
":",
"coordinates",
"=",
"[",
"tuple",
"(",
"func",
"(",
"c",
")",
")",
"for",
"c",
"in",
"obj",
"[",
"'coordinates'",
"]",
"]",
"elif",
"obj",
"[",
"'type'",
"]",
"in",
"[",
"'MultiLineString'",
",",
"'Polygon'",
"]",
":",
"coordinates",
"=",
"[",
"[",
"tuple",
"(",
"func",
"(",
"c",
")",
")",
"for",
"c",
"in",
"curve",
"]",
"for",
"curve",
"in",
"obj",
"[",
"'coordinates'",
"]",
"]",
"elif",
"obj",
"[",
"'type'",
"]",
"==",
"'MultiPolygon'",
":",
"coordinates",
"=",
"[",
"[",
"[",
"tuple",
"(",
"func",
"(",
"c",
")",
")",
"for",
"c",
"in",
"curve",
"]",
"for",
"curve",
"in",
"part",
"]",
"for",
"part",
"in",
"obj",
"[",
"'coordinates'",
"]",
"]",
"elif",
"obj",
"[",
"'type'",
"]",
"in",
"[",
"'Feature'",
",",
"'FeatureCollection'",
",",
"'GeometryCollection'",
"]",
":",
"return",
"map_geometries",
"(",
"lambda",
"g",
":",
"map_tuples",
"(",
"func",
",",
"g",
")",
",",
"obj",
")",
"else",
":",
"raise",
"ValueError",
"(",
"\"Invalid geometry object %s\"",
"%",
"repr",
"(",
"obj",
")",
")",
"return",
"{",
"'type'",
":",
"obj",
"[",
"'type'",
"]",
",",
"'coordinates'",
":",
"coordinates",
"}"
] | Returns the mapped coordinates from a Geometry after applying the provided
function to each coordinate.
:param func: Function to apply to tuples
:type func: function
:param obj: A geometry or feature to extract the coordinates from.
:type obj: Point, LineString, MultiPoint, MultiLineString, Polygon,
MultiPolygon
:return: The result of applying the function to each dimension in the
array.
:rtype: list
:raises ValueError: if the provided object is not GeoJSON. | [
"Returns",
"the",
"mapped",
"coordinates",
"from",
"a",
"Geometry",
"after",
"applying",
"the",
"provided",
"function",
"to",
"each",
"coordinate",
"."
] | 14abb31ba73a9f3cdb81d81c56656ea736f3c865 | https://github.com/jazzband/python-geojson/blob/14abb31ba73a9f3cdb81d81c56656ea736f3c865/geojson/utils.py#L58-L91 |
250,952 | jazzband/python-geojson | geojson/utils.py | map_geometries | def map_geometries(func, obj):
"""
Returns the result of passing every geometry in the given geojson object
through func.
:param func: Function to apply to tuples
:type func: function
:param obj: A geometry or feature to extract the coordinates from.
:type obj: GeoJSON
:return: The result of applying the function to each geometry
:rtype: list
:raises ValueError: if the provided object is not geojson.
"""
simple_types = [
'Point',
'LineString',
'MultiPoint',
'MultiLineString',
'Polygon',
'MultiPolygon',
]
if obj['type'] in simple_types:
return func(obj)
elif obj['type'] == 'GeometryCollection':
geoms = [func(geom) if geom else None for geom in obj['geometries']]
return {'type': obj['type'], 'geometries': geoms}
elif obj['type'] == 'Feature':
geom = func(obj['geometry']) if obj['geometry'] else None
return {
'type': obj['type'],
'geometry': geom,
'properties': obj['properties'],
}
elif obj['type'] == 'FeatureCollection':
feats = [map_geometries(func, feat) for feat in obj['features']]
return {'type': obj['type'], 'features': feats}
else:
raise ValueError("Invalid GeoJSON object %s" % repr(obj)) | python | def map_geometries(func, obj):
simple_types = [
'Point',
'LineString',
'MultiPoint',
'MultiLineString',
'Polygon',
'MultiPolygon',
]
if obj['type'] in simple_types:
return func(obj)
elif obj['type'] == 'GeometryCollection':
geoms = [func(geom) if geom else None for geom in obj['geometries']]
return {'type': obj['type'], 'geometries': geoms}
elif obj['type'] == 'Feature':
geom = func(obj['geometry']) if obj['geometry'] else None
return {
'type': obj['type'],
'geometry': geom,
'properties': obj['properties'],
}
elif obj['type'] == 'FeatureCollection':
feats = [map_geometries(func, feat) for feat in obj['features']]
return {'type': obj['type'], 'features': feats}
else:
raise ValueError("Invalid GeoJSON object %s" % repr(obj)) | [
"def",
"map_geometries",
"(",
"func",
",",
"obj",
")",
":",
"simple_types",
"=",
"[",
"'Point'",
",",
"'LineString'",
",",
"'MultiPoint'",
",",
"'MultiLineString'",
",",
"'Polygon'",
",",
"'MultiPolygon'",
",",
"]",
"if",
"obj",
"[",
"'type'",
"]",
"in",
"simple_types",
":",
"return",
"func",
"(",
"obj",
")",
"elif",
"obj",
"[",
"'type'",
"]",
"==",
"'GeometryCollection'",
":",
"geoms",
"=",
"[",
"func",
"(",
"geom",
")",
"if",
"geom",
"else",
"None",
"for",
"geom",
"in",
"obj",
"[",
"'geometries'",
"]",
"]",
"return",
"{",
"'type'",
":",
"obj",
"[",
"'type'",
"]",
",",
"'geometries'",
":",
"geoms",
"}",
"elif",
"obj",
"[",
"'type'",
"]",
"==",
"'Feature'",
":",
"geom",
"=",
"func",
"(",
"obj",
"[",
"'geometry'",
"]",
")",
"if",
"obj",
"[",
"'geometry'",
"]",
"else",
"None",
"return",
"{",
"'type'",
":",
"obj",
"[",
"'type'",
"]",
",",
"'geometry'",
":",
"geom",
",",
"'properties'",
":",
"obj",
"[",
"'properties'",
"]",
",",
"}",
"elif",
"obj",
"[",
"'type'",
"]",
"==",
"'FeatureCollection'",
":",
"feats",
"=",
"[",
"map_geometries",
"(",
"func",
",",
"feat",
")",
"for",
"feat",
"in",
"obj",
"[",
"'features'",
"]",
"]",
"return",
"{",
"'type'",
":",
"obj",
"[",
"'type'",
"]",
",",
"'features'",
":",
"feats",
"}",
"else",
":",
"raise",
"ValueError",
"(",
"\"Invalid GeoJSON object %s\"",
"%",
"repr",
"(",
"obj",
")",
")"
] | Returns the result of passing every geometry in the given geojson object
through func.
:param func: Function to apply to tuples
:type func: function
:param obj: A geometry or feature to extract the coordinates from.
:type obj: GeoJSON
:return: The result of applying the function to each geometry
:rtype: list
:raises ValueError: if the provided object is not geojson. | [
"Returns",
"the",
"result",
"of",
"passing",
"every",
"geometry",
"in",
"the",
"given",
"geojson",
"object",
"through",
"func",
"."
] | 14abb31ba73a9f3cdb81d81c56656ea736f3c865 | https://github.com/jazzband/python-geojson/blob/14abb31ba73a9f3cdb81d81c56656ea736f3c865/geojson/utils.py#L94-L132 |
250,953 | jazzband/python-geojson | geojson/base.py | GeoJSON.to_instance | def to_instance(cls, ob, default=None, strict=False):
"""Encode a GeoJSON dict into an GeoJSON object.
Assumes the caller knows that the dict should satisfy a GeoJSON type.
:param cls: Dict containing the elements to be encoded into a GeoJSON
object.
:type cls: dict
:param ob: GeoJSON object into which to encode the dict provided in
`cls`.
:type ob: GeoJSON
:param default: A default instance to append the content of the dict
to if none is provided.
:type default: GeoJSON
:param strict: Raise error if unable to coerce particular keys or
attributes to a valid GeoJSON structure.
:type strict: bool
:return: A GeoJSON object with the dict's elements as its constituents.
:rtype: GeoJSON
:raises TypeError: If the input dict contains items that are not valid
GeoJSON types.
:raises UnicodeEncodeError: If the input dict contains items of a type
that contain non-ASCII characters.
:raises AttributeError: If the input dict contains items that are not
valid GeoJSON types.
"""
if ob is None and default is not None:
instance = default()
elif isinstance(ob, GeoJSON):
instance = ob
else:
mapping = to_mapping(ob)
d = {}
for k in mapping:
d[k] = mapping[k]
try:
type_ = d.pop("type")
try:
type_ = str(type_)
except UnicodeEncodeError:
# If the type contains non-ascii characters, we can assume
# it's not a valid GeoJSON type
raise AttributeError(
"{0} is not a GeoJSON type").format(type_)
geojson_factory = getattr(geojson.factory, type_)
instance = geojson_factory(**d)
except (AttributeError, KeyError) as invalid:
if strict:
msg = "Cannot coerce %r into a valid GeoJSON structure: %s"
msg %= (ob, invalid)
raise ValueError(msg)
instance = ob
return instance | python | def to_instance(cls, ob, default=None, strict=False):
if ob is None and default is not None:
instance = default()
elif isinstance(ob, GeoJSON):
instance = ob
else:
mapping = to_mapping(ob)
d = {}
for k in mapping:
d[k] = mapping[k]
try:
type_ = d.pop("type")
try:
type_ = str(type_)
except UnicodeEncodeError:
# If the type contains non-ascii characters, we can assume
# it's not a valid GeoJSON type
raise AttributeError(
"{0} is not a GeoJSON type").format(type_)
geojson_factory = getattr(geojson.factory, type_)
instance = geojson_factory(**d)
except (AttributeError, KeyError) as invalid:
if strict:
msg = "Cannot coerce %r into a valid GeoJSON structure: %s"
msg %= (ob, invalid)
raise ValueError(msg)
instance = ob
return instance | [
"def",
"to_instance",
"(",
"cls",
",",
"ob",
",",
"default",
"=",
"None",
",",
"strict",
"=",
"False",
")",
":",
"if",
"ob",
"is",
"None",
"and",
"default",
"is",
"not",
"None",
":",
"instance",
"=",
"default",
"(",
")",
"elif",
"isinstance",
"(",
"ob",
",",
"GeoJSON",
")",
":",
"instance",
"=",
"ob",
"else",
":",
"mapping",
"=",
"to_mapping",
"(",
"ob",
")",
"d",
"=",
"{",
"}",
"for",
"k",
"in",
"mapping",
":",
"d",
"[",
"k",
"]",
"=",
"mapping",
"[",
"k",
"]",
"try",
":",
"type_",
"=",
"d",
".",
"pop",
"(",
"\"type\"",
")",
"try",
":",
"type_",
"=",
"str",
"(",
"type_",
")",
"except",
"UnicodeEncodeError",
":",
"# If the type contains non-ascii characters, we can assume",
"# it's not a valid GeoJSON type",
"raise",
"AttributeError",
"(",
"\"{0} is not a GeoJSON type\"",
")",
".",
"format",
"(",
"type_",
")",
"geojson_factory",
"=",
"getattr",
"(",
"geojson",
".",
"factory",
",",
"type_",
")",
"instance",
"=",
"geojson_factory",
"(",
"*",
"*",
"d",
")",
"except",
"(",
"AttributeError",
",",
"KeyError",
")",
"as",
"invalid",
":",
"if",
"strict",
":",
"msg",
"=",
"\"Cannot coerce %r into a valid GeoJSON structure: %s\"",
"msg",
"%=",
"(",
"ob",
",",
"invalid",
")",
"raise",
"ValueError",
"(",
"msg",
")",
"instance",
"=",
"ob",
"return",
"instance"
] | Encode a GeoJSON dict into an GeoJSON object.
Assumes the caller knows that the dict should satisfy a GeoJSON type.
:param cls: Dict containing the elements to be encoded into a GeoJSON
object.
:type cls: dict
:param ob: GeoJSON object into which to encode the dict provided in
`cls`.
:type ob: GeoJSON
:param default: A default instance to append the content of the dict
to if none is provided.
:type default: GeoJSON
:param strict: Raise error if unable to coerce particular keys or
attributes to a valid GeoJSON structure.
:type strict: bool
:return: A GeoJSON object with the dict's elements as its constituents.
:rtype: GeoJSON
:raises TypeError: If the input dict contains items that are not valid
GeoJSON types.
:raises UnicodeEncodeError: If the input dict contains items of a type
that contain non-ASCII characters.
:raises AttributeError: If the input dict contains items that are not
valid GeoJSON types. | [
"Encode",
"a",
"GeoJSON",
"dict",
"into",
"an",
"GeoJSON",
"object",
".",
"Assumes",
"the",
"caller",
"knows",
"that",
"the",
"dict",
"should",
"satisfy",
"a",
"GeoJSON",
"type",
"."
] | 14abb31ba73a9f3cdb81d81c56656ea736f3c865 | https://github.com/jazzband/python-geojson/blob/14abb31ba73a9f3cdb81d81c56656ea736f3c865/geojson/base.py#L71-L122 |
250,954 | jazzband/python-geojson | geojson/base.py | GeoJSON.check_list_errors | def check_list_errors(self, checkFunc, lst):
"""Validation helper function."""
# check for errors on each subitem, filter only subitems with errors
results = (checkFunc(i) for i in lst)
return [err for err in results if err] | python | def check_list_errors(self, checkFunc, lst):
# check for errors on each subitem, filter only subitems with errors
results = (checkFunc(i) for i in lst)
return [err for err in results if err] | [
"def",
"check_list_errors",
"(",
"self",
",",
"checkFunc",
",",
"lst",
")",
":",
"# check for errors on each subitem, filter only subitems with errors",
"results",
"=",
"(",
"checkFunc",
"(",
"i",
")",
"for",
"i",
"in",
"lst",
")",
"return",
"[",
"err",
"for",
"err",
"in",
"results",
"if",
"err",
"]"
] | Validation helper function. | [
"Validation",
"helper",
"function",
"."
] | 14abb31ba73a9f3cdb81d81c56656ea736f3c865 | https://github.com/jazzband/python-geojson/blob/14abb31ba73a9f3cdb81d81c56656ea736f3c865/geojson/base.py#L128-L132 |
250,955 | mkorpela/pabot | pabot/pabotlib.py | PabotLib.run_only_once | def run_only_once(self, keyword):
"""
Runs a keyword only once in one of the parallel processes.
As the keyword will be called
only in one process and the return value could basically be anything.
The "Run Only Once" can't return the actual return value.
If the keyword fails, "Run Only Once" fails.
Others executing "Run Only Once" wait before going through this
keyword before the actual command has been executed.
NOTE! This is a potential "Shoot yourself in to knee" keyword
Especially note that all the namespace changes are only visible
in the process that actually executed the keyword.
Also note that this might lead to odd situations if used inside
of other keywords.
Also at this point the keyword will be identified to be same
if it has the same name.
"""
lock_name = 'pabot_run_only_once_%s' % keyword
try:
self.acquire_lock(lock_name)
passed = self.get_parallel_value_for_key(lock_name)
if passed != '':
if passed == 'FAILED':
raise AssertionError('Keyword failed in other process')
return
BuiltIn().run_keyword(keyword)
self.set_parallel_value_for_key(lock_name, 'PASSED')
except:
self.set_parallel_value_for_key(lock_name, 'FAILED')
raise
finally:
self.release_lock(lock_name) | python | def run_only_once(self, keyword):
lock_name = 'pabot_run_only_once_%s' % keyword
try:
self.acquire_lock(lock_name)
passed = self.get_parallel_value_for_key(lock_name)
if passed != '':
if passed == 'FAILED':
raise AssertionError('Keyword failed in other process')
return
BuiltIn().run_keyword(keyword)
self.set_parallel_value_for_key(lock_name, 'PASSED')
except:
self.set_parallel_value_for_key(lock_name, 'FAILED')
raise
finally:
self.release_lock(lock_name) | [
"def",
"run_only_once",
"(",
"self",
",",
"keyword",
")",
":",
"lock_name",
"=",
"'pabot_run_only_once_%s'",
"%",
"keyword",
"try",
":",
"self",
".",
"acquire_lock",
"(",
"lock_name",
")",
"passed",
"=",
"self",
".",
"get_parallel_value_for_key",
"(",
"lock_name",
")",
"if",
"passed",
"!=",
"''",
":",
"if",
"passed",
"==",
"'FAILED'",
":",
"raise",
"AssertionError",
"(",
"'Keyword failed in other process'",
")",
"return",
"BuiltIn",
"(",
")",
".",
"run_keyword",
"(",
"keyword",
")",
"self",
".",
"set_parallel_value_for_key",
"(",
"lock_name",
",",
"'PASSED'",
")",
"except",
":",
"self",
".",
"set_parallel_value_for_key",
"(",
"lock_name",
",",
"'FAILED'",
")",
"raise",
"finally",
":",
"self",
".",
"release_lock",
"(",
"lock_name",
")"
] | Runs a keyword only once in one of the parallel processes.
As the keyword will be called
only in one process and the return value could basically be anything.
The "Run Only Once" can't return the actual return value.
If the keyword fails, "Run Only Once" fails.
Others executing "Run Only Once" wait before going through this
keyword before the actual command has been executed.
NOTE! This is a potential "Shoot yourself in to knee" keyword
Especially note that all the namespace changes are only visible
in the process that actually executed the keyword.
Also note that this might lead to odd situations if used inside
of other keywords.
Also at this point the keyword will be identified to be same
if it has the same name. | [
"Runs",
"a",
"keyword",
"only",
"once",
"in",
"one",
"of",
"the",
"parallel",
"processes",
".",
"As",
"the",
"keyword",
"will",
"be",
"called",
"only",
"in",
"one",
"process",
"and",
"the",
"return",
"value",
"could",
"basically",
"be",
"anything",
".",
"The",
"Run",
"Only",
"Once",
"can",
"t",
"return",
"the",
"actual",
"return",
"value",
".",
"If",
"the",
"keyword",
"fails",
"Run",
"Only",
"Once",
"fails",
".",
"Others",
"executing",
"Run",
"Only",
"Once",
"wait",
"before",
"going",
"through",
"this",
"keyword",
"before",
"the",
"actual",
"command",
"has",
"been",
"executed",
".",
"NOTE!",
"This",
"is",
"a",
"potential",
"Shoot",
"yourself",
"in",
"to",
"knee",
"keyword",
"Especially",
"note",
"that",
"all",
"the",
"namespace",
"changes",
"are",
"only",
"visible",
"in",
"the",
"process",
"that",
"actually",
"executed",
"the",
"keyword",
".",
"Also",
"note",
"that",
"this",
"might",
"lead",
"to",
"odd",
"situations",
"if",
"used",
"inside",
"of",
"other",
"keywords",
".",
"Also",
"at",
"this",
"point",
"the",
"keyword",
"will",
"be",
"identified",
"to",
"be",
"same",
"if",
"it",
"has",
"the",
"same",
"name",
"."
] | b7d85546a58e398d579bb14fd9135858ec08a031 | https://github.com/mkorpela/pabot/blob/b7d85546a58e398d579bb14fd9135858ec08a031/pabot/pabotlib.py#L136-L167 |
250,956 | mkorpela/pabot | pabot/pabotlib.py | PabotLib.set_parallel_value_for_key | def set_parallel_value_for_key(self, key, value):
"""
Set a globally available key and value that can be accessed
from all the pabot processes.
"""
if self._remotelib:
self._remotelib.run_keyword('set_parallel_value_for_key',
[key, value], {})
else:
_PabotLib.set_parallel_value_for_key(self, key, value) | python | def set_parallel_value_for_key(self, key, value):
if self._remotelib:
self._remotelib.run_keyword('set_parallel_value_for_key',
[key, value], {})
else:
_PabotLib.set_parallel_value_for_key(self, key, value) | [
"def",
"set_parallel_value_for_key",
"(",
"self",
",",
"key",
",",
"value",
")",
":",
"if",
"self",
".",
"_remotelib",
":",
"self",
".",
"_remotelib",
".",
"run_keyword",
"(",
"'set_parallel_value_for_key'",
",",
"[",
"key",
",",
"value",
"]",
",",
"{",
"}",
")",
"else",
":",
"_PabotLib",
".",
"set_parallel_value_for_key",
"(",
"self",
",",
"key",
",",
"value",
")"
] | Set a globally available key and value that can be accessed
from all the pabot processes. | [
"Set",
"a",
"globally",
"available",
"key",
"and",
"value",
"that",
"can",
"be",
"accessed",
"from",
"all",
"the",
"pabot",
"processes",
"."
] | b7d85546a58e398d579bb14fd9135858ec08a031 | https://github.com/mkorpela/pabot/blob/b7d85546a58e398d579bb14fd9135858ec08a031/pabot/pabotlib.py#L169-L178 |
250,957 | mkorpela/pabot | pabot/pabotlib.py | PabotLib.get_parallel_value_for_key | def get_parallel_value_for_key(self, key):
"""
Get the value for a key. If there is no value for the key then empty
string is returned.
"""
if self._remotelib:
return self._remotelib.run_keyword('get_parallel_value_for_key',
[key], {})
return _PabotLib.get_parallel_value_for_key(self, key) | python | def get_parallel_value_for_key(self, key):
if self._remotelib:
return self._remotelib.run_keyword('get_parallel_value_for_key',
[key], {})
return _PabotLib.get_parallel_value_for_key(self, key) | [
"def",
"get_parallel_value_for_key",
"(",
"self",
",",
"key",
")",
":",
"if",
"self",
".",
"_remotelib",
":",
"return",
"self",
".",
"_remotelib",
".",
"run_keyword",
"(",
"'get_parallel_value_for_key'",
",",
"[",
"key",
"]",
",",
"{",
"}",
")",
"return",
"_PabotLib",
".",
"get_parallel_value_for_key",
"(",
"self",
",",
"key",
")"
] | Get the value for a key. If there is no value for the key then empty
string is returned. | [
"Get",
"the",
"value",
"for",
"a",
"key",
".",
"If",
"there",
"is",
"no",
"value",
"for",
"the",
"key",
"then",
"empty",
"string",
"is",
"returned",
"."
] | b7d85546a58e398d579bb14fd9135858ec08a031 | https://github.com/mkorpela/pabot/blob/b7d85546a58e398d579bb14fd9135858ec08a031/pabot/pabotlib.py#L180-L188 |
250,958 | mkorpela/pabot | pabot/pabotlib.py | PabotLib.acquire_lock | def acquire_lock(self, name):
"""
Wait for a lock with name.
This will prevent other processes from acquiring the lock with
the name while it is held. Thus they will wait in the position
where they are acquiring the lock until the process that has it
releases it.
"""
if self._remotelib:
try:
while not self._remotelib.run_keyword('acquire_lock',
[name, self._my_id], {}):
time.sleep(0.1)
logger.debug('waiting for lock to release')
return True
except RuntimeError:
logger.warn('no connection')
self.__remotelib = None
return _PabotLib.acquire_lock(self, name, self._my_id) | python | def acquire_lock(self, name):
if self._remotelib:
try:
while not self._remotelib.run_keyword('acquire_lock',
[name, self._my_id], {}):
time.sleep(0.1)
logger.debug('waiting for lock to release')
return True
except RuntimeError:
logger.warn('no connection')
self.__remotelib = None
return _PabotLib.acquire_lock(self, name, self._my_id) | [
"def",
"acquire_lock",
"(",
"self",
",",
"name",
")",
":",
"if",
"self",
".",
"_remotelib",
":",
"try",
":",
"while",
"not",
"self",
".",
"_remotelib",
".",
"run_keyword",
"(",
"'acquire_lock'",
",",
"[",
"name",
",",
"self",
".",
"_my_id",
"]",
",",
"{",
"}",
")",
":",
"time",
".",
"sleep",
"(",
"0.1",
")",
"logger",
".",
"debug",
"(",
"'waiting for lock to release'",
")",
"return",
"True",
"except",
"RuntimeError",
":",
"logger",
".",
"warn",
"(",
"'no connection'",
")",
"self",
".",
"__remotelib",
"=",
"None",
"return",
"_PabotLib",
".",
"acquire_lock",
"(",
"self",
",",
"name",
",",
"self",
".",
"_my_id",
")"
] | Wait for a lock with name.
This will prevent other processes from acquiring the lock with
the name while it is held. Thus they will wait in the position
where they are acquiring the lock until the process that has it
releases it. | [
"Wait",
"for",
"a",
"lock",
"with",
"name",
".",
"This",
"will",
"prevent",
"other",
"processes",
"from",
"acquiring",
"the",
"lock",
"with",
"the",
"name",
"while",
"it",
"is",
"held",
".",
"Thus",
"they",
"will",
"wait",
"in",
"the",
"position",
"where",
"they",
"are",
"acquiring",
"the",
"lock",
"until",
"the",
"process",
"that",
"has",
"it",
"releases",
"it",
"."
] | b7d85546a58e398d579bb14fd9135858ec08a031 | https://github.com/mkorpela/pabot/blob/b7d85546a58e398d579bb14fd9135858ec08a031/pabot/pabotlib.py#L190-L208 |
250,959 | mkorpela/pabot | pabot/pabotlib.py | PabotLib.release_lock | def release_lock(self, name):
"""
Release a lock with name.
This will enable others to acquire the lock.
"""
if self._remotelib:
self._remotelib.run_keyword('release_lock',
[name, self._my_id], {})
else:
_PabotLib.release_lock(self, name, self._my_id) | python | def release_lock(self, name):
if self._remotelib:
self._remotelib.run_keyword('release_lock',
[name, self._my_id], {})
else:
_PabotLib.release_lock(self, name, self._my_id) | [
"def",
"release_lock",
"(",
"self",
",",
"name",
")",
":",
"if",
"self",
".",
"_remotelib",
":",
"self",
".",
"_remotelib",
".",
"run_keyword",
"(",
"'release_lock'",
",",
"[",
"name",
",",
"self",
".",
"_my_id",
"]",
",",
"{",
"}",
")",
"else",
":",
"_PabotLib",
".",
"release_lock",
"(",
"self",
",",
"name",
",",
"self",
".",
"_my_id",
")"
] | Release a lock with name.
This will enable others to acquire the lock. | [
"Release",
"a",
"lock",
"with",
"name",
".",
"This",
"will",
"enable",
"others",
"to",
"acquire",
"the",
"lock",
"."
] | b7d85546a58e398d579bb14fd9135858ec08a031 | https://github.com/mkorpela/pabot/blob/b7d85546a58e398d579bb14fd9135858ec08a031/pabot/pabotlib.py#L210-L219 |
250,960 | mkorpela/pabot | pabot/pabotlib.py | PabotLib.release_locks | def release_locks(self):
"""
Release all locks called by instance.
"""
if self._remotelib:
self._remotelib.run_keyword('release_locks',
[self._my_id], {})
else:
_PabotLib.release_locks(self, self._my_id) | python | def release_locks(self):
if self._remotelib:
self._remotelib.run_keyword('release_locks',
[self._my_id], {})
else:
_PabotLib.release_locks(self, self._my_id) | [
"def",
"release_locks",
"(",
"self",
")",
":",
"if",
"self",
".",
"_remotelib",
":",
"self",
".",
"_remotelib",
".",
"run_keyword",
"(",
"'release_locks'",
",",
"[",
"self",
".",
"_my_id",
"]",
",",
"{",
"}",
")",
"else",
":",
"_PabotLib",
".",
"release_locks",
"(",
"self",
",",
"self",
".",
"_my_id",
")"
] | Release all locks called by instance. | [
"Release",
"all",
"locks",
"called",
"by",
"instance",
"."
] | b7d85546a58e398d579bb14fd9135858ec08a031 | https://github.com/mkorpela/pabot/blob/b7d85546a58e398d579bb14fd9135858ec08a031/pabot/pabotlib.py#L221-L229 |
250,961 | mkorpela/pabot | pabot/pabotlib.py | PabotLib.acquire_value_set | def acquire_value_set(self, *tags):
"""
Reserve a set of values for this execution.
No other process can reserve the same set of values while the set is
reserved. Acquired value set needs to be released after use to allow
other processes to access it.
Add tags to limit the possible value sets that this returns.
"""
setname = self._acquire_value_set(*tags)
if setname is None:
raise ValueError("Could not aquire a value set")
return setname | python | def acquire_value_set(self, *tags):
setname = self._acquire_value_set(*tags)
if setname is None:
raise ValueError("Could not aquire a value set")
return setname | [
"def",
"acquire_value_set",
"(",
"self",
",",
"*",
"tags",
")",
":",
"setname",
"=",
"self",
".",
"_acquire_value_set",
"(",
"*",
"tags",
")",
"if",
"setname",
"is",
"None",
":",
"raise",
"ValueError",
"(",
"\"Could not aquire a value set\"",
")",
"return",
"setname"
] | Reserve a set of values for this execution.
No other process can reserve the same set of values while the set is
reserved. Acquired value set needs to be released after use to allow
other processes to access it.
Add tags to limit the possible value sets that this returns. | [
"Reserve",
"a",
"set",
"of",
"values",
"for",
"this",
"execution",
".",
"No",
"other",
"process",
"can",
"reserve",
"the",
"same",
"set",
"of",
"values",
"while",
"the",
"set",
"is",
"reserved",
".",
"Acquired",
"value",
"set",
"needs",
"to",
"be",
"released",
"after",
"use",
"to",
"allow",
"other",
"processes",
"to",
"access",
"it",
".",
"Add",
"tags",
"to",
"limit",
"the",
"possible",
"value",
"sets",
"that",
"this",
"returns",
"."
] | b7d85546a58e398d579bb14fd9135858ec08a031 | https://github.com/mkorpela/pabot/blob/b7d85546a58e398d579bb14fd9135858ec08a031/pabot/pabotlib.py#L231-L242 |
250,962 | mkorpela/pabot | pabot/pabotlib.py | PabotLib.get_value_from_set | def get_value_from_set(self, key):
"""
Get a value from previously reserved value set.
"""
#TODO: This should be done locally.
# We do not really need to call centralised server if the set is already
# reserved as the data there is immutable during execution
key = key.lower()
if self._remotelib:
while True:
value = self._remotelib.run_keyword('get_value_from_set',
[key, self._my_id], {})
if value:
return value
time.sleep(0.1)
logger.debug('waiting for a value')
else:
return _PabotLib.get_value_from_set(self, key, self._my_id) | python | def get_value_from_set(self, key):
#TODO: This should be done locally.
# We do not really need to call centralised server if the set is already
# reserved as the data there is immutable during execution
key = key.lower()
if self._remotelib:
while True:
value = self._remotelib.run_keyword('get_value_from_set',
[key, self._my_id], {})
if value:
return value
time.sleep(0.1)
logger.debug('waiting for a value')
else:
return _PabotLib.get_value_from_set(self, key, self._my_id) | [
"def",
"get_value_from_set",
"(",
"self",
",",
"key",
")",
":",
"#TODO: This should be done locally. ",
"# We do not really need to call centralised server if the set is already",
"# reserved as the data there is immutable during execution",
"key",
"=",
"key",
".",
"lower",
"(",
")",
"if",
"self",
".",
"_remotelib",
":",
"while",
"True",
":",
"value",
"=",
"self",
".",
"_remotelib",
".",
"run_keyword",
"(",
"'get_value_from_set'",
",",
"[",
"key",
",",
"self",
".",
"_my_id",
"]",
",",
"{",
"}",
")",
"if",
"value",
":",
"return",
"value",
"time",
".",
"sleep",
"(",
"0.1",
")",
"logger",
".",
"debug",
"(",
"'waiting for a value'",
")",
"else",
":",
"return",
"_PabotLib",
".",
"get_value_from_set",
"(",
"self",
",",
"key",
",",
"self",
".",
"_my_id",
")"
] | Get a value from previously reserved value set. | [
"Get",
"a",
"value",
"from",
"previously",
"reserved",
"value",
"set",
"."
] | b7d85546a58e398d579bb14fd9135858ec08a031 | https://github.com/mkorpela/pabot/blob/b7d85546a58e398d579bb14fd9135858ec08a031/pabot/pabotlib.py#L260-L277 |
250,963 | mkorpela/pabot | pabot/pabotlib.py | PabotLib.release_value_set | def release_value_set(self):
"""
Release a reserved value set so that other executions can use it also.
"""
if self._remotelib:
self._remotelib.run_keyword('release_value_set', [self._my_id], {})
else:
_PabotLib.release_value_set(self, self._my_id) | python | def release_value_set(self):
if self._remotelib:
self._remotelib.run_keyword('release_value_set', [self._my_id], {})
else:
_PabotLib.release_value_set(self, self._my_id) | [
"def",
"release_value_set",
"(",
"self",
")",
":",
"if",
"self",
".",
"_remotelib",
":",
"self",
".",
"_remotelib",
".",
"run_keyword",
"(",
"'release_value_set'",
",",
"[",
"self",
".",
"_my_id",
"]",
",",
"{",
"}",
")",
"else",
":",
"_PabotLib",
".",
"release_value_set",
"(",
"self",
",",
"self",
".",
"_my_id",
")"
] | Release a reserved value set so that other executions can use it also. | [
"Release",
"a",
"reserved",
"value",
"set",
"so",
"that",
"other",
"executions",
"can",
"use",
"it",
"also",
"."
] | b7d85546a58e398d579bb14fd9135858ec08a031 | https://github.com/mkorpela/pabot/blob/b7d85546a58e398d579bb14fd9135858ec08a031/pabot/pabotlib.py#L279-L286 |
250,964 | uber-common/opentracing-python-instrumentation | opentracing_instrumentation/client_hooks/__init__.py | install_all_patches | def install_all_patches():
"""
A convenience method that installs all available hooks.
If a specific module is not available on the path, it is ignored.
"""
from . import mysqldb
from . import psycopg2
from . import strict_redis
from . import sqlalchemy
from . import tornado_http
from . import urllib
from . import urllib2
from . import requests
mysqldb.install_patches()
psycopg2.install_patches()
strict_redis.install_patches()
sqlalchemy.install_patches()
tornado_http.install_patches()
urllib.install_patches()
urllib2.install_patches()
requests.install_patches() | python | def install_all_patches():
from . import mysqldb
from . import psycopg2
from . import strict_redis
from . import sqlalchemy
from . import tornado_http
from . import urllib
from . import urllib2
from . import requests
mysqldb.install_patches()
psycopg2.install_patches()
strict_redis.install_patches()
sqlalchemy.install_patches()
tornado_http.install_patches()
urllib.install_patches()
urllib2.install_patches()
requests.install_patches() | [
"def",
"install_all_patches",
"(",
")",
":",
"from",
".",
"import",
"mysqldb",
"from",
".",
"import",
"psycopg2",
"from",
".",
"import",
"strict_redis",
"from",
".",
"import",
"sqlalchemy",
"from",
".",
"import",
"tornado_http",
"from",
".",
"import",
"urllib",
"from",
".",
"import",
"urllib2",
"from",
".",
"import",
"requests",
"mysqldb",
".",
"install_patches",
"(",
")",
"psycopg2",
".",
"install_patches",
"(",
")",
"strict_redis",
".",
"install_patches",
"(",
")",
"sqlalchemy",
".",
"install_patches",
"(",
")",
"tornado_http",
".",
"install_patches",
"(",
")",
"urllib",
".",
"install_patches",
"(",
")",
"urllib2",
".",
"install_patches",
"(",
")",
"requests",
".",
"install_patches",
"(",
")"
] | A convenience method that installs all available hooks.
If a specific module is not available on the path, it is ignored. | [
"A",
"convenience",
"method",
"that",
"installs",
"all",
"available",
"hooks",
"."
] | 57b29fb9f647e073cde8c75155f4708cb5661d20 | https://github.com/uber-common/opentracing-python-instrumentation/blob/57b29fb9f647e073cde8c75155f4708cb5661d20/opentracing_instrumentation/client_hooks/__init__.py#L33-L55 |
250,965 | uber-common/opentracing-python-instrumentation | opentracing_instrumentation/client_hooks/__init__.py | install_patches | def install_patches(patchers='all'):
"""
Usually called from middleware to install client hooks
specified in the client_hooks section of the configuration.
:param patchers: a list of patchers to run. Acceptable values include:
* None - installs all client patches
* 'all' - installs all client patches
* empty list - does not install any patches
* list of function names - executes the functions
"""
if patchers is None or patchers == 'all':
install_all_patches()
return
if not _valid_args(patchers):
raise ValueError('patchers argument must be None, "all", or a list')
for patch_func_name in patchers:
logging.info('Loading client hook %s', patch_func_name)
patch_func = _load_symbol(patch_func_name)
logging.info('Applying client hook %s', patch_func_name)
patch_func() | python | def install_patches(patchers='all'):
if patchers is None or patchers == 'all':
install_all_patches()
return
if not _valid_args(patchers):
raise ValueError('patchers argument must be None, "all", or a list')
for patch_func_name in patchers:
logging.info('Loading client hook %s', patch_func_name)
patch_func = _load_symbol(patch_func_name)
logging.info('Applying client hook %s', patch_func_name)
patch_func() | [
"def",
"install_patches",
"(",
"patchers",
"=",
"'all'",
")",
":",
"if",
"patchers",
"is",
"None",
"or",
"patchers",
"==",
"'all'",
":",
"install_all_patches",
"(",
")",
"return",
"if",
"not",
"_valid_args",
"(",
"patchers",
")",
":",
"raise",
"ValueError",
"(",
"'patchers argument must be None, \"all\", or a list'",
")",
"for",
"patch_func_name",
"in",
"patchers",
":",
"logging",
".",
"info",
"(",
"'Loading client hook %s'",
",",
"patch_func_name",
")",
"patch_func",
"=",
"_load_symbol",
"(",
"patch_func_name",
")",
"logging",
".",
"info",
"(",
"'Applying client hook %s'",
",",
"patch_func_name",
")",
"patch_func",
"(",
")"
] | Usually called from middleware to install client hooks
specified in the client_hooks section of the configuration.
:param patchers: a list of patchers to run. Acceptable values include:
* None - installs all client patches
* 'all' - installs all client patches
* empty list - does not install any patches
* list of function names - executes the functions | [
"Usually",
"called",
"from",
"middleware",
"to",
"install",
"client",
"hooks",
"specified",
"in",
"the",
"client_hooks",
"section",
"of",
"the",
"configuration",
"."
] | 57b29fb9f647e073cde8c75155f4708cb5661d20 | https://github.com/uber-common/opentracing-python-instrumentation/blob/57b29fb9f647e073cde8c75155f4708cb5661d20/opentracing_instrumentation/client_hooks/__init__.py#L58-L79 |
250,966 | uber-common/opentracing-python-instrumentation | opentracing_instrumentation/client_hooks/__init__.py | install_client_interceptors | def install_client_interceptors(client_interceptors=()):
"""
Install client interceptors for the patchers.
:param client_interceptors: a list of client interceptors to install.
Should be a list of classes
"""
if not _valid_args(client_interceptors):
raise ValueError('client_interceptors argument must be a list')
from ..http_client import ClientInterceptors
for client_interceptor in client_interceptors:
logging.info('Loading client interceptor %s', client_interceptor)
interceptor_class = _load_symbol(client_interceptor)
logging.info('Adding client interceptor %s', client_interceptor)
ClientInterceptors.append(interceptor_class()) | python | def install_client_interceptors(client_interceptors=()):
if not _valid_args(client_interceptors):
raise ValueError('client_interceptors argument must be a list')
from ..http_client import ClientInterceptors
for client_interceptor in client_interceptors:
logging.info('Loading client interceptor %s', client_interceptor)
interceptor_class = _load_symbol(client_interceptor)
logging.info('Adding client interceptor %s', client_interceptor)
ClientInterceptors.append(interceptor_class()) | [
"def",
"install_client_interceptors",
"(",
"client_interceptors",
"=",
"(",
")",
")",
":",
"if",
"not",
"_valid_args",
"(",
"client_interceptors",
")",
":",
"raise",
"ValueError",
"(",
"'client_interceptors argument must be a list'",
")",
"from",
".",
".",
"http_client",
"import",
"ClientInterceptors",
"for",
"client_interceptor",
"in",
"client_interceptors",
":",
"logging",
".",
"info",
"(",
"'Loading client interceptor %s'",
",",
"client_interceptor",
")",
"interceptor_class",
"=",
"_load_symbol",
"(",
"client_interceptor",
")",
"logging",
".",
"info",
"(",
"'Adding client interceptor %s'",
",",
"client_interceptor",
")",
"ClientInterceptors",
".",
"append",
"(",
"interceptor_class",
"(",
")",
")"
] | Install client interceptors for the patchers.
:param client_interceptors: a list of client interceptors to install.
Should be a list of classes | [
"Install",
"client",
"interceptors",
"for",
"the",
"patchers",
"."
] | 57b29fb9f647e073cde8c75155f4708cb5661d20 | https://github.com/uber-common/opentracing-python-instrumentation/blob/57b29fb9f647e073cde8c75155f4708cb5661d20/opentracing_instrumentation/client_hooks/__init__.py#L82-L98 |
250,967 | uber-common/opentracing-python-instrumentation | opentracing_instrumentation/client_hooks/__init__.py | _load_symbol | def _load_symbol(name):
"""Load a symbol by name.
:param str name: The name to load, specified by `module.attr`.
:returns: The attribute value. If the specified module does not contain
the requested attribute then `None` is returned.
"""
module_name, key = name.rsplit('.', 1)
try:
module = importlib.import_module(module_name)
except ImportError as err:
# it's possible the symbol is a class method
module_name, class_name = module_name.rsplit('.', 1)
module = importlib.import_module(module_name)
cls = getattr(module, class_name, None)
if cls:
attr = getattr(cls, key, None)
else:
raise err
else:
attr = getattr(module, key, None)
if not callable(attr):
raise ValueError('%s is not callable (was %r)' % (name, attr))
return attr | python | def _load_symbol(name):
module_name, key = name.rsplit('.', 1)
try:
module = importlib.import_module(module_name)
except ImportError as err:
# it's possible the symbol is a class method
module_name, class_name = module_name.rsplit('.', 1)
module = importlib.import_module(module_name)
cls = getattr(module, class_name, None)
if cls:
attr = getattr(cls, key, None)
else:
raise err
else:
attr = getattr(module, key, None)
if not callable(attr):
raise ValueError('%s is not callable (was %r)' % (name, attr))
return attr | [
"def",
"_load_symbol",
"(",
"name",
")",
":",
"module_name",
",",
"key",
"=",
"name",
".",
"rsplit",
"(",
"'.'",
",",
"1",
")",
"try",
":",
"module",
"=",
"importlib",
".",
"import_module",
"(",
"module_name",
")",
"except",
"ImportError",
"as",
"err",
":",
"# it's possible the symbol is a class method",
"module_name",
",",
"class_name",
"=",
"module_name",
".",
"rsplit",
"(",
"'.'",
",",
"1",
")",
"module",
"=",
"importlib",
".",
"import_module",
"(",
"module_name",
")",
"cls",
"=",
"getattr",
"(",
"module",
",",
"class_name",
",",
"None",
")",
"if",
"cls",
":",
"attr",
"=",
"getattr",
"(",
"cls",
",",
"key",
",",
"None",
")",
"else",
":",
"raise",
"err",
"else",
":",
"attr",
"=",
"getattr",
"(",
"module",
",",
"key",
",",
"None",
")",
"if",
"not",
"callable",
"(",
"attr",
")",
":",
"raise",
"ValueError",
"(",
"'%s is not callable (was %r)'",
"%",
"(",
"name",
",",
"attr",
")",
")",
"return",
"attr"
] | Load a symbol by name.
:param str name: The name to load, specified by `module.attr`.
:returns: The attribute value. If the specified module does not contain
the requested attribute then `None` is returned. | [
"Load",
"a",
"symbol",
"by",
"name",
"."
] | 57b29fb9f647e073cde8c75155f4708cb5661d20 | https://github.com/uber-common/opentracing-python-instrumentation/blob/57b29fb9f647e073cde8c75155f4708cb5661d20/opentracing_instrumentation/client_hooks/__init__.py#L106-L129 |
250,968 | uber-common/opentracing-python-instrumentation | opentracing_instrumentation/request_context.py | span_in_stack_context | def span_in_stack_context(span):
"""
Create Tornado's StackContext that stores the given span in the
thread-local request context. This function is intended for use
in Tornado applications based on IOLoop, although will work fine
in single-threaded apps like Flask, albeit with more overhead.
## Usage example in Tornado application
Suppose you have a method `handle_request(request)` in the http server.
Instead of calling it directly, use a wrapper:
.. code-block:: python
from opentracing_instrumentation import request_context
@tornado.gen.coroutine
def handle_request_wrapper(request, actual_handler, *args, **kwargs)
request_wrapper = TornadoRequestWrapper(request=request)
span = http_server.before_request(request=request_wrapper)
with request_context.span_in_stack_context(span):
return actual_handler(*args, **kwargs)
:param span:
:return:
Return StackContext that wraps the request context.
"""
if not isinstance(opentracing.tracer.scope_manager, TornadoScopeManager):
raise RuntimeError('scope_manager is not TornadoScopeManager')
# Enter the newly created stack context so we have
# storage available for Span activation.
context = tracer_stack_context()
entered_context = _TracerEnteredStackContext(context)
if span is None:
return entered_context
opentracing.tracer.scope_manager.activate(span, False)
assert opentracing.tracer.active_span is not None
assert opentracing.tracer.active_span is span
return entered_context | python | def span_in_stack_context(span):
if not isinstance(opentracing.tracer.scope_manager, TornadoScopeManager):
raise RuntimeError('scope_manager is not TornadoScopeManager')
# Enter the newly created stack context so we have
# storage available for Span activation.
context = tracer_stack_context()
entered_context = _TracerEnteredStackContext(context)
if span is None:
return entered_context
opentracing.tracer.scope_manager.activate(span, False)
assert opentracing.tracer.active_span is not None
assert opentracing.tracer.active_span is span
return entered_context | [
"def",
"span_in_stack_context",
"(",
"span",
")",
":",
"if",
"not",
"isinstance",
"(",
"opentracing",
".",
"tracer",
".",
"scope_manager",
",",
"TornadoScopeManager",
")",
":",
"raise",
"RuntimeError",
"(",
"'scope_manager is not TornadoScopeManager'",
")",
"# Enter the newly created stack context so we have",
"# storage available for Span activation.",
"context",
"=",
"tracer_stack_context",
"(",
")",
"entered_context",
"=",
"_TracerEnteredStackContext",
"(",
"context",
")",
"if",
"span",
"is",
"None",
":",
"return",
"entered_context",
"opentracing",
".",
"tracer",
".",
"scope_manager",
".",
"activate",
"(",
"span",
",",
"False",
")",
"assert",
"opentracing",
".",
"tracer",
".",
"active_span",
"is",
"not",
"None",
"assert",
"opentracing",
".",
"tracer",
".",
"active_span",
"is",
"span",
"return",
"entered_context"
] | Create Tornado's StackContext that stores the given span in the
thread-local request context. This function is intended for use
in Tornado applications based on IOLoop, although will work fine
in single-threaded apps like Flask, albeit with more overhead.
## Usage example in Tornado application
Suppose you have a method `handle_request(request)` in the http server.
Instead of calling it directly, use a wrapper:
.. code-block:: python
from opentracing_instrumentation import request_context
@tornado.gen.coroutine
def handle_request_wrapper(request, actual_handler, *args, **kwargs)
request_wrapper = TornadoRequestWrapper(request=request)
span = http_server.before_request(request=request_wrapper)
with request_context.span_in_stack_context(span):
return actual_handler(*args, **kwargs)
:param span:
:return:
Return StackContext that wraps the request context. | [
"Create",
"Tornado",
"s",
"StackContext",
"that",
"stores",
"the",
"given",
"span",
"in",
"the",
"thread",
"-",
"local",
"request",
"context",
".",
"This",
"function",
"is",
"intended",
"for",
"use",
"in",
"Tornado",
"applications",
"based",
"on",
"IOLoop",
"although",
"will",
"work",
"fine",
"in",
"single",
"-",
"threaded",
"apps",
"like",
"Flask",
"albeit",
"with",
"more",
"overhead",
"."
] | 57b29fb9f647e073cde8c75155f4708cb5661d20 | https://github.com/uber-common/opentracing-python-instrumentation/blob/57b29fb9f647e073cde8c75155f4708cb5661d20/opentracing_instrumentation/request_context.py#L181-L226 |
250,969 | uber-common/opentracing-python-instrumentation | opentracing_instrumentation/local_span.py | traced_function | def traced_function(func=None, name=None, on_start=None,
require_active_trace=False):
"""
A decorator that enables tracing of the wrapped function or
Tornado co-routine provided there is a parent span already established.
.. code-block:: python
@traced_function
def my_function1(arg1, arg2=None)
...
:param func: decorated function or Tornado co-routine
:param name: optional name to use as the Span.operation_name.
If not provided, func.__name__ will be used.
:param on_start: an optional callback to be executed once the child span
is started, but before the decorated function is called. It can be
used to set any additional tags on the span, perhaps by inspecting
the decorated function arguments. The callback must have a signature
`(span, *args, *kwargs)`, where the last two collections are the
arguments passed to the actual decorated function.
.. code-block:: python
def extract_call_site_tag(span, *args, *kwargs)
if 'call_site_tag' in kwargs:
span.set_tag('call_site_tag', kwargs['call_site_tag'])
@traced_function(on_start=extract_call_site_tag)
@tornado.get.coroutine
def my_function(arg1, arg2=None, call_site_tag=None)
...
:param require_active_trace: controls what to do when there is no active
trace. If require_active_trace=True, then no span is created.
If require_active_trace=False, a new trace is started.
:return: returns a tracing decorator
"""
if func is None:
return functools.partial(traced_function, name=name,
on_start=on_start,
require_active_trace=require_active_trace)
if name:
operation_name = name
else:
operation_name = func.__name__
@functools.wraps(func)
def decorator(*args, **kwargs):
parent_span = get_current_span()
if parent_span is None and require_active_trace:
return func(*args, **kwargs)
span = utils.start_child_span(
operation_name=operation_name, parent=parent_span)
if callable(on_start):
on_start(span, *args, **kwargs)
# We explicitly invoke deactivation callback for the StackContext,
# because there are scenarios when it gets retained forever, for
# example when a Periodic Callback is scheduled lazily while in the
# scope of a tracing StackContext.
with span_in_stack_context(span) as deactivate_cb:
try:
res = func(*args, **kwargs)
# Tornado co-routines usually return futures, so we must wait
# until the future is completed, in order to accurately
# capture the function's execution time.
if tornado.concurrent.is_future(res):
def done_callback(future):
deactivate_cb()
exception = future.exception()
if exception is not None:
span.log(event='exception', payload=exception)
span.set_tag('error', 'true')
span.finish()
res.add_done_callback(done_callback)
else:
deactivate_cb()
span.finish()
return res
except Exception as e:
deactivate_cb()
span.log(event='exception', payload=e)
span.set_tag('error', 'true')
span.finish()
raise
return decorator | python | def traced_function(func=None, name=None, on_start=None,
require_active_trace=False):
if func is None:
return functools.partial(traced_function, name=name,
on_start=on_start,
require_active_trace=require_active_trace)
if name:
operation_name = name
else:
operation_name = func.__name__
@functools.wraps(func)
def decorator(*args, **kwargs):
parent_span = get_current_span()
if parent_span is None and require_active_trace:
return func(*args, **kwargs)
span = utils.start_child_span(
operation_name=operation_name, parent=parent_span)
if callable(on_start):
on_start(span, *args, **kwargs)
# We explicitly invoke deactivation callback for the StackContext,
# because there are scenarios when it gets retained forever, for
# example when a Periodic Callback is scheduled lazily while in the
# scope of a tracing StackContext.
with span_in_stack_context(span) as deactivate_cb:
try:
res = func(*args, **kwargs)
# Tornado co-routines usually return futures, so we must wait
# until the future is completed, in order to accurately
# capture the function's execution time.
if tornado.concurrent.is_future(res):
def done_callback(future):
deactivate_cb()
exception = future.exception()
if exception is not None:
span.log(event='exception', payload=exception)
span.set_tag('error', 'true')
span.finish()
res.add_done_callback(done_callback)
else:
deactivate_cb()
span.finish()
return res
except Exception as e:
deactivate_cb()
span.log(event='exception', payload=e)
span.set_tag('error', 'true')
span.finish()
raise
return decorator | [
"def",
"traced_function",
"(",
"func",
"=",
"None",
",",
"name",
"=",
"None",
",",
"on_start",
"=",
"None",
",",
"require_active_trace",
"=",
"False",
")",
":",
"if",
"func",
"is",
"None",
":",
"return",
"functools",
".",
"partial",
"(",
"traced_function",
",",
"name",
"=",
"name",
",",
"on_start",
"=",
"on_start",
",",
"require_active_trace",
"=",
"require_active_trace",
")",
"if",
"name",
":",
"operation_name",
"=",
"name",
"else",
":",
"operation_name",
"=",
"func",
".",
"__name__",
"@",
"functools",
".",
"wraps",
"(",
"func",
")",
"def",
"decorator",
"(",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"parent_span",
"=",
"get_current_span",
"(",
")",
"if",
"parent_span",
"is",
"None",
"and",
"require_active_trace",
":",
"return",
"func",
"(",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
"span",
"=",
"utils",
".",
"start_child_span",
"(",
"operation_name",
"=",
"operation_name",
",",
"parent",
"=",
"parent_span",
")",
"if",
"callable",
"(",
"on_start",
")",
":",
"on_start",
"(",
"span",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
"# We explicitly invoke deactivation callback for the StackContext,",
"# because there are scenarios when it gets retained forever, for",
"# example when a Periodic Callback is scheduled lazily while in the",
"# scope of a tracing StackContext.",
"with",
"span_in_stack_context",
"(",
"span",
")",
"as",
"deactivate_cb",
":",
"try",
":",
"res",
"=",
"func",
"(",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
"# Tornado co-routines usually return futures, so we must wait",
"# until the future is completed, in order to accurately",
"# capture the function's execution time.",
"if",
"tornado",
".",
"concurrent",
".",
"is_future",
"(",
"res",
")",
":",
"def",
"done_callback",
"(",
"future",
")",
":",
"deactivate_cb",
"(",
")",
"exception",
"=",
"future",
".",
"exception",
"(",
")",
"if",
"exception",
"is",
"not",
"None",
":",
"span",
".",
"log",
"(",
"event",
"=",
"'exception'",
",",
"payload",
"=",
"exception",
")",
"span",
".",
"set_tag",
"(",
"'error'",
",",
"'true'",
")",
"span",
".",
"finish",
"(",
")",
"res",
".",
"add_done_callback",
"(",
"done_callback",
")",
"else",
":",
"deactivate_cb",
"(",
")",
"span",
".",
"finish",
"(",
")",
"return",
"res",
"except",
"Exception",
"as",
"e",
":",
"deactivate_cb",
"(",
")",
"span",
".",
"log",
"(",
"event",
"=",
"'exception'",
",",
"payload",
"=",
"e",
")",
"span",
".",
"set_tag",
"(",
"'error'",
",",
"'true'",
")",
"span",
".",
"finish",
"(",
")",
"raise",
"return",
"decorator"
] | A decorator that enables tracing of the wrapped function or
Tornado co-routine provided there is a parent span already established.
.. code-block:: python
@traced_function
def my_function1(arg1, arg2=None)
...
:param func: decorated function or Tornado co-routine
:param name: optional name to use as the Span.operation_name.
If not provided, func.__name__ will be used.
:param on_start: an optional callback to be executed once the child span
is started, but before the decorated function is called. It can be
used to set any additional tags on the span, perhaps by inspecting
the decorated function arguments. The callback must have a signature
`(span, *args, *kwargs)`, where the last two collections are the
arguments passed to the actual decorated function.
.. code-block:: python
def extract_call_site_tag(span, *args, *kwargs)
if 'call_site_tag' in kwargs:
span.set_tag('call_site_tag', kwargs['call_site_tag'])
@traced_function(on_start=extract_call_site_tag)
@tornado.get.coroutine
def my_function(arg1, arg2=None, call_site_tag=None)
...
:param require_active_trace: controls what to do when there is no active
trace. If require_active_trace=True, then no span is created.
If require_active_trace=False, a new trace is started.
:return: returns a tracing decorator | [
"A",
"decorator",
"that",
"enables",
"tracing",
"of",
"the",
"wrapped",
"function",
"or",
"Tornado",
"co",
"-",
"routine",
"provided",
"there",
"is",
"a",
"parent",
"span",
"already",
"established",
"."
] | 57b29fb9f647e073cde8c75155f4708cb5661d20 | https://github.com/uber-common/opentracing-python-instrumentation/blob/57b29fb9f647e073cde8c75155f4708cb5661d20/opentracing_instrumentation/local_span.py#L64-L153 |
250,970 | uber-common/opentracing-python-instrumentation | opentracing_instrumentation/utils.py | start_child_span | def start_child_span(operation_name, tracer=None, parent=None, tags=None):
"""
Start a new span as a child of parent_span. If parent_span is None,
start a new root span.
:param operation_name: operation name
:param tracer: Tracer or None (defaults to opentracing.tracer)
:param parent: parent Span or None
:param tags: optional tags
:return: new span
"""
tracer = tracer or opentracing.tracer
return tracer.start_span(
operation_name=operation_name,
child_of=parent.context if parent else None,
tags=tags
) | python | def start_child_span(operation_name, tracer=None, parent=None, tags=None):
tracer = tracer or opentracing.tracer
return tracer.start_span(
operation_name=operation_name,
child_of=parent.context if parent else None,
tags=tags
) | [
"def",
"start_child_span",
"(",
"operation_name",
",",
"tracer",
"=",
"None",
",",
"parent",
"=",
"None",
",",
"tags",
"=",
"None",
")",
":",
"tracer",
"=",
"tracer",
"or",
"opentracing",
".",
"tracer",
"return",
"tracer",
".",
"start_span",
"(",
"operation_name",
"=",
"operation_name",
",",
"child_of",
"=",
"parent",
".",
"context",
"if",
"parent",
"else",
"None",
",",
"tags",
"=",
"tags",
")"
] | Start a new span as a child of parent_span. If parent_span is None,
start a new root span.
:param operation_name: operation name
:param tracer: Tracer or None (defaults to opentracing.tracer)
:param parent: parent Span or None
:param tags: optional tags
:return: new span | [
"Start",
"a",
"new",
"span",
"as",
"a",
"child",
"of",
"parent_span",
".",
"If",
"parent_span",
"is",
"None",
"start",
"a",
"new",
"root",
"span",
"."
] | 57b29fb9f647e073cde8c75155f4708cb5661d20 | https://github.com/uber-common/opentracing-python-instrumentation/blob/57b29fb9f647e073cde8c75155f4708cb5661d20/opentracing_instrumentation/utils.py#L25-L41 |
250,971 | uber-common/opentracing-python-instrumentation | opentracing_instrumentation/http_server.py | before_request | def before_request(request, tracer=None):
"""
Attempts to extract a tracing span from incoming request.
If no tracing context is passed in the headers, or the data
cannot be parsed, a new root span is started.
:param request: HTTP request with `.headers` property exposed
that satisfies a regular dictionary interface
:param tracer: optional tracer instance to use. If not specified
the global opentracing.tracer will be used.
:return: returns a new, already started span.
"""
if tracer is None: # pragma: no cover
tracer = opentracing.tracer
# we need to prepare tags upfront, mainly because RPC_SERVER tag must be
# set when starting the span, to support Zipkin's one-span-per-RPC model
tags_dict = {
tags.SPAN_KIND: tags.SPAN_KIND_RPC_SERVER,
tags.HTTP_URL: request.full_url,
}
remote_ip = request.remote_ip
if remote_ip:
tags_dict[tags.PEER_HOST_IPV4] = remote_ip
caller_name = request.caller_name
if caller_name:
tags_dict[tags.PEER_SERVICE] = caller_name
remote_port = request.remote_port
if remote_port:
tags_dict[tags.PEER_PORT] = remote_port
operation = request.operation
try:
carrier = {}
for key, value in six.iteritems(request.headers):
carrier[key] = value
parent_ctx = tracer.extract(
format=Format.HTTP_HEADERS, carrier=carrier
)
except Exception as e:
logging.exception('trace extract failed: %s' % e)
parent_ctx = None
span = tracer.start_span(
operation_name=operation,
child_of=parent_ctx,
tags=tags_dict)
return span | python | def before_request(request, tracer=None):
if tracer is None: # pragma: no cover
tracer = opentracing.tracer
# we need to prepare tags upfront, mainly because RPC_SERVER tag must be
# set when starting the span, to support Zipkin's one-span-per-RPC model
tags_dict = {
tags.SPAN_KIND: tags.SPAN_KIND_RPC_SERVER,
tags.HTTP_URL: request.full_url,
}
remote_ip = request.remote_ip
if remote_ip:
tags_dict[tags.PEER_HOST_IPV4] = remote_ip
caller_name = request.caller_name
if caller_name:
tags_dict[tags.PEER_SERVICE] = caller_name
remote_port = request.remote_port
if remote_port:
tags_dict[tags.PEER_PORT] = remote_port
operation = request.operation
try:
carrier = {}
for key, value in six.iteritems(request.headers):
carrier[key] = value
parent_ctx = tracer.extract(
format=Format.HTTP_HEADERS, carrier=carrier
)
except Exception as e:
logging.exception('trace extract failed: %s' % e)
parent_ctx = None
span = tracer.start_span(
operation_name=operation,
child_of=parent_ctx,
tags=tags_dict)
return span | [
"def",
"before_request",
"(",
"request",
",",
"tracer",
"=",
"None",
")",
":",
"if",
"tracer",
"is",
"None",
":",
"# pragma: no cover",
"tracer",
"=",
"opentracing",
".",
"tracer",
"# we need to prepare tags upfront, mainly because RPC_SERVER tag must be",
"# set when starting the span, to support Zipkin's one-span-per-RPC model",
"tags_dict",
"=",
"{",
"tags",
".",
"SPAN_KIND",
":",
"tags",
".",
"SPAN_KIND_RPC_SERVER",
",",
"tags",
".",
"HTTP_URL",
":",
"request",
".",
"full_url",
",",
"}",
"remote_ip",
"=",
"request",
".",
"remote_ip",
"if",
"remote_ip",
":",
"tags_dict",
"[",
"tags",
".",
"PEER_HOST_IPV4",
"]",
"=",
"remote_ip",
"caller_name",
"=",
"request",
".",
"caller_name",
"if",
"caller_name",
":",
"tags_dict",
"[",
"tags",
".",
"PEER_SERVICE",
"]",
"=",
"caller_name",
"remote_port",
"=",
"request",
".",
"remote_port",
"if",
"remote_port",
":",
"tags_dict",
"[",
"tags",
".",
"PEER_PORT",
"]",
"=",
"remote_port",
"operation",
"=",
"request",
".",
"operation",
"try",
":",
"carrier",
"=",
"{",
"}",
"for",
"key",
",",
"value",
"in",
"six",
".",
"iteritems",
"(",
"request",
".",
"headers",
")",
":",
"carrier",
"[",
"key",
"]",
"=",
"value",
"parent_ctx",
"=",
"tracer",
".",
"extract",
"(",
"format",
"=",
"Format",
".",
"HTTP_HEADERS",
",",
"carrier",
"=",
"carrier",
")",
"except",
"Exception",
"as",
"e",
":",
"logging",
".",
"exception",
"(",
"'trace extract failed: %s'",
"%",
"e",
")",
"parent_ctx",
"=",
"None",
"span",
"=",
"tracer",
".",
"start_span",
"(",
"operation_name",
"=",
"operation",
",",
"child_of",
"=",
"parent_ctx",
",",
"tags",
"=",
"tags_dict",
")",
"return",
"span"
] | Attempts to extract a tracing span from incoming request.
If no tracing context is passed in the headers, or the data
cannot be parsed, a new root span is started.
:param request: HTTP request with `.headers` property exposed
that satisfies a regular dictionary interface
:param tracer: optional tracer instance to use. If not specified
the global opentracing.tracer will be used.
:return: returns a new, already started span. | [
"Attempts",
"to",
"extract",
"a",
"tracing",
"span",
"from",
"incoming",
"request",
".",
"If",
"no",
"tracing",
"context",
"is",
"passed",
"in",
"the",
"headers",
"or",
"the",
"data",
"cannot",
"be",
"parsed",
"a",
"new",
"root",
"span",
"is",
"started",
"."
] | 57b29fb9f647e073cde8c75155f4708cb5661d20 | https://github.com/uber-common/opentracing-python-instrumentation/blob/57b29fb9f647e073cde8c75155f4708cb5661d20/opentracing_instrumentation/http_server.py#L35-L86 |
250,972 | uber-common/opentracing-python-instrumentation | opentracing_instrumentation/http_server.py | WSGIRequestWrapper._parse_wsgi_headers | def _parse_wsgi_headers(wsgi_environ):
"""
HTTP headers are presented in WSGI environment with 'HTTP_' prefix.
This method finds those headers, removes the prefix, converts
underscores to dashes, and converts to lower case.
:param wsgi_environ:
:return: returns a dictionary of headers
"""
prefix = 'HTTP_'
p_len = len(prefix)
# use .items() despite suspected memory pressure bc GC occasionally
# collects wsgi_environ.iteritems() during iteration.
headers = {
key[p_len:].replace('_', '-').lower():
val for (key, val) in wsgi_environ.items()
if key.startswith(prefix)}
return headers | python | def _parse_wsgi_headers(wsgi_environ):
prefix = 'HTTP_'
p_len = len(prefix)
# use .items() despite suspected memory pressure bc GC occasionally
# collects wsgi_environ.iteritems() during iteration.
headers = {
key[p_len:].replace('_', '-').lower():
val for (key, val) in wsgi_environ.items()
if key.startswith(prefix)}
return headers | [
"def",
"_parse_wsgi_headers",
"(",
"wsgi_environ",
")",
":",
"prefix",
"=",
"'HTTP_'",
"p_len",
"=",
"len",
"(",
"prefix",
")",
"# use .items() despite suspected memory pressure bc GC occasionally",
"# collects wsgi_environ.iteritems() during iteration.",
"headers",
"=",
"{",
"key",
"[",
"p_len",
":",
"]",
".",
"replace",
"(",
"'_'",
",",
"'-'",
")",
".",
"lower",
"(",
")",
":",
"val",
"for",
"(",
"key",
",",
"val",
")",
"in",
"wsgi_environ",
".",
"items",
"(",
")",
"if",
"key",
".",
"startswith",
"(",
"prefix",
")",
"}",
"return",
"headers"
] | HTTP headers are presented in WSGI environment with 'HTTP_' prefix.
This method finds those headers, removes the prefix, converts
underscores to dashes, and converts to lower case.
:param wsgi_environ:
:return: returns a dictionary of headers | [
"HTTP",
"headers",
"are",
"presented",
"in",
"WSGI",
"environment",
"with",
"HTTP_",
"prefix",
".",
"This",
"method",
"finds",
"those",
"headers",
"removes",
"the",
"prefix",
"converts",
"underscores",
"to",
"dashes",
"and",
"converts",
"to",
"lower",
"case",
"."
] | 57b29fb9f647e073cde8c75155f4708cb5661d20 | https://github.com/uber-common/opentracing-python-instrumentation/blob/57b29fb9f647e073cde8c75155f4708cb5661d20/opentracing_instrumentation/http_server.py#L174-L191 |
250,973 | uber-common/opentracing-python-instrumentation | opentracing_instrumentation/interceptors.py | ClientInterceptors.append | def append(cls, interceptor):
"""
Add interceptor to the end of the internal list.
Note: Raises ``ValueError`` if interceptor
does not extend ``OpenTracingInterceptor``
"""
cls._check(interceptor)
cls._interceptors.append(interceptor) | python | def append(cls, interceptor):
cls._check(interceptor)
cls._interceptors.append(interceptor) | [
"def",
"append",
"(",
"cls",
",",
"interceptor",
")",
":",
"cls",
".",
"_check",
"(",
"interceptor",
")",
"cls",
".",
"_interceptors",
".",
"append",
"(",
"interceptor",
")"
] | Add interceptor to the end of the internal list.
Note: Raises ``ValueError`` if interceptor
does not extend ``OpenTracingInterceptor`` | [
"Add",
"interceptor",
"to",
"the",
"end",
"of",
"the",
"internal",
"list",
"."
] | 57b29fb9f647e073cde8c75155f4708cb5661d20 | https://github.com/uber-common/opentracing-python-instrumentation/blob/57b29fb9f647e073cde8c75155f4708cb5661d20/opentracing_instrumentation/interceptors.py#L80-L88 |
250,974 | uber-common/opentracing-python-instrumentation | opentracing_instrumentation/interceptors.py | ClientInterceptors.insert | def insert(cls, index, interceptor):
"""
Add interceptor to the given index in the internal list.
Note: Raises ``ValueError`` if interceptor
does not extend ``OpenTracingInterceptor``
"""
cls._check(interceptor)
cls._interceptors.insert(index, interceptor) | python | def insert(cls, index, interceptor):
cls._check(interceptor)
cls._interceptors.insert(index, interceptor) | [
"def",
"insert",
"(",
"cls",
",",
"index",
",",
"interceptor",
")",
":",
"cls",
".",
"_check",
"(",
"interceptor",
")",
"cls",
".",
"_interceptors",
".",
"insert",
"(",
"index",
",",
"interceptor",
")"
] | Add interceptor to the given index in the internal list.
Note: Raises ``ValueError`` if interceptor
does not extend ``OpenTracingInterceptor`` | [
"Add",
"interceptor",
"to",
"the",
"given",
"index",
"in",
"the",
"internal",
"list",
"."
] | 57b29fb9f647e073cde8c75155f4708cb5661d20 | https://github.com/uber-common/opentracing-python-instrumentation/blob/57b29fb9f647e073cde8c75155f4708cb5661d20/opentracing_instrumentation/interceptors.py#L91-L99 |
250,975 | uber-common/opentracing-python-instrumentation | opentracing_instrumentation/client_hooks/_singleton.py | singleton | def singleton(func):
"""
This decorator allows you to make sure that a function is called once and
only once. Note that recursive functions will still work.
WARNING: Not thread-safe!!!
"""
@functools.wraps(func)
def wrapper(*args, **kwargs):
if wrapper.__call_state__ == CALLED:
return
ret = func(*args, **kwargs)
wrapper.__call_state__ = CALLED
return ret
def reset():
wrapper.__call_state__ = NOT_CALLED
wrapper.reset = reset
reset()
# save original func to be able to patch and restore multiple times from
# unit tests
wrapper.__original_func = func
return wrapper | python | def singleton(func):
@functools.wraps(func)
def wrapper(*args, **kwargs):
if wrapper.__call_state__ == CALLED:
return
ret = func(*args, **kwargs)
wrapper.__call_state__ = CALLED
return ret
def reset():
wrapper.__call_state__ = NOT_CALLED
wrapper.reset = reset
reset()
# save original func to be able to patch and restore multiple times from
# unit tests
wrapper.__original_func = func
return wrapper | [
"def",
"singleton",
"(",
"func",
")",
":",
"@",
"functools",
".",
"wraps",
"(",
"func",
")",
"def",
"wrapper",
"(",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"if",
"wrapper",
".",
"__call_state__",
"==",
"CALLED",
":",
"return",
"ret",
"=",
"func",
"(",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
"wrapper",
".",
"__call_state__",
"=",
"CALLED",
"return",
"ret",
"def",
"reset",
"(",
")",
":",
"wrapper",
".",
"__call_state__",
"=",
"NOT_CALLED",
"wrapper",
".",
"reset",
"=",
"reset",
"reset",
"(",
")",
"# save original func to be able to patch and restore multiple times from",
"# unit tests",
"wrapper",
".",
"__original_func",
"=",
"func",
"return",
"wrapper"
] | This decorator allows you to make sure that a function is called once and
only once. Note that recursive functions will still work.
WARNING: Not thread-safe!!! | [
"This",
"decorator",
"allows",
"you",
"to",
"make",
"sure",
"that",
"a",
"function",
"is",
"called",
"once",
"and",
"only",
"once",
".",
"Note",
"that",
"recursive",
"functions",
"will",
"still",
"work",
"."
] | 57b29fb9f647e073cde8c75155f4708cb5661d20 | https://github.com/uber-common/opentracing-python-instrumentation/blob/57b29fb9f647e073cde8c75155f4708cb5661d20/opentracing_instrumentation/client_hooks/_singleton.py#L30-L55 |
250,976 | ANTsX/ANTsPy | ants/utils/smooth_image.py | smooth_image | def smooth_image(image, sigma, sigma_in_physical_coordinates=True, FWHM=False, max_kernel_width=32):
"""
Smooth an image
ANTsR function: `smoothImage`
Arguments
---------
image
Image to smooth
sigma
Smoothing factor. Can be scalar, in which case the same sigma is applied to each dimension, or a vector of length dim(inimage) to specify a unique smoothness for each dimension.
sigma_in_physical_coordinates : boolean
If true, the smoothing factor is in millimeters; if false, it is in pixels.
FWHM : boolean
If true, sigma is interpreted as the full-width-half-max (FWHM) of the filter, not the sigma of a Gaussian kernel.
max_kernel_width : scalar
Maximum kernel width
Returns
-------
ANTsImage
Example
-------
>>> import ants
>>> image = ants.image_read( ants.get_ants_data('r16'))
>>> simage = ants.smooth_image(image, (1.2,1.5))
"""
if image.components == 1:
return _smooth_image_helper(image, sigma, sigma_in_physical_coordinates, FWHM, max_kernel_width)
else:
imagelist = utils.split_channels(image)
newimages = []
for image in imagelist:
newimage = _smooth_image_helper(image, sigma, sigma_in_physical_coordinates, FWHM, max_kernel_width)
newimages.append(newimage)
return utils.merge_channels(newimages) | python | def smooth_image(image, sigma, sigma_in_physical_coordinates=True, FWHM=False, max_kernel_width=32):
if image.components == 1:
return _smooth_image_helper(image, sigma, sigma_in_physical_coordinates, FWHM, max_kernel_width)
else:
imagelist = utils.split_channels(image)
newimages = []
for image in imagelist:
newimage = _smooth_image_helper(image, sigma, sigma_in_physical_coordinates, FWHM, max_kernel_width)
newimages.append(newimage)
return utils.merge_channels(newimages) | [
"def",
"smooth_image",
"(",
"image",
",",
"sigma",
",",
"sigma_in_physical_coordinates",
"=",
"True",
",",
"FWHM",
"=",
"False",
",",
"max_kernel_width",
"=",
"32",
")",
":",
"if",
"image",
".",
"components",
"==",
"1",
":",
"return",
"_smooth_image_helper",
"(",
"image",
",",
"sigma",
",",
"sigma_in_physical_coordinates",
",",
"FWHM",
",",
"max_kernel_width",
")",
"else",
":",
"imagelist",
"=",
"utils",
".",
"split_channels",
"(",
"image",
")",
"newimages",
"=",
"[",
"]",
"for",
"image",
"in",
"imagelist",
":",
"newimage",
"=",
"_smooth_image_helper",
"(",
"image",
",",
"sigma",
",",
"sigma_in_physical_coordinates",
",",
"FWHM",
",",
"max_kernel_width",
")",
"newimages",
".",
"append",
"(",
"newimage",
")",
"return",
"utils",
".",
"merge_channels",
"(",
"newimages",
")"
] | Smooth an image
ANTsR function: `smoothImage`
Arguments
---------
image
Image to smooth
sigma
Smoothing factor. Can be scalar, in which case the same sigma is applied to each dimension, or a vector of length dim(inimage) to specify a unique smoothness for each dimension.
sigma_in_physical_coordinates : boolean
If true, the smoothing factor is in millimeters; if false, it is in pixels.
FWHM : boolean
If true, sigma is interpreted as the full-width-half-max (FWHM) of the filter, not the sigma of a Gaussian kernel.
max_kernel_width : scalar
Maximum kernel width
Returns
-------
ANTsImage
Example
-------
>>> import ants
>>> image = ants.image_read( ants.get_ants_data('r16'))
>>> simage = ants.smooth_image(image, (1.2,1.5)) | [
"Smooth",
"an",
"image"
] | 638020af2cdfc5ff4bdb9809ffe67aa505727a3b | https://github.com/ANTsX/ANTsPy/blob/638020af2cdfc5ff4bdb9809ffe67aa505727a3b/ants/utils/smooth_image.py#L34-L75 |
250,977 | ANTsX/ANTsPy | ants/registration/build_template.py | build_template | def build_template(
initial_template=None,
image_list=None,
iterations = 3,
gradient_step = 0.2,
**kwargs ):
"""
Estimate an optimal template from an input image_list
ANTsR function: N/A
Arguments
---------
initial_template : ANTsImage
initialization for the template building
image_list : ANTsImages
images from which to estimate template
iterations : integer
number of template building iterations
gradient_step : scalar
for shape update gradient
kwargs : keyword args
extra arguments passed to ants registration
Returns
-------
ANTsImage
Example
-------
>>> import ants
>>> image = ants.image_read( ants.get_ants_data('r16') , 'float')
>>> image2 = ants.image_read( ants.get_ants_data('r27') , 'float')
>>> image3 = ants.image_read( ants.get_ants_data('r85') , 'float')
>>> timage = ants.build_template( image_list = ( image, image2, image3 ) )
"""
wt = 1.0 / len( image_list )
if initial_template is None:
initial_template = image_list[ 0 ] * 0
for i in range( len( image_list ) ):
initial_template = initial_template + image_list[ i ] * wt
xavg = initial_template.clone()
for i in range( iterations ):
for k in range( len( image_list ) ):
w1 = registration( xavg, image_list[k],
type_of_transform='SyN', **kwargs )
if k == 0:
wavg = iio.image_read( w1['fwdtransforms'][0] ) * wt
xavgNew = w1['warpedmovout'] * wt
else:
wavg = wavg + iio.image_read( w1['fwdtransforms'][0] ) * wt
xavgNew = xavgNew + w1['warpedmovout'] * wt
print( wavg.abs().mean() )
wscl = (-1.0) * gradient_step
wavg = wavg * wscl
wavgfn = mktemp(suffix='.nii.gz')
iio.image_write(wavg, wavgfn)
xavg = apply_transforms( xavg, xavg, wavgfn )
return xavg | python | def build_template(
initial_template=None,
image_list=None,
iterations = 3,
gradient_step = 0.2,
**kwargs ):
wt = 1.0 / len( image_list )
if initial_template is None:
initial_template = image_list[ 0 ] * 0
for i in range( len( image_list ) ):
initial_template = initial_template + image_list[ i ] * wt
xavg = initial_template.clone()
for i in range( iterations ):
for k in range( len( image_list ) ):
w1 = registration( xavg, image_list[k],
type_of_transform='SyN', **kwargs )
if k == 0:
wavg = iio.image_read( w1['fwdtransforms'][0] ) * wt
xavgNew = w1['warpedmovout'] * wt
else:
wavg = wavg + iio.image_read( w1['fwdtransforms'][0] ) * wt
xavgNew = xavgNew + w1['warpedmovout'] * wt
print( wavg.abs().mean() )
wscl = (-1.0) * gradient_step
wavg = wavg * wscl
wavgfn = mktemp(suffix='.nii.gz')
iio.image_write(wavg, wavgfn)
xavg = apply_transforms( xavg, xavg, wavgfn )
return xavg | [
"def",
"build_template",
"(",
"initial_template",
"=",
"None",
",",
"image_list",
"=",
"None",
",",
"iterations",
"=",
"3",
",",
"gradient_step",
"=",
"0.2",
",",
"*",
"*",
"kwargs",
")",
":",
"wt",
"=",
"1.0",
"/",
"len",
"(",
"image_list",
")",
"if",
"initial_template",
"is",
"None",
":",
"initial_template",
"=",
"image_list",
"[",
"0",
"]",
"*",
"0",
"for",
"i",
"in",
"range",
"(",
"len",
"(",
"image_list",
")",
")",
":",
"initial_template",
"=",
"initial_template",
"+",
"image_list",
"[",
"i",
"]",
"*",
"wt",
"xavg",
"=",
"initial_template",
".",
"clone",
"(",
")",
"for",
"i",
"in",
"range",
"(",
"iterations",
")",
":",
"for",
"k",
"in",
"range",
"(",
"len",
"(",
"image_list",
")",
")",
":",
"w1",
"=",
"registration",
"(",
"xavg",
",",
"image_list",
"[",
"k",
"]",
",",
"type_of_transform",
"=",
"'SyN'",
",",
"*",
"*",
"kwargs",
")",
"if",
"k",
"==",
"0",
":",
"wavg",
"=",
"iio",
".",
"image_read",
"(",
"w1",
"[",
"'fwdtransforms'",
"]",
"[",
"0",
"]",
")",
"*",
"wt",
"xavgNew",
"=",
"w1",
"[",
"'warpedmovout'",
"]",
"*",
"wt",
"else",
":",
"wavg",
"=",
"wavg",
"+",
"iio",
".",
"image_read",
"(",
"w1",
"[",
"'fwdtransforms'",
"]",
"[",
"0",
"]",
")",
"*",
"wt",
"xavgNew",
"=",
"xavgNew",
"+",
"w1",
"[",
"'warpedmovout'",
"]",
"*",
"wt",
"print",
"(",
"wavg",
".",
"abs",
"(",
")",
".",
"mean",
"(",
")",
")",
"wscl",
"=",
"(",
"-",
"1.0",
")",
"*",
"gradient_step",
"wavg",
"=",
"wavg",
"*",
"wscl",
"wavgfn",
"=",
"mktemp",
"(",
"suffix",
"=",
"'.nii.gz'",
")",
"iio",
".",
"image_write",
"(",
"wavg",
",",
"wavgfn",
")",
"xavg",
"=",
"apply_transforms",
"(",
"xavg",
",",
"xavg",
",",
"wavgfn",
")",
"return",
"xavg"
] | Estimate an optimal template from an input image_list
ANTsR function: N/A
Arguments
---------
initial_template : ANTsImage
initialization for the template building
image_list : ANTsImages
images from which to estimate template
iterations : integer
number of template building iterations
gradient_step : scalar
for shape update gradient
kwargs : keyword args
extra arguments passed to ants registration
Returns
-------
ANTsImage
Example
-------
>>> import ants
>>> image = ants.image_read( ants.get_ants_data('r16') , 'float')
>>> image2 = ants.image_read( ants.get_ants_data('r27') , 'float')
>>> image3 = ants.image_read( ants.get_ants_data('r85') , 'float')
>>> timage = ants.build_template( image_list = ( image, image2, image3 ) ) | [
"Estimate",
"an",
"optimal",
"template",
"from",
"an",
"input",
"image_list"
] | 638020af2cdfc5ff4bdb9809ffe67aa505727a3b | https://github.com/ANTsX/ANTsPy/blob/638020af2cdfc5ff4bdb9809ffe67aa505727a3b/ants/registration/build_template.py#L13-L77 |
250,978 | ANTsX/ANTsPy | ants/registration/resample_image.py | resample_image | def resample_image(image, resample_params, use_voxels=False, interp_type=1):
"""
Resample image by spacing or number of voxels with
various interpolators. Works with multi-channel images.
ANTsR function: `resampleImage`
Arguments
---------
image : ANTsImage
input image
resample_params : tuple/list
vector of size dimension with numeric values
use_voxels : boolean
True means interpret resample params as voxel counts
interp_type : integer
one of 0 (linear), 1 (nearest neighbor), 2 (gaussian), 3 (windowed sinc), 4 (bspline)
Returns
-------
ANTsImage
Example
-------
>>> import ants
>>> fi = ants.image_read( ants.get_ants_data("r16"))
>>> finn = ants.resample_image(fi,(50,60),True,0)
>>> filin = ants.resample_image(fi,(1.5,1.5),False,1)
"""
if image.components == 1:
inimage = image.clone('float')
outimage = image.clone('float')
rsampar = 'x'.join([str(rp) for rp in resample_params])
args = [image.dimension, inimage, outimage, rsampar, int(use_voxels), interp_type]
processed_args = utils._int_antsProcessArguments(args)
libfn = utils.get_lib_fn('ResampleImage')
libfn(processed_args)
outimage = outimage.clone(image.pixeltype)
return outimage
else:
raise ValueError('images with more than 1 component not currently supported') | python | def resample_image(image, resample_params, use_voxels=False, interp_type=1):
if image.components == 1:
inimage = image.clone('float')
outimage = image.clone('float')
rsampar = 'x'.join([str(rp) for rp in resample_params])
args = [image.dimension, inimage, outimage, rsampar, int(use_voxels), interp_type]
processed_args = utils._int_antsProcessArguments(args)
libfn = utils.get_lib_fn('ResampleImage')
libfn(processed_args)
outimage = outimage.clone(image.pixeltype)
return outimage
else:
raise ValueError('images with more than 1 component not currently supported') | [
"def",
"resample_image",
"(",
"image",
",",
"resample_params",
",",
"use_voxels",
"=",
"False",
",",
"interp_type",
"=",
"1",
")",
":",
"if",
"image",
".",
"components",
"==",
"1",
":",
"inimage",
"=",
"image",
".",
"clone",
"(",
"'float'",
")",
"outimage",
"=",
"image",
".",
"clone",
"(",
"'float'",
")",
"rsampar",
"=",
"'x'",
".",
"join",
"(",
"[",
"str",
"(",
"rp",
")",
"for",
"rp",
"in",
"resample_params",
"]",
")",
"args",
"=",
"[",
"image",
".",
"dimension",
",",
"inimage",
",",
"outimage",
",",
"rsampar",
",",
"int",
"(",
"use_voxels",
")",
",",
"interp_type",
"]",
"processed_args",
"=",
"utils",
".",
"_int_antsProcessArguments",
"(",
"args",
")",
"libfn",
"=",
"utils",
".",
"get_lib_fn",
"(",
"'ResampleImage'",
")",
"libfn",
"(",
"processed_args",
")",
"outimage",
"=",
"outimage",
".",
"clone",
"(",
"image",
".",
"pixeltype",
")",
"return",
"outimage",
"else",
":",
"raise",
"ValueError",
"(",
"'images with more than 1 component not currently supported'",
")"
] | Resample image by spacing or number of voxels with
various interpolators. Works with multi-channel images.
ANTsR function: `resampleImage`
Arguments
---------
image : ANTsImage
input image
resample_params : tuple/list
vector of size dimension with numeric values
use_voxels : boolean
True means interpret resample params as voxel counts
interp_type : integer
one of 0 (linear), 1 (nearest neighbor), 2 (gaussian), 3 (windowed sinc), 4 (bspline)
Returns
-------
ANTsImage
Example
-------
>>> import ants
>>> fi = ants.image_read( ants.get_ants_data("r16"))
>>> finn = ants.resample_image(fi,(50,60),True,0)
>>> filin = ants.resample_image(fi,(1.5,1.5),False,1) | [
"Resample",
"image",
"by",
"spacing",
"or",
"number",
"of",
"voxels",
"with",
"various",
"interpolators",
".",
"Works",
"with",
"multi",
"-",
"channel",
"images",
"."
] | 638020af2cdfc5ff4bdb9809ffe67aa505727a3b | https://github.com/ANTsX/ANTsPy/blob/638020af2cdfc5ff4bdb9809ffe67aa505727a3b/ants/registration/resample_image.py#L12-L56 |
250,979 | ANTsX/ANTsPy | ants/core/ants_transform.py | apply_ants_transform | def apply_ants_transform(transform, data, data_type="point", reference=None, **kwargs):
"""
Apply ANTsTransform to data
ANTsR function: `applyAntsrTransform`
Arguments
---------
transform : ANTsTransform
transform to apply to image
data : ndarray/list/tuple
data to which transform will be applied
data_type : string
type of data
Options :
'point'
'vector'
'image'
reference : ANTsImage
target space for transforming image
kwargs : kwargs
additional options passed to `apply_ants_transform_to_image`
Returns
-------
ANTsImage if data_type == 'point'
OR
tuple if data_type == 'point' or data_type == 'vector'
"""
return transform.apply(data, data_type, reference, **kwargs) | python | def apply_ants_transform(transform, data, data_type="point", reference=None, **kwargs):
return transform.apply(data, data_type, reference, **kwargs) | [
"def",
"apply_ants_transform",
"(",
"transform",
",",
"data",
",",
"data_type",
"=",
"\"point\"",
",",
"reference",
"=",
"None",
",",
"*",
"*",
"kwargs",
")",
":",
"return",
"transform",
".",
"apply",
"(",
"data",
",",
"data_type",
",",
"reference",
",",
"*",
"*",
"kwargs",
")"
] | Apply ANTsTransform to data
ANTsR function: `applyAntsrTransform`
Arguments
---------
transform : ANTsTransform
transform to apply to image
data : ndarray/list/tuple
data to which transform will be applied
data_type : string
type of data
Options :
'point'
'vector'
'image'
reference : ANTsImage
target space for transforming image
kwargs : kwargs
additional options passed to `apply_ants_transform_to_image`
Returns
-------
ANTsImage if data_type == 'point'
OR
tuple if data_type == 'point' or data_type == 'vector' | [
"Apply",
"ANTsTransform",
"to",
"data"
] | 638020af2cdfc5ff4bdb9809ffe67aa505727a3b | https://github.com/ANTsX/ANTsPy/blob/638020af2cdfc5ff4bdb9809ffe67aa505727a3b/ants/core/ants_transform.py#L225-L258 |
250,980 | ANTsX/ANTsPy | ants/core/ants_transform.py | compose_ants_transforms | def compose_ants_transforms(transform_list):
"""
Compose multiple ANTsTransform's together
ANTsR function: `composeAntsrTransforms`
Arguments
---------
transform_list : list/tuple of ANTsTransform object
list of transforms to compose together
Returns
-------
ANTsTransform
one transform that contains all given transforms
Example
-------
>>> import ants
>>> img = ants.image_read(ants.get_ants_data("r16")).clone('float')
>>> tx = ants.new_ants_transform(dimension=2)
>>> tx.set_parameters((0.9,0,0,1.1,10,11))
>>> inv_tx = tx.invert()
>>> single_tx = ants.compose_ants_transforms([tx, inv_tx])
>>> img_orig = single_tx.apply_to_image(img, img)
>>> rRotGenerator = ants.contrib.RandomRotate2D( ( 0, 40 ), reference=img )
>>> rShearGenerator=ants.contrib.RandomShear2D( (0,50), reference=img )
>>> tx1 = rRotGenerator.transform()
>>> tx2 = rShearGenerator.transform()
>>> rSrR = ants.compose_ants_transforms([tx1, tx2])
>>> rSrR.apply_to_image( img )
"""
precision = transform_list[0].precision
dimension = transform_list[0].dimension
for tx in transform_list:
if precision != tx.precision:
raise ValueError('All transforms must have the same precision')
if dimension != tx.dimension:
raise ValueError('All transforms must have the same dimension')
tx_ptr_list = list(reversed([tf.pointer for tf in transform_list]))
libfn = utils.get_lib_fn('composeTransforms%s' % (transform_list[0]._libsuffix))
itk_composed_tx = libfn(tx_ptr_list, precision, dimension)
return ANTsTransform(precision=precision, dimension=dimension,
transform_type='CompositeTransform', pointer=itk_composed_tx) | python | def compose_ants_transforms(transform_list):
precision = transform_list[0].precision
dimension = transform_list[0].dimension
for tx in transform_list:
if precision != tx.precision:
raise ValueError('All transforms must have the same precision')
if dimension != tx.dimension:
raise ValueError('All transforms must have the same dimension')
tx_ptr_list = list(reversed([tf.pointer for tf in transform_list]))
libfn = utils.get_lib_fn('composeTransforms%s' % (transform_list[0]._libsuffix))
itk_composed_tx = libfn(tx_ptr_list, precision, dimension)
return ANTsTransform(precision=precision, dimension=dimension,
transform_type='CompositeTransform', pointer=itk_composed_tx) | [
"def",
"compose_ants_transforms",
"(",
"transform_list",
")",
":",
"precision",
"=",
"transform_list",
"[",
"0",
"]",
".",
"precision",
"dimension",
"=",
"transform_list",
"[",
"0",
"]",
".",
"dimension",
"for",
"tx",
"in",
"transform_list",
":",
"if",
"precision",
"!=",
"tx",
".",
"precision",
":",
"raise",
"ValueError",
"(",
"'All transforms must have the same precision'",
")",
"if",
"dimension",
"!=",
"tx",
".",
"dimension",
":",
"raise",
"ValueError",
"(",
"'All transforms must have the same dimension'",
")",
"tx_ptr_list",
"=",
"list",
"(",
"reversed",
"(",
"[",
"tf",
".",
"pointer",
"for",
"tf",
"in",
"transform_list",
"]",
")",
")",
"libfn",
"=",
"utils",
".",
"get_lib_fn",
"(",
"'composeTransforms%s'",
"%",
"(",
"transform_list",
"[",
"0",
"]",
".",
"_libsuffix",
")",
")",
"itk_composed_tx",
"=",
"libfn",
"(",
"tx_ptr_list",
",",
"precision",
",",
"dimension",
")",
"return",
"ANTsTransform",
"(",
"precision",
"=",
"precision",
",",
"dimension",
"=",
"dimension",
",",
"transform_type",
"=",
"'CompositeTransform'",
",",
"pointer",
"=",
"itk_composed_tx",
")"
] | Compose multiple ANTsTransform's together
ANTsR function: `composeAntsrTransforms`
Arguments
---------
transform_list : list/tuple of ANTsTransform object
list of transforms to compose together
Returns
-------
ANTsTransform
one transform that contains all given transforms
Example
-------
>>> import ants
>>> img = ants.image_read(ants.get_ants_data("r16")).clone('float')
>>> tx = ants.new_ants_transform(dimension=2)
>>> tx.set_parameters((0.9,0,0,1.1,10,11))
>>> inv_tx = tx.invert()
>>> single_tx = ants.compose_ants_transforms([tx, inv_tx])
>>> img_orig = single_tx.apply_to_image(img, img)
>>> rRotGenerator = ants.contrib.RandomRotate2D( ( 0, 40 ), reference=img )
>>> rShearGenerator=ants.contrib.RandomShear2D( (0,50), reference=img )
>>> tx1 = rRotGenerator.transform()
>>> tx2 = rShearGenerator.transform()
>>> rSrR = ants.compose_ants_transforms([tx1, tx2])
>>> rSrR.apply_to_image( img ) | [
"Compose",
"multiple",
"ANTsTransform",
"s",
"together"
] | 638020af2cdfc5ff4bdb9809ffe67aa505727a3b | https://github.com/ANTsX/ANTsPy/blob/638020af2cdfc5ff4bdb9809ffe67aa505727a3b/ants/core/ants_transform.py#L355-L400 |
250,981 | ANTsX/ANTsPy | ants/core/ants_transform.py | transform_index_to_physical_point | def transform_index_to_physical_point(image, index):
"""
Get spatial point from index of an image.
ANTsR function: `antsTransformIndexToPhysicalPoint`
Arguments
---------
img : ANTsImage
image to get values from
index : list or tuple or numpy.ndarray
location in image
Returns
-------
tuple
Example
-------
>>> import ants
>>> import numpy as np
>>> img = ants.make_image((10,10),np.random.randn(100))
>>> pt = ants.transform_index_to_physical_point(img, (2,2))
"""
if not isinstance(image, iio.ANTsImage):
raise ValueError('image must be ANTsImage type')
if isinstance(index, np.ndarray):
index = index.tolist()
if not isinstance(index, (tuple,list)):
raise ValueError('index must be tuple or list')
if len(index) != image.dimension:
raise ValueError('len(index) != image.dimension')
index = [i+1 for i in index]
ndim = image.dimension
ptype = image.pixeltype
libfn = utils.get_lib_fn('TransformIndexToPhysicalPoint%s%i' % (utils.short_ptype(ptype), ndim))
point = libfn(image.pointer, [list(index)])
return np.array(point[0]) | python | def transform_index_to_physical_point(image, index):
if not isinstance(image, iio.ANTsImage):
raise ValueError('image must be ANTsImage type')
if isinstance(index, np.ndarray):
index = index.tolist()
if not isinstance(index, (tuple,list)):
raise ValueError('index must be tuple or list')
if len(index) != image.dimension:
raise ValueError('len(index) != image.dimension')
index = [i+1 for i in index]
ndim = image.dimension
ptype = image.pixeltype
libfn = utils.get_lib_fn('TransformIndexToPhysicalPoint%s%i' % (utils.short_ptype(ptype), ndim))
point = libfn(image.pointer, [list(index)])
return np.array(point[0]) | [
"def",
"transform_index_to_physical_point",
"(",
"image",
",",
"index",
")",
":",
"if",
"not",
"isinstance",
"(",
"image",
",",
"iio",
".",
"ANTsImage",
")",
":",
"raise",
"ValueError",
"(",
"'image must be ANTsImage type'",
")",
"if",
"isinstance",
"(",
"index",
",",
"np",
".",
"ndarray",
")",
":",
"index",
"=",
"index",
".",
"tolist",
"(",
")",
"if",
"not",
"isinstance",
"(",
"index",
",",
"(",
"tuple",
",",
"list",
")",
")",
":",
"raise",
"ValueError",
"(",
"'index must be tuple or list'",
")",
"if",
"len",
"(",
"index",
")",
"!=",
"image",
".",
"dimension",
":",
"raise",
"ValueError",
"(",
"'len(index) != image.dimension'",
")",
"index",
"=",
"[",
"i",
"+",
"1",
"for",
"i",
"in",
"index",
"]",
"ndim",
"=",
"image",
".",
"dimension",
"ptype",
"=",
"image",
".",
"pixeltype",
"libfn",
"=",
"utils",
".",
"get_lib_fn",
"(",
"'TransformIndexToPhysicalPoint%s%i'",
"%",
"(",
"utils",
".",
"short_ptype",
"(",
"ptype",
")",
",",
"ndim",
")",
")",
"point",
"=",
"libfn",
"(",
"image",
".",
"pointer",
",",
"[",
"list",
"(",
"index",
")",
"]",
")",
"return",
"np",
".",
"array",
"(",
"point",
"[",
"0",
"]",
")"
] | Get spatial point from index of an image.
ANTsR function: `antsTransformIndexToPhysicalPoint`
Arguments
---------
img : ANTsImage
image to get values from
index : list or tuple or numpy.ndarray
location in image
Returns
-------
tuple
Example
-------
>>> import ants
>>> import numpy as np
>>> img = ants.make_image((10,10),np.random.randn(100))
>>> pt = ants.transform_index_to_physical_point(img, (2,2)) | [
"Get",
"spatial",
"point",
"from",
"index",
"of",
"an",
"image",
"."
] | 638020af2cdfc5ff4bdb9809ffe67aa505727a3b | https://github.com/ANTsX/ANTsPy/blob/638020af2cdfc5ff4bdb9809ffe67aa505727a3b/ants/core/ants_transform.py#L403-L446 |
250,982 | ANTsX/ANTsPy | ants/core/ants_transform.py | ANTsTransform.invert | def invert(self):
""" Invert the transform """
libfn = utils.get_lib_fn('inverseTransform%s' % (self._libsuffix))
inv_tx_ptr = libfn(self.pointer)
new_tx = ANTsTransform(precision=self.precision, dimension=self.dimension,
transform_type=self.transform_type, pointer=inv_tx_ptr)
return new_tx | python | def invert(self):
libfn = utils.get_lib_fn('inverseTransform%s' % (self._libsuffix))
inv_tx_ptr = libfn(self.pointer)
new_tx = ANTsTransform(precision=self.precision, dimension=self.dimension,
transform_type=self.transform_type, pointer=inv_tx_ptr)
return new_tx | [
"def",
"invert",
"(",
"self",
")",
":",
"libfn",
"=",
"utils",
".",
"get_lib_fn",
"(",
"'inverseTransform%s'",
"%",
"(",
"self",
".",
"_libsuffix",
")",
")",
"inv_tx_ptr",
"=",
"libfn",
"(",
"self",
".",
"pointer",
")",
"new_tx",
"=",
"ANTsTransform",
"(",
"precision",
"=",
"self",
".",
"precision",
",",
"dimension",
"=",
"self",
".",
"dimension",
",",
"transform_type",
"=",
"self",
".",
"transform_type",
",",
"pointer",
"=",
"inv_tx_ptr",
")",
"return",
"new_tx"
] | Invert the transform | [
"Invert",
"the",
"transform"
] | 638020af2cdfc5ff4bdb9809ffe67aa505727a3b | https://github.com/ANTsX/ANTsPy/blob/638020af2cdfc5ff4bdb9809ffe67aa505727a3b/ants/core/ants_transform.py#L87-L95 |
250,983 | ANTsX/ANTsPy | ants/core/ants_transform.py | ANTsTransform.apply | def apply(self, data, data_type='point', reference=None, **kwargs):
"""
Apply transform to data
"""
if data_type == 'point':
return self.apply_to_point(data)
elif data_type == 'vector':
return self.apply_to_vector(data)
elif data_type == 'image':
return self.apply_to_image(data, reference, **kwargs) | python | def apply(self, data, data_type='point', reference=None, **kwargs):
if data_type == 'point':
return self.apply_to_point(data)
elif data_type == 'vector':
return self.apply_to_vector(data)
elif data_type == 'image':
return self.apply_to_image(data, reference, **kwargs) | [
"def",
"apply",
"(",
"self",
",",
"data",
",",
"data_type",
"=",
"'point'",
",",
"reference",
"=",
"None",
",",
"*",
"*",
"kwargs",
")",
":",
"if",
"data_type",
"==",
"'point'",
":",
"return",
"self",
".",
"apply_to_point",
"(",
"data",
")",
"elif",
"data_type",
"==",
"'vector'",
":",
"return",
"self",
".",
"apply_to_vector",
"(",
"data",
")",
"elif",
"data_type",
"==",
"'image'",
":",
"return",
"self",
".",
"apply_to_image",
"(",
"data",
",",
"reference",
",",
"*",
"*",
"kwargs",
")"
] | Apply transform to data | [
"Apply",
"transform",
"to",
"data"
] | 638020af2cdfc5ff4bdb9809ffe67aa505727a3b | https://github.com/ANTsX/ANTsPy/blob/638020af2cdfc5ff4bdb9809ffe67aa505727a3b/ants/core/ants_transform.py#L97-L107 |
250,984 | ANTsX/ANTsPy | ants/core/ants_transform.py | ANTsTransform.apply_to_point | def apply_to_point(self, point):
"""
Apply transform to a point
Arguments
---------
point : list/tuple
point to which the transform will be applied
Returns
-------
list : transformed point
Example
-------
>>> import ants
>>> tx = ants.new_ants_transform()
>>> params = tx.parameters
>>> tx.set_parameters(params*2)
>>> pt2 = tx.apply_to_point((1,2,3)) # should be (2,4,6)
"""
libfn = utils.get_lib_fn('transformPoint%s' % (self._libsuffix))
return tuple(libfn(self.pointer, point)) | python | def apply_to_point(self, point):
libfn = utils.get_lib_fn('transformPoint%s' % (self._libsuffix))
return tuple(libfn(self.pointer, point)) | [
"def",
"apply_to_point",
"(",
"self",
",",
"point",
")",
":",
"libfn",
"=",
"utils",
".",
"get_lib_fn",
"(",
"'transformPoint%s'",
"%",
"(",
"self",
".",
"_libsuffix",
")",
")",
"return",
"tuple",
"(",
"libfn",
"(",
"self",
".",
"pointer",
",",
"point",
")",
")"
] | Apply transform to a point
Arguments
---------
point : list/tuple
point to which the transform will be applied
Returns
-------
list : transformed point
Example
-------
>>> import ants
>>> tx = ants.new_ants_transform()
>>> params = tx.parameters
>>> tx.set_parameters(params*2)
>>> pt2 = tx.apply_to_point((1,2,3)) # should be (2,4,6) | [
"Apply",
"transform",
"to",
"a",
"point"
] | 638020af2cdfc5ff4bdb9809ffe67aa505727a3b | https://github.com/ANTsX/ANTsPy/blob/638020af2cdfc5ff4bdb9809ffe67aa505727a3b/ants/core/ants_transform.py#L109-L131 |
250,985 | ANTsX/ANTsPy | ants/core/ants_transform.py | ANTsTransform.apply_to_vector | def apply_to_vector(self, vector):
"""
Apply transform to a vector
Arguments
---------
vector : list/tuple
vector to which the transform will be applied
Returns
-------
list : transformed vector
"""
if isinstance(vector, np.ndarray):
vector = vector.tolist()
libfn = utils.get_lib_fn('transformVector%s' % (self._libsuffix))
return np.asarray(libfn(self.pointer, vector)) | python | def apply_to_vector(self, vector):
if isinstance(vector, np.ndarray):
vector = vector.tolist()
libfn = utils.get_lib_fn('transformVector%s' % (self._libsuffix))
return np.asarray(libfn(self.pointer, vector)) | [
"def",
"apply_to_vector",
"(",
"self",
",",
"vector",
")",
":",
"if",
"isinstance",
"(",
"vector",
",",
"np",
".",
"ndarray",
")",
":",
"vector",
"=",
"vector",
".",
"tolist",
"(",
")",
"libfn",
"=",
"utils",
".",
"get_lib_fn",
"(",
"'transformVector%s'",
"%",
"(",
"self",
".",
"_libsuffix",
")",
")",
"return",
"np",
".",
"asarray",
"(",
"libfn",
"(",
"self",
".",
"pointer",
",",
"vector",
")",
")"
] | Apply transform to a vector
Arguments
---------
vector : list/tuple
vector to which the transform will be applied
Returns
-------
list : transformed vector | [
"Apply",
"transform",
"to",
"a",
"vector"
] | 638020af2cdfc5ff4bdb9809ffe67aa505727a3b | https://github.com/ANTsX/ANTsPy/blob/638020af2cdfc5ff4bdb9809ffe67aa505727a3b/ants/core/ants_transform.py#L133-L150 |
250,986 | ANTsX/ANTsPy | ants/viz/plot.py | plot_hist | def plot_hist(image, threshold=0., fit_line=False, normfreq=True,
## plot label arguments
title=None, grid=True, xlabel=None, ylabel=None,
## other plot arguments
facecolor='green', alpha=0.75):
"""
Plot a histogram from an ANTsImage
Arguments
---------
image : ANTsImage
image from which histogram will be created
"""
img_arr = image.numpy().flatten()
img_arr = img_arr[np.abs(img_arr) > threshold]
if normfreq != False:
normfreq = 1. if normfreq == True else normfreq
n, bins, patches = plt.hist(img_arr, 50, normed=normfreq, facecolor=facecolor, alpha=alpha)
if fit_line:
# add a 'best fit' line
y = mlab.normpdf( bins, img_arr.mean(), img_arr.std())
l = plt.plot(bins, y, 'r--', linewidth=1)
if xlabel is not None:
plt.xlabel(xlabel)
if ylabel is not None:
plt.ylabel(ylabel)
if title is not None:
plt.title(title)
plt.grid(grid)
plt.show() | python | def plot_hist(image, threshold=0., fit_line=False, normfreq=True,
## plot label arguments
title=None, grid=True, xlabel=None, ylabel=None,
## other plot arguments
facecolor='green', alpha=0.75):
img_arr = image.numpy().flatten()
img_arr = img_arr[np.abs(img_arr) > threshold]
if normfreq != False:
normfreq = 1. if normfreq == True else normfreq
n, bins, patches = plt.hist(img_arr, 50, normed=normfreq, facecolor=facecolor, alpha=alpha)
if fit_line:
# add a 'best fit' line
y = mlab.normpdf( bins, img_arr.mean(), img_arr.std())
l = plt.plot(bins, y, 'r--', linewidth=1)
if xlabel is not None:
plt.xlabel(xlabel)
if ylabel is not None:
plt.ylabel(ylabel)
if title is not None:
plt.title(title)
plt.grid(grid)
plt.show() | [
"def",
"plot_hist",
"(",
"image",
",",
"threshold",
"=",
"0.",
",",
"fit_line",
"=",
"False",
",",
"normfreq",
"=",
"True",
",",
"## plot label arguments",
"title",
"=",
"None",
",",
"grid",
"=",
"True",
",",
"xlabel",
"=",
"None",
",",
"ylabel",
"=",
"None",
",",
"## other plot arguments",
"facecolor",
"=",
"'green'",
",",
"alpha",
"=",
"0.75",
")",
":",
"img_arr",
"=",
"image",
".",
"numpy",
"(",
")",
".",
"flatten",
"(",
")",
"img_arr",
"=",
"img_arr",
"[",
"np",
".",
"abs",
"(",
"img_arr",
")",
">",
"threshold",
"]",
"if",
"normfreq",
"!=",
"False",
":",
"normfreq",
"=",
"1.",
"if",
"normfreq",
"==",
"True",
"else",
"normfreq",
"n",
",",
"bins",
",",
"patches",
"=",
"plt",
".",
"hist",
"(",
"img_arr",
",",
"50",
",",
"normed",
"=",
"normfreq",
",",
"facecolor",
"=",
"facecolor",
",",
"alpha",
"=",
"alpha",
")",
"if",
"fit_line",
":",
"# add a 'best fit' line",
"y",
"=",
"mlab",
".",
"normpdf",
"(",
"bins",
",",
"img_arr",
".",
"mean",
"(",
")",
",",
"img_arr",
".",
"std",
"(",
")",
")",
"l",
"=",
"plt",
".",
"plot",
"(",
"bins",
",",
"y",
",",
"'r--'",
",",
"linewidth",
"=",
"1",
")",
"if",
"xlabel",
"is",
"not",
"None",
":",
"plt",
".",
"xlabel",
"(",
"xlabel",
")",
"if",
"ylabel",
"is",
"not",
"None",
":",
"plt",
".",
"ylabel",
"(",
"ylabel",
")",
"if",
"title",
"is",
"not",
"None",
":",
"plt",
".",
"title",
"(",
"title",
")",
"plt",
".",
"grid",
"(",
"grid",
")",
"plt",
".",
"show",
"(",
")"
] | Plot a histogram from an ANTsImage
Arguments
---------
image : ANTsImage
image from which histogram will be created | [
"Plot",
"a",
"histogram",
"from",
"an",
"ANTsImage"
] | 638020af2cdfc5ff4bdb9809ffe67aa505727a3b | https://github.com/ANTsX/ANTsPy/blob/638020af2cdfc5ff4bdb9809ffe67aa505727a3b/ants/viz/plot.py#L91-L124 |
250,987 | ANTsX/ANTsPy | ants/utils/morphology.py | morphology | def morphology(image, operation, radius, mtype='binary', value=1,
shape='ball', radius_is_parametric=False, thickness=1,
lines=3, include_center=False):
"""
Apply morphological operations to an image
ANTsR function: `morphology`
Arguments
---------
input : ANTsImage
input image
operation : string
operation to apply
"close" Morpholgical closing
"dilate" Morpholgical dilation
"erode" Morpholgical erosion
"open" Morpholgical opening
radius : scalar
radius of structuring element
mtype : string
type of morphology
"binary" Binary operation on a single value
"grayscale" Grayscale operations
value : scalar
value to operation on (type='binary' only)
shape : string
shape of the structuring element ( type='binary' only )
"ball" spherical structuring element
"box" box shaped structuring element
"cross" cross shaped structuring element
"annulus" annulus shaped structuring element
"polygon" polygon structuring element
radius_is_parametric : boolean
used parametric radius boolean (shape='ball' and shape='annulus' only)
thickness : scalar
thickness (shape='annulus' only)
lines : integer
number of lines in polygon (shape='polygon' only)
include_center : boolean
include center of annulus boolean (shape='annulus' only)
Returns
-------
ANTsImage
Example
-------
>>> import ants
>>> fi = ants.image_read( ants.get_ants_data('r16') , 2 )
>>> mask = ants.get_mask( fi )
>>> dilated_ball = ants.morphology( mask, operation='dilate', radius=3, mtype='binary', shape='ball')
>>> eroded_box = ants.morphology( mask, operation='erode', radius=3, mtype='binary', shape='box')
>>> opened_annulus = ants.morphology( mask, operation='open', radius=5, mtype='binary', shape='annulus', thickness=2)
"""
if image.components > 1:
raise ValueError('multichannel images not yet supported')
_sflag_dict = {'ball': 1, 'box': 2, 'cross': 3, 'annulus': 4, 'polygon': 5}
sFlag = _sflag_dict.get(shape, 0)
if sFlag == 0:
raise ValueError('invalid element shape')
radius_is_parametric = radius_is_parametric * 1
include_center = include_center * 1
if (mtype == 'binary'):
if (operation == 'dilate'):
if (sFlag == 5):
ret = iMath(image, 'MD', radius, value, sFlag, lines)
else:
ret = iMath(image, 'MD', radius, value, sFlag, radius_is_parametric, thickness, include_center)
elif (operation == 'erode'):
if (sFlag == 5):
ret = iMath(image, 'ME', radius, value, sFlag, lines)
else:
ret = iMath(image, 'ME', radius, value, sFlag, radius_is_parametric, thickness, include_center)
elif (operation == 'open'):
if (sFlag == 5):
ret = iMath(image, 'MO', radius, value, sFlag, lines)
else:
ret = iMath(image, 'MO', radius, value, sFlag, radius_is_parametric, thickness, include_center)
elif (operation == 'close'):
if (sFlag == 5):
ret = iMath(image, 'MC', radius, value, sFlag, lines)
else:
ret = iMath(image, 'MC', radius, value, sFlag, radius_is_parametric, thickness, include_center)
else:
raise ValueError('Invalid morphology operation')
elif (mtype == 'grayscale'):
if (operation == 'dilate'):
ret = iMath(image, 'GD', radius)
elif (operation == 'erode'):
ret = iMath(image, 'GE', radius)
elif (operation == 'open'):
ret = iMath(image, 'GO', radius)
elif (operation == 'close'):
ret = iMath(image, 'GC', radius)
else:
raise ValueError('Invalid morphology operation')
else:
raise ValueError('Invalid morphology type')
return ret | python | def morphology(image, operation, radius, mtype='binary', value=1,
shape='ball', radius_is_parametric=False, thickness=1,
lines=3, include_center=False):
if image.components > 1:
raise ValueError('multichannel images not yet supported')
_sflag_dict = {'ball': 1, 'box': 2, 'cross': 3, 'annulus': 4, 'polygon': 5}
sFlag = _sflag_dict.get(shape, 0)
if sFlag == 0:
raise ValueError('invalid element shape')
radius_is_parametric = radius_is_parametric * 1
include_center = include_center * 1
if (mtype == 'binary'):
if (operation == 'dilate'):
if (sFlag == 5):
ret = iMath(image, 'MD', radius, value, sFlag, lines)
else:
ret = iMath(image, 'MD', radius, value, sFlag, radius_is_parametric, thickness, include_center)
elif (operation == 'erode'):
if (sFlag == 5):
ret = iMath(image, 'ME', radius, value, sFlag, lines)
else:
ret = iMath(image, 'ME', radius, value, sFlag, radius_is_parametric, thickness, include_center)
elif (operation == 'open'):
if (sFlag == 5):
ret = iMath(image, 'MO', radius, value, sFlag, lines)
else:
ret = iMath(image, 'MO', radius, value, sFlag, radius_is_parametric, thickness, include_center)
elif (operation == 'close'):
if (sFlag == 5):
ret = iMath(image, 'MC', radius, value, sFlag, lines)
else:
ret = iMath(image, 'MC', radius, value, sFlag, radius_is_parametric, thickness, include_center)
else:
raise ValueError('Invalid morphology operation')
elif (mtype == 'grayscale'):
if (operation == 'dilate'):
ret = iMath(image, 'GD', radius)
elif (operation == 'erode'):
ret = iMath(image, 'GE', radius)
elif (operation == 'open'):
ret = iMath(image, 'GO', radius)
elif (operation == 'close'):
ret = iMath(image, 'GC', radius)
else:
raise ValueError('Invalid morphology operation')
else:
raise ValueError('Invalid morphology type')
return ret | [
"def",
"morphology",
"(",
"image",
",",
"operation",
",",
"radius",
",",
"mtype",
"=",
"'binary'",
",",
"value",
"=",
"1",
",",
"shape",
"=",
"'ball'",
",",
"radius_is_parametric",
"=",
"False",
",",
"thickness",
"=",
"1",
",",
"lines",
"=",
"3",
",",
"include_center",
"=",
"False",
")",
":",
"if",
"image",
".",
"components",
">",
"1",
":",
"raise",
"ValueError",
"(",
"'multichannel images not yet supported'",
")",
"_sflag_dict",
"=",
"{",
"'ball'",
":",
"1",
",",
"'box'",
":",
"2",
",",
"'cross'",
":",
"3",
",",
"'annulus'",
":",
"4",
",",
"'polygon'",
":",
"5",
"}",
"sFlag",
"=",
"_sflag_dict",
".",
"get",
"(",
"shape",
",",
"0",
")",
"if",
"sFlag",
"==",
"0",
":",
"raise",
"ValueError",
"(",
"'invalid element shape'",
")",
"radius_is_parametric",
"=",
"radius_is_parametric",
"*",
"1",
"include_center",
"=",
"include_center",
"*",
"1",
"if",
"(",
"mtype",
"==",
"'binary'",
")",
":",
"if",
"(",
"operation",
"==",
"'dilate'",
")",
":",
"if",
"(",
"sFlag",
"==",
"5",
")",
":",
"ret",
"=",
"iMath",
"(",
"image",
",",
"'MD'",
",",
"radius",
",",
"value",
",",
"sFlag",
",",
"lines",
")",
"else",
":",
"ret",
"=",
"iMath",
"(",
"image",
",",
"'MD'",
",",
"radius",
",",
"value",
",",
"sFlag",
",",
"radius_is_parametric",
",",
"thickness",
",",
"include_center",
")",
"elif",
"(",
"operation",
"==",
"'erode'",
")",
":",
"if",
"(",
"sFlag",
"==",
"5",
")",
":",
"ret",
"=",
"iMath",
"(",
"image",
",",
"'ME'",
",",
"radius",
",",
"value",
",",
"sFlag",
",",
"lines",
")",
"else",
":",
"ret",
"=",
"iMath",
"(",
"image",
",",
"'ME'",
",",
"radius",
",",
"value",
",",
"sFlag",
",",
"radius_is_parametric",
",",
"thickness",
",",
"include_center",
")",
"elif",
"(",
"operation",
"==",
"'open'",
")",
":",
"if",
"(",
"sFlag",
"==",
"5",
")",
":",
"ret",
"=",
"iMath",
"(",
"image",
",",
"'MO'",
",",
"radius",
",",
"value",
",",
"sFlag",
",",
"lines",
")",
"else",
":",
"ret",
"=",
"iMath",
"(",
"image",
",",
"'MO'",
",",
"radius",
",",
"value",
",",
"sFlag",
",",
"radius_is_parametric",
",",
"thickness",
",",
"include_center",
")",
"elif",
"(",
"operation",
"==",
"'close'",
")",
":",
"if",
"(",
"sFlag",
"==",
"5",
")",
":",
"ret",
"=",
"iMath",
"(",
"image",
",",
"'MC'",
",",
"radius",
",",
"value",
",",
"sFlag",
",",
"lines",
")",
"else",
":",
"ret",
"=",
"iMath",
"(",
"image",
",",
"'MC'",
",",
"radius",
",",
"value",
",",
"sFlag",
",",
"radius_is_parametric",
",",
"thickness",
",",
"include_center",
")",
"else",
":",
"raise",
"ValueError",
"(",
"'Invalid morphology operation'",
")",
"elif",
"(",
"mtype",
"==",
"'grayscale'",
")",
":",
"if",
"(",
"operation",
"==",
"'dilate'",
")",
":",
"ret",
"=",
"iMath",
"(",
"image",
",",
"'GD'",
",",
"radius",
")",
"elif",
"(",
"operation",
"==",
"'erode'",
")",
":",
"ret",
"=",
"iMath",
"(",
"image",
",",
"'GE'",
",",
"radius",
")",
"elif",
"(",
"operation",
"==",
"'open'",
")",
":",
"ret",
"=",
"iMath",
"(",
"image",
",",
"'GO'",
",",
"radius",
")",
"elif",
"(",
"operation",
"==",
"'close'",
")",
":",
"ret",
"=",
"iMath",
"(",
"image",
",",
"'GC'",
",",
"radius",
")",
"else",
":",
"raise",
"ValueError",
"(",
"'Invalid morphology operation'",
")",
"else",
":",
"raise",
"ValueError",
"(",
"'Invalid morphology type'",
")",
"return",
"ret"
] | Apply morphological operations to an image
ANTsR function: `morphology`
Arguments
---------
input : ANTsImage
input image
operation : string
operation to apply
"close" Morpholgical closing
"dilate" Morpholgical dilation
"erode" Morpholgical erosion
"open" Morpholgical opening
radius : scalar
radius of structuring element
mtype : string
type of morphology
"binary" Binary operation on a single value
"grayscale" Grayscale operations
value : scalar
value to operation on (type='binary' only)
shape : string
shape of the structuring element ( type='binary' only )
"ball" spherical structuring element
"box" box shaped structuring element
"cross" cross shaped structuring element
"annulus" annulus shaped structuring element
"polygon" polygon structuring element
radius_is_parametric : boolean
used parametric radius boolean (shape='ball' and shape='annulus' only)
thickness : scalar
thickness (shape='annulus' only)
lines : integer
number of lines in polygon (shape='polygon' only)
include_center : boolean
include center of annulus boolean (shape='annulus' only)
Returns
-------
ANTsImage
Example
-------
>>> import ants
>>> fi = ants.image_read( ants.get_ants_data('r16') , 2 )
>>> mask = ants.get_mask( fi )
>>> dilated_ball = ants.morphology( mask, operation='dilate', radius=3, mtype='binary', shape='ball')
>>> eroded_box = ants.morphology( mask, operation='erode', radius=3, mtype='binary', shape='box')
>>> opened_annulus = ants.morphology( mask, operation='open', radius=5, mtype='binary', shape='annulus', thickness=2) | [
"Apply",
"morphological",
"operations",
"to",
"an",
"image"
] | 638020af2cdfc5ff4bdb9809ffe67aa505727a3b | https://github.com/ANTsX/ANTsPy/blob/638020af2cdfc5ff4bdb9809ffe67aa505727a3b/ants/utils/morphology.py#L8-L120 |
250,988 | ANTsX/ANTsPy | ants/utils/scalar_rgb_vector.py | rgb_to_vector | def rgb_to_vector(image):
"""
Convert an RGB ANTsImage to a Vector ANTsImage
Arguments
---------
image : ANTsImage
RGB image to be converted
Returns
-------
ANTsImage
Example
-------
>>> import ants
>>> mni = ants.image_read(ants.get_data('mni'))
>>> mni_rgb = mni.scalar_to_rgb()
>>> mni_vector = mni.rgb_to_vector()
>>> mni_rgb2 = mni.vector_to_rgb()
"""
if image.pixeltype != 'unsigned char':
image = image.clone('unsigned char')
idim = image.dimension
libfn = utils.get_lib_fn('RgbToVector%i' % idim)
new_ptr = libfn(image.pointer)
new_img = iio.ANTsImage(pixeltype=image.pixeltype, dimension=image.dimension,
components=3, pointer=new_ptr, is_rgb=False)
return new_img | python | def rgb_to_vector(image):
if image.pixeltype != 'unsigned char':
image = image.clone('unsigned char')
idim = image.dimension
libfn = utils.get_lib_fn('RgbToVector%i' % idim)
new_ptr = libfn(image.pointer)
new_img = iio.ANTsImage(pixeltype=image.pixeltype, dimension=image.dimension,
components=3, pointer=new_ptr, is_rgb=False)
return new_img | [
"def",
"rgb_to_vector",
"(",
"image",
")",
":",
"if",
"image",
".",
"pixeltype",
"!=",
"'unsigned char'",
":",
"image",
"=",
"image",
".",
"clone",
"(",
"'unsigned char'",
")",
"idim",
"=",
"image",
".",
"dimension",
"libfn",
"=",
"utils",
".",
"get_lib_fn",
"(",
"'RgbToVector%i'",
"%",
"idim",
")",
"new_ptr",
"=",
"libfn",
"(",
"image",
".",
"pointer",
")",
"new_img",
"=",
"iio",
".",
"ANTsImage",
"(",
"pixeltype",
"=",
"image",
".",
"pixeltype",
",",
"dimension",
"=",
"image",
".",
"dimension",
",",
"components",
"=",
"3",
",",
"pointer",
"=",
"new_ptr",
",",
"is_rgb",
"=",
"False",
")",
"return",
"new_img"
] | Convert an RGB ANTsImage to a Vector ANTsImage
Arguments
---------
image : ANTsImage
RGB image to be converted
Returns
-------
ANTsImage
Example
-------
>>> import ants
>>> mni = ants.image_read(ants.get_data('mni'))
>>> mni_rgb = mni.scalar_to_rgb()
>>> mni_vector = mni.rgb_to_vector()
>>> mni_rgb2 = mni.vector_to_rgb() | [
"Convert",
"an",
"RGB",
"ANTsImage",
"to",
"a",
"Vector",
"ANTsImage"
] | 638020af2cdfc5ff4bdb9809ffe67aa505727a3b | https://github.com/ANTsX/ANTsPy/blob/638020af2cdfc5ff4bdb9809ffe67aa505727a3b/ants/utils/scalar_rgb_vector.py#L78-L106 |
250,989 | ANTsX/ANTsPy | ants/utils/scalar_rgb_vector.py | vector_to_rgb | def vector_to_rgb(image):
"""
Convert an Vector ANTsImage to a RGB ANTsImage
Arguments
---------
image : ANTsImage
RGB image to be converted
Returns
-------
ANTsImage
Example
-------
>>> import ants
>>> img = ants.image_read(ants.get_data('r16'), pixeltype='unsigned char')
>>> img_rgb = img.clone().scalar_to_rgb()
>>> img_vec = img_rgb.rgb_to_vector()
>>> img_rgb2 = img_vec.vector_to_rgb()
"""
if image.pixeltype != 'unsigned char':
image = image.clone('unsigned char')
idim = image.dimension
libfn = utils.get_lib_fn('VectorToRgb%i' % idim)
new_ptr = libfn(image.pointer)
new_img = iio.ANTsImage(pixeltype=image.pixeltype, dimension=image.dimension,
components=3, pointer=new_ptr, is_rgb=True)
return new_img | python | def vector_to_rgb(image):
if image.pixeltype != 'unsigned char':
image = image.clone('unsigned char')
idim = image.dimension
libfn = utils.get_lib_fn('VectorToRgb%i' % idim)
new_ptr = libfn(image.pointer)
new_img = iio.ANTsImage(pixeltype=image.pixeltype, dimension=image.dimension,
components=3, pointer=new_ptr, is_rgb=True)
return new_img | [
"def",
"vector_to_rgb",
"(",
"image",
")",
":",
"if",
"image",
".",
"pixeltype",
"!=",
"'unsigned char'",
":",
"image",
"=",
"image",
".",
"clone",
"(",
"'unsigned char'",
")",
"idim",
"=",
"image",
".",
"dimension",
"libfn",
"=",
"utils",
".",
"get_lib_fn",
"(",
"'VectorToRgb%i'",
"%",
"idim",
")",
"new_ptr",
"=",
"libfn",
"(",
"image",
".",
"pointer",
")",
"new_img",
"=",
"iio",
".",
"ANTsImage",
"(",
"pixeltype",
"=",
"image",
".",
"pixeltype",
",",
"dimension",
"=",
"image",
".",
"dimension",
",",
"components",
"=",
"3",
",",
"pointer",
"=",
"new_ptr",
",",
"is_rgb",
"=",
"True",
")",
"return",
"new_img"
] | Convert an Vector ANTsImage to a RGB ANTsImage
Arguments
---------
image : ANTsImage
RGB image to be converted
Returns
-------
ANTsImage
Example
-------
>>> import ants
>>> img = ants.image_read(ants.get_data('r16'), pixeltype='unsigned char')
>>> img_rgb = img.clone().scalar_to_rgb()
>>> img_vec = img_rgb.rgb_to_vector()
>>> img_rgb2 = img_vec.vector_to_rgb() | [
"Convert",
"an",
"Vector",
"ANTsImage",
"to",
"a",
"RGB",
"ANTsImage"
] | 638020af2cdfc5ff4bdb9809ffe67aa505727a3b | https://github.com/ANTsX/ANTsPy/blob/638020af2cdfc5ff4bdb9809ffe67aa505727a3b/ants/utils/scalar_rgb_vector.py#L109-L137 |
250,990 | ANTsX/ANTsPy | ants/utils/quantile.py | quantile | def quantile(image, q, nonzero=True):
"""
Get the quantile values from an ANTsImage
"""
img_arr = image.numpy()
if isinstance(q, (list,tuple)):
q = [qq*100. if qq <= 1. else qq for qq in q]
if nonzero:
img_arr = img_arr[img_arr>0]
vals = [np.percentile(img_arr, qq) for qq in q]
return tuple(vals)
elif isinstance(q, (float,int)):
if q <= 1.:
q = q*100.
if nonzero:
img_arr = img_arr[img_arr>0]
return np.percentile(img_arr[img_arr>0], q)
else:
raise ValueError('q argument must be list/tuple or float/int') | python | def quantile(image, q, nonzero=True):
img_arr = image.numpy()
if isinstance(q, (list,tuple)):
q = [qq*100. if qq <= 1. else qq for qq in q]
if nonzero:
img_arr = img_arr[img_arr>0]
vals = [np.percentile(img_arr, qq) for qq in q]
return tuple(vals)
elif isinstance(q, (float,int)):
if q <= 1.:
q = q*100.
if nonzero:
img_arr = img_arr[img_arr>0]
return np.percentile(img_arr[img_arr>0], q)
else:
raise ValueError('q argument must be list/tuple or float/int') | [
"def",
"quantile",
"(",
"image",
",",
"q",
",",
"nonzero",
"=",
"True",
")",
":",
"img_arr",
"=",
"image",
".",
"numpy",
"(",
")",
"if",
"isinstance",
"(",
"q",
",",
"(",
"list",
",",
"tuple",
")",
")",
":",
"q",
"=",
"[",
"qq",
"*",
"100.",
"if",
"qq",
"<=",
"1.",
"else",
"qq",
"for",
"qq",
"in",
"q",
"]",
"if",
"nonzero",
":",
"img_arr",
"=",
"img_arr",
"[",
"img_arr",
">",
"0",
"]",
"vals",
"=",
"[",
"np",
".",
"percentile",
"(",
"img_arr",
",",
"qq",
")",
"for",
"qq",
"in",
"q",
"]",
"return",
"tuple",
"(",
"vals",
")",
"elif",
"isinstance",
"(",
"q",
",",
"(",
"float",
",",
"int",
")",
")",
":",
"if",
"q",
"<=",
"1.",
":",
"q",
"=",
"q",
"*",
"100.",
"if",
"nonzero",
":",
"img_arr",
"=",
"img_arr",
"[",
"img_arr",
">",
"0",
"]",
"return",
"np",
".",
"percentile",
"(",
"img_arr",
"[",
"img_arr",
">",
"0",
"]",
",",
"q",
")",
"else",
":",
"raise",
"ValueError",
"(",
"'q argument must be list/tuple or float/int'",
")"
] | Get the quantile values from an ANTsImage | [
"Get",
"the",
"quantile",
"values",
"from",
"an",
"ANTsImage"
] | 638020af2cdfc5ff4bdb9809ffe67aa505727a3b | https://github.com/ANTsX/ANTsPy/blob/638020af2cdfc5ff4bdb9809ffe67aa505727a3b/ants/utils/quantile.py#L153-L171 |
250,991 | ANTsX/ANTsPy | ants/utils/quantile.py | bandpass_filter_matrix | def bandpass_filter_matrix( matrix,
tr=1, lowf=0.01, highf=0.1, order = 3):
"""
Bandpass filter the input time series image
ANTsR function: `frequencyFilterfMRI`
Arguments
---------
image: input time series image
tr: sampling time interval (inverse of sampling rate)
lowf: low frequency cutoff
highf: high frequency cutoff
order: order of the butterworth filter run using `filtfilt`
Returns
-------
filtered matrix
Example
-------
>>> import numpy as np
>>> import ants
>>> import matplotlib.pyplot as plt
>>> brainSignal = np.random.randn( 400, 1000 )
>>> tr = 1
>>> filtered = ants.bandpass_filter_matrix( brainSignal, tr = tr )
>>> nsamples = brainSignal.shape[0]
>>> t = np.linspace(0, tr*nsamples, nsamples, endpoint=False)
>>> k = 20
>>> plt.plot(t, brainSignal[:,k], label='Noisy signal')
>>> plt.plot(t, filtered[:,k], label='Filtered signal')
>>> plt.xlabel('time (seconds)')
>>> plt.grid(True)
>>> plt.axis('tight')
>>> plt.legend(loc='upper left')
>>> plt.show()
"""
from scipy.signal import butter, filtfilt
def butter_bandpass(lowcut, highcut, fs, order ):
nyq = 0.5 * fs
low = lowcut / nyq
high = highcut / nyq
b, a = butter(order, [low, high], btype='band')
return b, a
def butter_bandpass_filter(data, lowcut, highcut, fs, order ):
b, a = butter_bandpass(lowcut, highcut, fs, order=order)
y = filtfilt(b, a, data)
return y
fs = 1/tr # sampling rate based on tr
nsamples = matrix.shape[0]
ncolumns = matrix.shape[1]
matrixOut = matrix.copy()
for k in range( ncolumns ):
matrixOut[:,k] = butter_bandpass_filter(
matrix[:,k], lowf, highf, fs, order=order )
return matrixOut | python | def bandpass_filter_matrix( matrix,
tr=1, lowf=0.01, highf=0.1, order = 3):
from scipy.signal import butter, filtfilt
def butter_bandpass(lowcut, highcut, fs, order ):
nyq = 0.5 * fs
low = lowcut / nyq
high = highcut / nyq
b, a = butter(order, [low, high], btype='band')
return b, a
def butter_bandpass_filter(data, lowcut, highcut, fs, order ):
b, a = butter_bandpass(lowcut, highcut, fs, order=order)
y = filtfilt(b, a, data)
return y
fs = 1/tr # sampling rate based on tr
nsamples = matrix.shape[0]
ncolumns = matrix.shape[1]
matrixOut = matrix.copy()
for k in range( ncolumns ):
matrixOut[:,k] = butter_bandpass_filter(
matrix[:,k], lowf, highf, fs, order=order )
return matrixOut | [
"def",
"bandpass_filter_matrix",
"(",
"matrix",
",",
"tr",
"=",
"1",
",",
"lowf",
"=",
"0.01",
",",
"highf",
"=",
"0.1",
",",
"order",
"=",
"3",
")",
":",
"from",
"scipy",
".",
"signal",
"import",
"butter",
",",
"filtfilt",
"def",
"butter_bandpass",
"(",
"lowcut",
",",
"highcut",
",",
"fs",
",",
"order",
")",
":",
"nyq",
"=",
"0.5",
"*",
"fs",
"low",
"=",
"lowcut",
"/",
"nyq",
"high",
"=",
"highcut",
"/",
"nyq",
"b",
",",
"a",
"=",
"butter",
"(",
"order",
",",
"[",
"low",
",",
"high",
"]",
",",
"btype",
"=",
"'band'",
")",
"return",
"b",
",",
"a",
"def",
"butter_bandpass_filter",
"(",
"data",
",",
"lowcut",
",",
"highcut",
",",
"fs",
",",
"order",
")",
":",
"b",
",",
"a",
"=",
"butter_bandpass",
"(",
"lowcut",
",",
"highcut",
",",
"fs",
",",
"order",
"=",
"order",
")",
"y",
"=",
"filtfilt",
"(",
"b",
",",
"a",
",",
"data",
")",
"return",
"y",
"fs",
"=",
"1",
"/",
"tr",
"# sampling rate based on tr",
"nsamples",
"=",
"matrix",
".",
"shape",
"[",
"0",
"]",
"ncolumns",
"=",
"matrix",
".",
"shape",
"[",
"1",
"]",
"matrixOut",
"=",
"matrix",
".",
"copy",
"(",
")",
"for",
"k",
"in",
"range",
"(",
"ncolumns",
")",
":",
"matrixOut",
"[",
":",
",",
"k",
"]",
"=",
"butter_bandpass_filter",
"(",
"matrix",
"[",
":",
",",
"k",
"]",
",",
"lowf",
",",
"highf",
",",
"fs",
",",
"order",
"=",
"order",
")",
"return",
"matrixOut"
] | Bandpass filter the input time series image
ANTsR function: `frequencyFilterfMRI`
Arguments
---------
image: input time series image
tr: sampling time interval (inverse of sampling rate)
lowf: low frequency cutoff
highf: high frequency cutoff
order: order of the butterworth filter run using `filtfilt`
Returns
-------
filtered matrix
Example
-------
>>> import numpy as np
>>> import ants
>>> import matplotlib.pyplot as plt
>>> brainSignal = np.random.randn( 400, 1000 )
>>> tr = 1
>>> filtered = ants.bandpass_filter_matrix( brainSignal, tr = tr )
>>> nsamples = brainSignal.shape[0]
>>> t = np.linspace(0, tr*nsamples, nsamples, endpoint=False)
>>> k = 20
>>> plt.plot(t, brainSignal[:,k], label='Noisy signal')
>>> plt.plot(t, filtered[:,k], label='Filtered signal')
>>> plt.xlabel('time (seconds)')
>>> plt.grid(True)
>>> plt.axis('tight')
>>> plt.legend(loc='upper left')
>>> plt.show() | [
"Bandpass",
"filter",
"the",
"input",
"time",
"series",
"image"
] | 638020af2cdfc5ff4bdb9809ffe67aa505727a3b | https://github.com/ANTsX/ANTsPy/blob/638020af2cdfc5ff4bdb9809ffe67aa505727a3b/ants/utils/quantile.py#L227-L292 |
250,992 | ANTsX/ANTsPy | ants/utils/quantile.py | compcor | def compcor( boldImage, ncompcor=4, quantile=0.975, mask=None, filter_type=False, degree=2 ):
"""
Compute noise components from the input image
ANTsR function: `compcor`
this is adapted from nipy code https://github.com/nipy/nipype/blob/e29ac95fc0fc00fedbcaa0adaf29d5878408ca7c/nipype/algorithms/confounds.py
Arguments
---------
boldImage: input time series image
ncompcor: number of noise components to return
quantile: quantile defining high-variance
mask: mask defining brain or specific tissues
filter_type: type off filter to apply to time series before computing
noise components.
'polynomial' - Legendre polynomial basis
False - None (mean-removal only)
degree: order of polynomial used to remove trends from the timeseries
Returns
-------
dictionary containing:
components: a numpy array
basis: a numpy array containing the (non-constant) filter regressors
Example
-------
>>> cc = ants.compcor( ants.image_read(ants.get_ants_data("ch2")) )
"""
def compute_tSTD(M, quantile, x=0, axis=0):
stdM = np.std(M, axis=axis)
# set bad values to x
stdM[stdM == 0] = x
stdM[np.isnan(stdM)] = x
tt = round( quantile*100 )
threshold_std = np.percentile( stdM, tt )
# threshold_std = quantile( stdM, quantile )
return { 'tSTD': stdM, 'threshold_std': threshold_std}
if mask is None:
temp = utils.slice_image( boldImage, axis=boldImage.dimension-1, idx=0 )
mask = utils.get_mask( temp )
imagematrix = core.timeseries_to_matrix( boldImage, mask )
temp = compute_tSTD( imagematrix, quantile, 0 )
tsnrmask = core.make_image( mask, temp['tSTD'] )
tsnrmask = utils.threshold_image( tsnrmask, temp['threshold_std'], temp['tSTD'].max() )
M = core.timeseries_to_matrix( boldImage, tsnrmask )
components = None
basis = np.array([])
if filter_type in ('polynomial', False):
M, basis = regress_poly(degree, M)
# M = M / compute_tSTD(M, 1.)['tSTD']
# "The covariance matrix C = MMT was constructed and decomposed into its
# principal components using a singular value decomposition."
u, _, _ = linalg.svd(M, full_matrices=False)
if components is None:
components = u[:, :ncompcor]
else:
components = np.hstack((components, u[:, :ncompcor]))
if components is None and ncompcor > 0:
raise ValueError('No components found')
return { 'components': components, 'basis': basis } | python | def compcor( boldImage, ncompcor=4, quantile=0.975, mask=None, filter_type=False, degree=2 ):
def compute_tSTD(M, quantile, x=0, axis=0):
stdM = np.std(M, axis=axis)
# set bad values to x
stdM[stdM == 0] = x
stdM[np.isnan(stdM)] = x
tt = round( quantile*100 )
threshold_std = np.percentile( stdM, tt )
# threshold_std = quantile( stdM, quantile )
return { 'tSTD': stdM, 'threshold_std': threshold_std}
if mask is None:
temp = utils.slice_image( boldImage, axis=boldImage.dimension-1, idx=0 )
mask = utils.get_mask( temp )
imagematrix = core.timeseries_to_matrix( boldImage, mask )
temp = compute_tSTD( imagematrix, quantile, 0 )
tsnrmask = core.make_image( mask, temp['tSTD'] )
tsnrmask = utils.threshold_image( tsnrmask, temp['threshold_std'], temp['tSTD'].max() )
M = core.timeseries_to_matrix( boldImage, tsnrmask )
components = None
basis = np.array([])
if filter_type in ('polynomial', False):
M, basis = regress_poly(degree, M)
# M = M / compute_tSTD(M, 1.)['tSTD']
# "The covariance matrix C = MMT was constructed and decomposed into its
# principal components using a singular value decomposition."
u, _, _ = linalg.svd(M, full_matrices=False)
if components is None:
components = u[:, :ncompcor]
else:
components = np.hstack((components, u[:, :ncompcor]))
if components is None and ncompcor > 0:
raise ValueError('No components found')
return { 'components': components, 'basis': basis } | [
"def",
"compcor",
"(",
"boldImage",
",",
"ncompcor",
"=",
"4",
",",
"quantile",
"=",
"0.975",
",",
"mask",
"=",
"None",
",",
"filter_type",
"=",
"False",
",",
"degree",
"=",
"2",
")",
":",
"def",
"compute_tSTD",
"(",
"M",
",",
"quantile",
",",
"x",
"=",
"0",
",",
"axis",
"=",
"0",
")",
":",
"stdM",
"=",
"np",
".",
"std",
"(",
"M",
",",
"axis",
"=",
"axis",
")",
"# set bad values to x",
"stdM",
"[",
"stdM",
"==",
"0",
"]",
"=",
"x",
"stdM",
"[",
"np",
".",
"isnan",
"(",
"stdM",
")",
"]",
"=",
"x",
"tt",
"=",
"round",
"(",
"quantile",
"*",
"100",
")",
"threshold_std",
"=",
"np",
".",
"percentile",
"(",
"stdM",
",",
"tt",
")",
"# threshold_std = quantile( stdM, quantile )",
"return",
"{",
"'tSTD'",
":",
"stdM",
",",
"'threshold_std'",
":",
"threshold_std",
"}",
"if",
"mask",
"is",
"None",
":",
"temp",
"=",
"utils",
".",
"slice_image",
"(",
"boldImage",
",",
"axis",
"=",
"boldImage",
".",
"dimension",
"-",
"1",
",",
"idx",
"=",
"0",
")",
"mask",
"=",
"utils",
".",
"get_mask",
"(",
"temp",
")",
"imagematrix",
"=",
"core",
".",
"timeseries_to_matrix",
"(",
"boldImage",
",",
"mask",
")",
"temp",
"=",
"compute_tSTD",
"(",
"imagematrix",
",",
"quantile",
",",
"0",
")",
"tsnrmask",
"=",
"core",
".",
"make_image",
"(",
"mask",
",",
"temp",
"[",
"'tSTD'",
"]",
")",
"tsnrmask",
"=",
"utils",
".",
"threshold_image",
"(",
"tsnrmask",
",",
"temp",
"[",
"'threshold_std'",
"]",
",",
"temp",
"[",
"'tSTD'",
"]",
".",
"max",
"(",
")",
")",
"M",
"=",
"core",
".",
"timeseries_to_matrix",
"(",
"boldImage",
",",
"tsnrmask",
")",
"components",
"=",
"None",
"basis",
"=",
"np",
".",
"array",
"(",
"[",
"]",
")",
"if",
"filter_type",
"in",
"(",
"'polynomial'",
",",
"False",
")",
":",
"M",
",",
"basis",
"=",
"regress_poly",
"(",
"degree",
",",
"M",
")",
"# M = M / compute_tSTD(M, 1.)['tSTD']",
"# \"The covariance matrix C = MMT was constructed and decomposed into its",
"# principal components using a singular value decomposition.\"",
"u",
",",
"_",
",",
"_",
"=",
"linalg",
".",
"svd",
"(",
"M",
",",
"full_matrices",
"=",
"False",
")",
"if",
"components",
"is",
"None",
":",
"components",
"=",
"u",
"[",
":",
",",
":",
"ncompcor",
"]",
"else",
":",
"components",
"=",
"np",
".",
"hstack",
"(",
"(",
"components",
",",
"u",
"[",
":",
",",
":",
"ncompcor",
"]",
")",
")",
"if",
"components",
"is",
"None",
"and",
"ncompcor",
">",
"0",
":",
"raise",
"ValueError",
"(",
"'No components found'",
")",
"return",
"{",
"'components'",
":",
"components",
",",
"'basis'",
":",
"basis",
"}"
] | Compute noise components from the input image
ANTsR function: `compcor`
this is adapted from nipy code https://github.com/nipy/nipype/blob/e29ac95fc0fc00fedbcaa0adaf29d5878408ca7c/nipype/algorithms/confounds.py
Arguments
---------
boldImage: input time series image
ncompcor: number of noise components to return
quantile: quantile defining high-variance
mask: mask defining brain or specific tissues
filter_type: type off filter to apply to time series before computing
noise components.
'polynomial' - Legendre polynomial basis
False - None (mean-removal only)
degree: order of polynomial used to remove trends from the timeseries
Returns
-------
dictionary containing:
components: a numpy array
basis: a numpy array containing the (non-constant) filter regressors
Example
-------
>>> cc = ants.compcor( ants.image_read(ants.get_ants_data("ch2")) ) | [
"Compute",
"noise",
"components",
"from",
"the",
"input",
"image"
] | 638020af2cdfc5ff4bdb9809ffe67aa505727a3b | https://github.com/ANTsX/ANTsPy/blob/638020af2cdfc5ff4bdb9809ffe67aa505727a3b/ants/utils/quantile.py#L295-L367 |
250,993 | ANTsX/ANTsPy | ants/utils/bias_correction.py | n3_bias_field_correction | def n3_bias_field_correction(image, downsample_factor=3):
"""
N3 Bias Field Correction
ANTsR function: `n3BiasFieldCorrection`
Arguments
---------
image : ANTsImage
image to be bias corrected
downsample_factor : scalar
how much to downsample image before performing bias correction
Returns
-------
ANTsImage
Example
-------
>>> import ants
>>> image = ants.image_read( ants.get_ants_data('r16') )
>>> image_n3 = ants.n3_bias_field_correction(image)
"""
outimage = image.clone()
args = [image.dimension, image, outimage, downsample_factor]
processed_args = pargs._int_antsProcessArguments(args)
libfn = utils.get_lib_fn('N3BiasFieldCorrection')
libfn(processed_args)
return outimage | python | def n3_bias_field_correction(image, downsample_factor=3):
outimage = image.clone()
args = [image.dimension, image, outimage, downsample_factor]
processed_args = pargs._int_antsProcessArguments(args)
libfn = utils.get_lib_fn('N3BiasFieldCorrection')
libfn(processed_args)
return outimage | [
"def",
"n3_bias_field_correction",
"(",
"image",
",",
"downsample_factor",
"=",
"3",
")",
":",
"outimage",
"=",
"image",
".",
"clone",
"(",
")",
"args",
"=",
"[",
"image",
".",
"dimension",
",",
"image",
",",
"outimage",
",",
"downsample_factor",
"]",
"processed_args",
"=",
"pargs",
".",
"_int_antsProcessArguments",
"(",
"args",
")",
"libfn",
"=",
"utils",
".",
"get_lib_fn",
"(",
"'N3BiasFieldCorrection'",
")",
"libfn",
"(",
"processed_args",
")",
"return",
"outimage"
] | N3 Bias Field Correction
ANTsR function: `n3BiasFieldCorrection`
Arguments
---------
image : ANTsImage
image to be bias corrected
downsample_factor : scalar
how much to downsample image before performing bias correction
Returns
-------
ANTsImage
Example
-------
>>> import ants
>>> image = ants.image_read( ants.get_ants_data('r16') )
>>> image_n3 = ants.n3_bias_field_correction(image) | [
"N3",
"Bias",
"Field",
"Correction"
] | 638020af2cdfc5ff4bdb9809ffe67aa505727a3b | https://github.com/ANTsX/ANTsPy/blob/638020af2cdfc5ff4bdb9809ffe67aa505727a3b/ants/utils/bias_correction.py#L17-L46 |
250,994 | ANTsX/ANTsPy | ants/utils/bias_correction.py | n4_bias_field_correction | def n4_bias_field_correction(image, mask=None, shrink_factor=4,
convergence={'iters':[50,50,50,50], 'tol':1e-07},
spline_param=200, verbose=False, weight_mask=None):
"""
N4 Bias Field Correction
ANTsR function: `n4BiasFieldCorrection`
Arguments
---------
image : ANTsImage
image to bias correct
mask : ANTsImage
input mask, if one is not passed one will be made
shrink_factor : scalar
Shrink factor for multi-resolution correction, typically integer less than 4
convergence : dict w/ keys `iters` and `tol`
iters : vector of maximum number of iterations for each shrinkage factor
tol : the convergence tolerance.
spline_param : integer
Parameter controlling number of control points in spline. Either single value, indicating how many control points, or vector with one entry per dimension of image, indicating the spacing in each direction.
verbose : boolean
enables verbose output.
weight_mask : ANTsImage (optional)
antsImage of weight mask
Returns
-------
ANTsImage
Example
-------
>>> image = ants.image_read( ants.get_ants_data('r16') )
>>> image_n4 = ants.n4_bias_field_correction(image)
"""
if image.pixeltype != 'float':
image = image.clone('float')
iters = convergence['iters']
tol = convergence['tol']
if mask is None:
mask = get_mask(image)
N4_CONVERGENCE_1 = '[%s, %.10f]' % ('x'.join([str(it) for it in iters]), tol)
N4_SHRINK_FACTOR_1 = str(shrink_factor)
if (not isinstance(spline_param, (list,tuple))) or (len(spline_param) == 1):
N4_BSPLINE_PARAMS = '[%i]' % spline_param
elif (isinstance(spline_param, (list,tuple))) and (len(spline_param) == image.dimension):
N4_BSPLINE_PARAMS = '[%s]' % ('x'.join([str(sp) for sp in spline_param]))
else:
raise ValueError('Length of splineParam must either be 1 or dimensionality of image')
if weight_mask is not None:
if not isinstance(weight_mask, iio.ANTsImage):
raise ValueError('Weight Image must be an antsImage')
outimage = image.clone()
kwargs = {
'd': outimage.dimension,
'i': image,
'w': weight_mask,
's': N4_SHRINK_FACTOR_1,
'c': N4_CONVERGENCE_1,
'b': N4_BSPLINE_PARAMS,
'x': mask,
'o': outimage,
'v': int(verbose)
}
processed_args = pargs._int_antsProcessArguments(kwargs)
libfn = utils.get_lib_fn('N4BiasFieldCorrection')
libfn(processed_args)
return outimage | python | def n4_bias_field_correction(image, mask=None, shrink_factor=4,
convergence={'iters':[50,50,50,50], 'tol':1e-07},
spline_param=200, verbose=False, weight_mask=None):
if image.pixeltype != 'float':
image = image.clone('float')
iters = convergence['iters']
tol = convergence['tol']
if mask is None:
mask = get_mask(image)
N4_CONVERGENCE_1 = '[%s, %.10f]' % ('x'.join([str(it) for it in iters]), tol)
N4_SHRINK_FACTOR_1 = str(shrink_factor)
if (not isinstance(spline_param, (list,tuple))) or (len(spline_param) == 1):
N4_BSPLINE_PARAMS = '[%i]' % spline_param
elif (isinstance(spline_param, (list,tuple))) and (len(spline_param) == image.dimension):
N4_BSPLINE_PARAMS = '[%s]' % ('x'.join([str(sp) for sp in spline_param]))
else:
raise ValueError('Length of splineParam must either be 1 or dimensionality of image')
if weight_mask is not None:
if not isinstance(weight_mask, iio.ANTsImage):
raise ValueError('Weight Image must be an antsImage')
outimage = image.clone()
kwargs = {
'd': outimage.dimension,
'i': image,
'w': weight_mask,
's': N4_SHRINK_FACTOR_1,
'c': N4_CONVERGENCE_1,
'b': N4_BSPLINE_PARAMS,
'x': mask,
'o': outimage,
'v': int(verbose)
}
processed_args = pargs._int_antsProcessArguments(kwargs)
libfn = utils.get_lib_fn('N4BiasFieldCorrection')
libfn(processed_args)
return outimage | [
"def",
"n4_bias_field_correction",
"(",
"image",
",",
"mask",
"=",
"None",
",",
"shrink_factor",
"=",
"4",
",",
"convergence",
"=",
"{",
"'iters'",
":",
"[",
"50",
",",
"50",
",",
"50",
",",
"50",
"]",
",",
"'tol'",
":",
"1e-07",
"}",
",",
"spline_param",
"=",
"200",
",",
"verbose",
"=",
"False",
",",
"weight_mask",
"=",
"None",
")",
":",
"if",
"image",
".",
"pixeltype",
"!=",
"'float'",
":",
"image",
"=",
"image",
".",
"clone",
"(",
"'float'",
")",
"iters",
"=",
"convergence",
"[",
"'iters'",
"]",
"tol",
"=",
"convergence",
"[",
"'tol'",
"]",
"if",
"mask",
"is",
"None",
":",
"mask",
"=",
"get_mask",
"(",
"image",
")",
"N4_CONVERGENCE_1",
"=",
"'[%s, %.10f]'",
"%",
"(",
"'x'",
".",
"join",
"(",
"[",
"str",
"(",
"it",
")",
"for",
"it",
"in",
"iters",
"]",
")",
",",
"tol",
")",
"N4_SHRINK_FACTOR_1",
"=",
"str",
"(",
"shrink_factor",
")",
"if",
"(",
"not",
"isinstance",
"(",
"spline_param",
",",
"(",
"list",
",",
"tuple",
")",
")",
")",
"or",
"(",
"len",
"(",
"spline_param",
")",
"==",
"1",
")",
":",
"N4_BSPLINE_PARAMS",
"=",
"'[%i]'",
"%",
"spline_param",
"elif",
"(",
"isinstance",
"(",
"spline_param",
",",
"(",
"list",
",",
"tuple",
")",
")",
")",
"and",
"(",
"len",
"(",
"spline_param",
")",
"==",
"image",
".",
"dimension",
")",
":",
"N4_BSPLINE_PARAMS",
"=",
"'[%s]'",
"%",
"(",
"'x'",
".",
"join",
"(",
"[",
"str",
"(",
"sp",
")",
"for",
"sp",
"in",
"spline_param",
"]",
")",
")",
"else",
":",
"raise",
"ValueError",
"(",
"'Length of splineParam must either be 1 or dimensionality of image'",
")",
"if",
"weight_mask",
"is",
"not",
"None",
":",
"if",
"not",
"isinstance",
"(",
"weight_mask",
",",
"iio",
".",
"ANTsImage",
")",
":",
"raise",
"ValueError",
"(",
"'Weight Image must be an antsImage'",
")",
"outimage",
"=",
"image",
".",
"clone",
"(",
")",
"kwargs",
"=",
"{",
"'d'",
":",
"outimage",
".",
"dimension",
",",
"'i'",
":",
"image",
",",
"'w'",
":",
"weight_mask",
",",
"'s'",
":",
"N4_SHRINK_FACTOR_1",
",",
"'c'",
":",
"N4_CONVERGENCE_1",
",",
"'b'",
":",
"N4_BSPLINE_PARAMS",
",",
"'x'",
":",
"mask",
",",
"'o'",
":",
"outimage",
",",
"'v'",
":",
"int",
"(",
"verbose",
")",
"}",
"processed_args",
"=",
"pargs",
".",
"_int_antsProcessArguments",
"(",
"kwargs",
")",
"libfn",
"=",
"utils",
".",
"get_lib_fn",
"(",
"'N4BiasFieldCorrection'",
")",
"libfn",
"(",
"processed_args",
")",
"return",
"outimage"
] | N4 Bias Field Correction
ANTsR function: `n4BiasFieldCorrection`
Arguments
---------
image : ANTsImage
image to bias correct
mask : ANTsImage
input mask, if one is not passed one will be made
shrink_factor : scalar
Shrink factor for multi-resolution correction, typically integer less than 4
convergence : dict w/ keys `iters` and `tol`
iters : vector of maximum number of iterations for each shrinkage factor
tol : the convergence tolerance.
spline_param : integer
Parameter controlling number of control points in spline. Either single value, indicating how many control points, or vector with one entry per dimension of image, indicating the spacing in each direction.
verbose : boolean
enables verbose output.
weight_mask : ANTsImage (optional)
antsImage of weight mask
Returns
-------
ANTsImage
Example
-------
>>> image = ants.image_read( ants.get_ants_data('r16') )
>>> image_n4 = ants.n4_bias_field_correction(image) | [
"N4",
"Bias",
"Field",
"Correction"
] | 638020af2cdfc5ff4bdb9809ffe67aa505727a3b | https://github.com/ANTsX/ANTsPy/blob/638020af2cdfc5ff4bdb9809ffe67aa505727a3b/ants/utils/bias_correction.py#L49-L126 |
250,995 | ANTsX/ANTsPy | ants/utils/bias_correction.py | abp_n4 | def abp_n4(image, intensity_truncation=(0.025,0.975,256), mask=None, usen3=False):
"""
Truncate outlier intensities and bias correct with the N4 algorithm.
ANTsR function: `abpN4`
Arguments
---------
image : ANTsImage
image to correct and truncate
intensity_truncation : 3-tuple
quantiles for intensity truncation
mask : ANTsImage (optional)
mask for bias correction
usen3 : boolean
if True, use N3 bias correction instead of N4
Returns
-------
ANTsImage
Example
-------
>>> import ants
>>> image = ants.image_read(ants.get_ants_data('r16'))
>>> image2 = ants.abp_n4(image)
"""
if (not isinstance(intensity_truncation, (list,tuple))) or (len(intensity_truncation) != 3):
raise ValueError('intensity_truncation must be list/tuple with 3 values')
outimage = iMath(image, 'TruncateIntensity',
intensity_truncation[0], intensity_truncation[1], intensity_truncation[2])
if usen3 == True:
outimage = n3_bias_field_correction(outimage, 4)
outimage = n3_bias_field_correction(outimage, 2)
return outimage
else:
outimage = n4_bias_field_correction(outimage, mask)
return outimage | python | def abp_n4(image, intensity_truncation=(0.025,0.975,256), mask=None, usen3=False):
if (not isinstance(intensity_truncation, (list,tuple))) or (len(intensity_truncation) != 3):
raise ValueError('intensity_truncation must be list/tuple with 3 values')
outimage = iMath(image, 'TruncateIntensity',
intensity_truncation[0], intensity_truncation[1], intensity_truncation[2])
if usen3 == True:
outimage = n3_bias_field_correction(outimage, 4)
outimage = n3_bias_field_correction(outimage, 2)
return outimage
else:
outimage = n4_bias_field_correction(outimage, mask)
return outimage | [
"def",
"abp_n4",
"(",
"image",
",",
"intensity_truncation",
"=",
"(",
"0.025",
",",
"0.975",
",",
"256",
")",
",",
"mask",
"=",
"None",
",",
"usen3",
"=",
"False",
")",
":",
"if",
"(",
"not",
"isinstance",
"(",
"intensity_truncation",
",",
"(",
"list",
",",
"tuple",
")",
")",
")",
"or",
"(",
"len",
"(",
"intensity_truncation",
")",
"!=",
"3",
")",
":",
"raise",
"ValueError",
"(",
"'intensity_truncation must be list/tuple with 3 values'",
")",
"outimage",
"=",
"iMath",
"(",
"image",
",",
"'TruncateIntensity'",
",",
"intensity_truncation",
"[",
"0",
"]",
",",
"intensity_truncation",
"[",
"1",
"]",
",",
"intensity_truncation",
"[",
"2",
"]",
")",
"if",
"usen3",
"==",
"True",
":",
"outimage",
"=",
"n3_bias_field_correction",
"(",
"outimage",
",",
"4",
")",
"outimage",
"=",
"n3_bias_field_correction",
"(",
"outimage",
",",
"2",
")",
"return",
"outimage",
"else",
":",
"outimage",
"=",
"n4_bias_field_correction",
"(",
"outimage",
",",
"mask",
")",
"return",
"outimage"
] | Truncate outlier intensities and bias correct with the N4 algorithm.
ANTsR function: `abpN4`
Arguments
---------
image : ANTsImage
image to correct and truncate
intensity_truncation : 3-tuple
quantiles for intensity truncation
mask : ANTsImage (optional)
mask for bias correction
usen3 : boolean
if True, use N3 bias correction instead of N4
Returns
-------
ANTsImage
Example
-------
>>> import ants
>>> image = ants.image_read(ants.get_ants_data('r16'))
>>> image2 = ants.abp_n4(image) | [
"Truncate",
"outlier",
"intensities",
"and",
"bias",
"correct",
"with",
"the",
"N4",
"algorithm",
"."
] | 638020af2cdfc5ff4bdb9809ffe67aa505727a3b | https://github.com/ANTsX/ANTsPy/blob/638020af2cdfc5ff4bdb9809ffe67aa505727a3b/ants/utils/bias_correction.py#L129-L169 |
250,996 | ANTsX/ANTsPy | ants/registration/metrics.py | image_mutual_information | def image_mutual_information(image1, image2):
"""
Compute mutual information between two ANTsImage types
ANTsR function: `antsImageMutualInformation`
Arguments
---------
image1 : ANTsImage
image 1
image2 : ANTsImage
image 2
Returns
-------
scalar
Example
-------
>>> import ants
>>> fi = ants.image_read( ants.get_ants_data('r16') ).clone('float')
>>> mi = ants.image_read( ants.get_ants_data('r64') ).clone('float')
>>> mival = ants.image_mutual_information(fi, mi) # -0.1796141
"""
if (image1.pixeltype != 'float') or (image2.pixeltype != 'float'):
raise ValueError('Both images must have float pixeltype')
if image1.dimension != image2.dimension:
raise ValueError('Both images must have same dimension')
libfn = utils.get_lib_fn('antsImageMutualInformation%iD' % image1.dimension)
return libfn(image1.pointer, image2.pointer) | python | def image_mutual_information(image1, image2):
if (image1.pixeltype != 'float') or (image2.pixeltype != 'float'):
raise ValueError('Both images must have float pixeltype')
if image1.dimension != image2.dimension:
raise ValueError('Both images must have same dimension')
libfn = utils.get_lib_fn('antsImageMutualInformation%iD' % image1.dimension)
return libfn(image1.pointer, image2.pointer) | [
"def",
"image_mutual_information",
"(",
"image1",
",",
"image2",
")",
":",
"if",
"(",
"image1",
".",
"pixeltype",
"!=",
"'float'",
")",
"or",
"(",
"image2",
".",
"pixeltype",
"!=",
"'float'",
")",
":",
"raise",
"ValueError",
"(",
"'Both images must have float pixeltype'",
")",
"if",
"image1",
".",
"dimension",
"!=",
"image2",
".",
"dimension",
":",
"raise",
"ValueError",
"(",
"'Both images must have same dimension'",
")",
"libfn",
"=",
"utils",
".",
"get_lib_fn",
"(",
"'antsImageMutualInformation%iD'",
"%",
"image1",
".",
"dimension",
")",
"return",
"libfn",
"(",
"image1",
".",
"pointer",
",",
"image2",
".",
"pointer",
")"
] | Compute mutual information between two ANTsImage types
ANTsR function: `antsImageMutualInformation`
Arguments
---------
image1 : ANTsImage
image 1
image2 : ANTsImage
image 2
Returns
-------
scalar
Example
-------
>>> import ants
>>> fi = ants.image_read( ants.get_ants_data('r16') ).clone('float')
>>> mi = ants.image_read( ants.get_ants_data('r64') ).clone('float')
>>> mival = ants.image_mutual_information(fi, mi) # -0.1796141 | [
"Compute",
"mutual",
"information",
"between",
"two",
"ANTsImage",
"types"
] | 638020af2cdfc5ff4bdb9809ffe67aa505727a3b | https://github.com/ANTsX/ANTsPy/blob/638020af2cdfc5ff4bdb9809ffe67aa505727a3b/ants/registration/metrics.py#L10-L42 |
250,997 | ANTsX/ANTsPy | ants/utils/get_mask.py | get_mask | def get_mask(image, low_thresh=None, high_thresh=None, cleanup=2):
"""
Get a binary mask image from the given image after thresholding
ANTsR function: `getMask`
Arguments
---------
image : ANTsImage
image from which mask will be computed. Can be an antsImage of 2, 3 or 4 dimensions.
low_thresh : scalar (optional)
An inclusive lower threshold for voxels to be included in the mask.
If not given, defaults to image mean.
high_thresh : scalar (optional)
An inclusive upper threshold for voxels to be included in the mask.
If not given, defaults to image max
cleanup : integer
If > 0, morphological operations will be applied to clean up the mask by eroding away small or weakly-connected areas, and closing holes.
If cleanup is >0, the following steps are applied
1. Erosion with radius 2 voxels
2. Retain largest component
3. Dilation with radius 1 voxel
4. Morphological closing
Returns
-------
ANTsImage
Example
-------
>>> import ants
>>> image = ants.image_read( ants.get_ants_data('r16') )
>>> mask = ants.get_mask(image)
"""
cleanup = int(cleanup)
if isinstance(image, iio.ANTsImage):
if image.pixeltype != 'float':
image = image.clone('float')
if low_thresh is None:
low_thresh = image.mean()
if high_thresh is None:
high_thresh = image.max()
mask_image = threshold_image(image, low_thresh, high_thresh)
if cleanup > 0:
mask_image = iMath(mask_image, 'ME', cleanup)
mask_image = iMath(mask_image, 'GetLargestComponent')
mask_image = iMath(mask_image, 'MD', cleanup)
mask_image = iMath(mask_image, 'FillHoles').threshold_image( 1, 2 )
while ((mask_image.min() == mask_image.max()) and (cleanup > 0)):
cleanup = cleanup - 1
mask_image = threshold_image(image, low_thresh, high_thresh)
if cleanup > 0:
mask_image = iMath(mask_image, 'ME', cleanup)
mask_image = iMath(mask_image, 'MD', cleanup)
mask_image = iMath(mask_image, 'FillHoles').threshold_image( 1, 2 )
#if cleanup == 0:
# clustlab = label_clusters(mask_image, 1)
# mask_image = threshold_image(clustlab, 1, 1)
return mask_image | python | def get_mask(image, low_thresh=None, high_thresh=None, cleanup=2):
cleanup = int(cleanup)
if isinstance(image, iio.ANTsImage):
if image.pixeltype != 'float':
image = image.clone('float')
if low_thresh is None:
low_thresh = image.mean()
if high_thresh is None:
high_thresh = image.max()
mask_image = threshold_image(image, low_thresh, high_thresh)
if cleanup > 0:
mask_image = iMath(mask_image, 'ME', cleanup)
mask_image = iMath(mask_image, 'GetLargestComponent')
mask_image = iMath(mask_image, 'MD', cleanup)
mask_image = iMath(mask_image, 'FillHoles').threshold_image( 1, 2 )
while ((mask_image.min() == mask_image.max()) and (cleanup > 0)):
cleanup = cleanup - 1
mask_image = threshold_image(image, low_thresh, high_thresh)
if cleanup > 0:
mask_image = iMath(mask_image, 'ME', cleanup)
mask_image = iMath(mask_image, 'MD', cleanup)
mask_image = iMath(mask_image, 'FillHoles').threshold_image( 1, 2 )
#if cleanup == 0:
# clustlab = label_clusters(mask_image, 1)
# mask_image = threshold_image(clustlab, 1, 1)
return mask_image | [
"def",
"get_mask",
"(",
"image",
",",
"low_thresh",
"=",
"None",
",",
"high_thresh",
"=",
"None",
",",
"cleanup",
"=",
"2",
")",
":",
"cleanup",
"=",
"int",
"(",
"cleanup",
")",
"if",
"isinstance",
"(",
"image",
",",
"iio",
".",
"ANTsImage",
")",
":",
"if",
"image",
".",
"pixeltype",
"!=",
"'float'",
":",
"image",
"=",
"image",
".",
"clone",
"(",
"'float'",
")",
"if",
"low_thresh",
"is",
"None",
":",
"low_thresh",
"=",
"image",
".",
"mean",
"(",
")",
"if",
"high_thresh",
"is",
"None",
":",
"high_thresh",
"=",
"image",
".",
"max",
"(",
")",
"mask_image",
"=",
"threshold_image",
"(",
"image",
",",
"low_thresh",
",",
"high_thresh",
")",
"if",
"cleanup",
">",
"0",
":",
"mask_image",
"=",
"iMath",
"(",
"mask_image",
",",
"'ME'",
",",
"cleanup",
")",
"mask_image",
"=",
"iMath",
"(",
"mask_image",
",",
"'GetLargestComponent'",
")",
"mask_image",
"=",
"iMath",
"(",
"mask_image",
",",
"'MD'",
",",
"cleanup",
")",
"mask_image",
"=",
"iMath",
"(",
"mask_image",
",",
"'FillHoles'",
")",
".",
"threshold_image",
"(",
"1",
",",
"2",
")",
"while",
"(",
"(",
"mask_image",
".",
"min",
"(",
")",
"==",
"mask_image",
".",
"max",
"(",
")",
")",
"and",
"(",
"cleanup",
">",
"0",
")",
")",
":",
"cleanup",
"=",
"cleanup",
"-",
"1",
"mask_image",
"=",
"threshold_image",
"(",
"image",
",",
"low_thresh",
",",
"high_thresh",
")",
"if",
"cleanup",
">",
"0",
":",
"mask_image",
"=",
"iMath",
"(",
"mask_image",
",",
"'ME'",
",",
"cleanup",
")",
"mask_image",
"=",
"iMath",
"(",
"mask_image",
",",
"'MD'",
",",
"cleanup",
")",
"mask_image",
"=",
"iMath",
"(",
"mask_image",
",",
"'FillHoles'",
")",
".",
"threshold_image",
"(",
"1",
",",
"2",
")",
"#if cleanup == 0:",
"# clustlab = label_clusters(mask_image, 1)",
"# mask_image = threshold_image(clustlab, 1, 1)",
"return",
"mask_image"
] | Get a binary mask image from the given image after thresholding
ANTsR function: `getMask`
Arguments
---------
image : ANTsImage
image from which mask will be computed. Can be an antsImage of 2, 3 or 4 dimensions.
low_thresh : scalar (optional)
An inclusive lower threshold for voxels to be included in the mask.
If not given, defaults to image mean.
high_thresh : scalar (optional)
An inclusive upper threshold for voxels to be included in the mask.
If not given, defaults to image max
cleanup : integer
If > 0, morphological operations will be applied to clean up the mask by eroding away small or weakly-connected areas, and closing holes.
If cleanup is >0, the following steps are applied
1. Erosion with radius 2 voxels
2. Retain largest component
3. Dilation with radius 1 voxel
4. Morphological closing
Returns
-------
ANTsImage
Example
-------
>>> import ants
>>> image = ants.image_read( ants.get_ants_data('r16') )
>>> mask = ants.get_mask(image) | [
"Get",
"a",
"binary",
"mask",
"image",
"from",
"the",
"given",
"image",
"after",
"thresholding"
] | 638020af2cdfc5ff4bdb9809ffe67aa505727a3b | https://github.com/ANTsX/ANTsPy/blob/638020af2cdfc5ff4bdb9809ffe67aa505727a3b/ants/utils/get_mask.py#L13-L78 |
250,998 | ANTsX/ANTsPy | ants/utils/label_image_centroids.py | label_image_centroids | def label_image_centroids(image, physical=False, convex=True, verbose=False):
"""
Converts a label image to coordinates summarizing their positions
ANTsR function: `labelImageCentroids`
Arguments
---------
image : ANTsImage
image of integer labels
physical : boolean
whether you want physical space coordinates or not
convex : boolean
if True, return centroid
if False return point with min average distance to other points with same label
Returns
-------
dictionary w/ following key-value pairs:
`labels` : 1D-ndarray
array of label values
`vertices` : pd.DataFrame
coordinates of label centroids
Example
-------
>>> import ants
>>> import numpy as np
>>> image = ants.from_numpy(np.asarray([[[0,2],[1,3]],[[4,6],[5,7]]]).astype('float32'))
>>> labels = ants.label_image_centroids(image)
"""
d = image.shape
if len(d) != 3:
raise ValueError('image must be 3 dimensions')
xcoords = np.asarray(np.arange(d[0]).tolist()*(d[1]*d[2]))
ycoords = np.asarray(np.repeat(np.arange(d[1]),d[0]).tolist()*d[2])
zcoords = np.asarray(np.repeat(np.arange(d[1]), d[0]*d[2]))
labels = image.numpy()
mylabels = np.sort(np.unique(labels[labels > 0])).astype('int')
n_labels = len(mylabels)
xc = np.zeros(n_labels)
yc = np.zeros(n_labels)
zc = np.zeros(n_labels)
if convex:
for i in mylabels:
idx = (labels == i).flatten()
xc[i-1] = np.mean(xcoords[idx])
yc[i-1] = np.mean(ycoords[idx])
zc[i-1] = np.mean(zcoords[idx])
else:
for i in mylabels:
idx = (labels == i).flatten()
xci = xcoords[idx]
yci = ycoords[idx]
zci = zcoords[idx]
dist = np.zeros(len(xci))
for j in range(len(xci)):
dist[j] = np.mean(np.sqrt((xci[j] - xci)**2 + (yci[j] - yci)**2 + (zci[j] - zci)**2))
mid = np.where(dist==np.min(dist))
xc[i-1] = xci[mid]
yc[i-1] = yci[mid]
zc[i-1] = zci[mid]
centroids = np.vstack([xc,yc,zc]).T
#if physical:
# centroids = tio.transform_index_to_physical_point(image, centroids)
return {
'labels': mylabels,
'vertices': centroids
} | python | def label_image_centroids(image, physical=False, convex=True, verbose=False):
d = image.shape
if len(d) != 3:
raise ValueError('image must be 3 dimensions')
xcoords = np.asarray(np.arange(d[0]).tolist()*(d[1]*d[2]))
ycoords = np.asarray(np.repeat(np.arange(d[1]),d[0]).tolist()*d[2])
zcoords = np.asarray(np.repeat(np.arange(d[1]), d[0]*d[2]))
labels = image.numpy()
mylabels = np.sort(np.unique(labels[labels > 0])).astype('int')
n_labels = len(mylabels)
xc = np.zeros(n_labels)
yc = np.zeros(n_labels)
zc = np.zeros(n_labels)
if convex:
for i in mylabels:
idx = (labels == i).flatten()
xc[i-1] = np.mean(xcoords[idx])
yc[i-1] = np.mean(ycoords[idx])
zc[i-1] = np.mean(zcoords[idx])
else:
for i in mylabels:
idx = (labels == i).flatten()
xci = xcoords[idx]
yci = ycoords[idx]
zci = zcoords[idx]
dist = np.zeros(len(xci))
for j in range(len(xci)):
dist[j] = np.mean(np.sqrt((xci[j] - xci)**2 + (yci[j] - yci)**2 + (zci[j] - zci)**2))
mid = np.where(dist==np.min(dist))
xc[i-1] = xci[mid]
yc[i-1] = yci[mid]
zc[i-1] = zci[mid]
centroids = np.vstack([xc,yc,zc]).T
#if physical:
# centroids = tio.transform_index_to_physical_point(image, centroids)
return {
'labels': mylabels,
'vertices': centroids
} | [
"def",
"label_image_centroids",
"(",
"image",
",",
"physical",
"=",
"False",
",",
"convex",
"=",
"True",
",",
"verbose",
"=",
"False",
")",
":",
"d",
"=",
"image",
".",
"shape",
"if",
"len",
"(",
"d",
")",
"!=",
"3",
":",
"raise",
"ValueError",
"(",
"'image must be 3 dimensions'",
")",
"xcoords",
"=",
"np",
".",
"asarray",
"(",
"np",
".",
"arange",
"(",
"d",
"[",
"0",
"]",
")",
".",
"tolist",
"(",
")",
"*",
"(",
"d",
"[",
"1",
"]",
"*",
"d",
"[",
"2",
"]",
")",
")",
"ycoords",
"=",
"np",
".",
"asarray",
"(",
"np",
".",
"repeat",
"(",
"np",
".",
"arange",
"(",
"d",
"[",
"1",
"]",
")",
",",
"d",
"[",
"0",
"]",
")",
".",
"tolist",
"(",
")",
"*",
"d",
"[",
"2",
"]",
")",
"zcoords",
"=",
"np",
".",
"asarray",
"(",
"np",
".",
"repeat",
"(",
"np",
".",
"arange",
"(",
"d",
"[",
"1",
"]",
")",
",",
"d",
"[",
"0",
"]",
"*",
"d",
"[",
"2",
"]",
")",
")",
"labels",
"=",
"image",
".",
"numpy",
"(",
")",
"mylabels",
"=",
"np",
".",
"sort",
"(",
"np",
".",
"unique",
"(",
"labels",
"[",
"labels",
">",
"0",
"]",
")",
")",
".",
"astype",
"(",
"'int'",
")",
"n_labels",
"=",
"len",
"(",
"mylabels",
")",
"xc",
"=",
"np",
".",
"zeros",
"(",
"n_labels",
")",
"yc",
"=",
"np",
".",
"zeros",
"(",
"n_labels",
")",
"zc",
"=",
"np",
".",
"zeros",
"(",
"n_labels",
")",
"if",
"convex",
":",
"for",
"i",
"in",
"mylabels",
":",
"idx",
"=",
"(",
"labels",
"==",
"i",
")",
".",
"flatten",
"(",
")",
"xc",
"[",
"i",
"-",
"1",
"]",
"=",
"np",
".",
"mean",
"(",
"xcoords",
"[",
"idx",
"]",
")",
"yc",
"[",
"i",
"-",
"1",
"]",
"=",
"np",
".",
"mean",
"(",
"ycoords",
"[",
"idx",
"]",
")",
"zc",
"[",
"i",
"-",
"1",
"]",
"=",
"np",
".",
"mean",
"(",
"zcoords",
"[",
"idx",
"]",
")",
"else",
":",
"for",
"i",
"in",
"mylabels",
":",
"idx",
"=",
"(",
"labels",
"==",
"i",
")",
".",
"flatten",
"(",
")",
"xci",
"=",
"xcoords",
"[",
"idx",
"]",
"yci",
"=",
"ycoords",
"[",
"idx",
"]",
"zci",
"=",
"zcoords",
"[",
"idx",
"]",
"dist",
"=",
"np",
".",
"zeros",
"(",
"len",
"(",
"xci",
")",
")",
"for",
"j",
"in",
"range",
"(",
"len",
"(",
"xci",
")",
")",
":",
"dist",
"[",
"j",
"]",
"=",
"np",
".",
"mean",
"(",
"np",
".",
"sqrt",
"(",
"(",
"xci",
"[",
"j",
"]",
"-",
"xci",
")",
"**",
"2",
"+",
"(",
"yci",
"[",
"j",
"]",
"-",
"yci",
")",
"**",
"2",
"+",
"(",
"zci",
"[",
"j",
"]",
"-",
"zci",
")",
"**",
"2",
")",
")",
"mid",
"=",
"np",
".",
"where",
"(",
"dist",
"==",
"np",
".",
"min",
"(",
"dist",
")",
")",
"xc",
"[",
"i",
"-",
"1",
"]",
"=",
"xci",
"[",
"mid",
"]",
"yc",
"[",
"i",
"-",
"1",
"]",
"=",
"yci",
"[",
"mid",
"]",
"zc",
"[",
"i",
"-",
"1",
"]",
"=",
"zci",
"[",
"mid",
"]",
"centroids",
"=",
"np",
".",
"vstack",
"(",
"[",
"xc",
",",
"yc",
",",
"zc",
"]",
")",
".",
"T",
"#if physical:",
"# centroids = tio.transform_index_to_physical_point(image, centroids)",
"return",
"{",
"'labels'",
":",
"mylabels",
",",
"'vertices'",
":",
"centroids",
"}"
] | Converts a label image to coordinates summarizing their positions
ANTsR function: `labelImageCentroids`
Arguments
---------
image : ANTsImage
image of integer labels
physical : boolean
whether you want physical space coordinates or not
convex : boolean
if True, return centroid
if False return point with min average distance to other points with same label
Returns
-------
dictionary w/ following key-value pairs:
`labels` : 1D-ndarray
array of label values
`vertices` : pd.DataFrame
coordinates of label centroids
Example
-------
>>> import ants
>>> import numpy as np
>>> image = ants.from_numpy(np.asarray([[[0,2],[1,3]],[[4,6],[5,7]]]).astype('float32'))
>>> labels = ants.label_image_centroids(image) | [
"Converts",
"a",
"label",
"image",
"to",
"coordinates",
"summarizing",
"their",
"positions"
] | 638020af2cdfc5ff4bdb9809ffe67aa505727a3b | https://github.com/ANTsX/ANTsPy/blob/638020af2cdfc5ff4bdb9809ffe67aa505727a3b/ants/utils/label_image_centroids.py#L10-L89 |
250,999 | ANTsX/ANTsPy | ants/contrib/sampling/transforms.py | MultiResolutionImage.transform | def transform(self, X, y=None):
"""
Generate a set of multi-resolution ANTsImage types
Arguments
---------
X : ANTsImage
image to transform
y : ANTsImage (optional)
another image to transform
Example
-------
>>> import ants
>>> multires = ants.contrib.MultiResolutionImage(levels=4)
>>> img = ants.image_read(ants.get_data('r16'))
>>> imgs = multires.transform(img)
"""
insuffix = X._libsuffix
multires_fn = utils.get_lib_fn('multiResolutionAntsImage%s' % (insuffix))
casted_ptrs = multires_fn(X.pointer, self.levels)
imgs = []
for casted_ptr in casted_ptrs:
img = iio.ANTsImage(pixeltype=X.pixeltype, dimension=X.dimension,
components=X.components, pointer=casted_ptr)
if self.keep_shape:
img = img.resample_image_to_target(X)
imgs.append(img)
return imgs | python | def transform(self, X, y=None):
insuffix = X._libsuffix
multires_fn = utils.get_lib_fn('multiResolutionAntsImage%s' % (insuffix))
casted_ptrs = multires_fn(X.pointer, self.levels)
imgs = []
for casted_ptr in casted_ptrs:
img = iio.ANTsImage(pixeltype=X.pixeltype, dimension=X.dimension,
components=X.components, pointer=casted_ptr)
if self.keep_shape:
img = img.resample_image_to_target(X)
imgs.append(img)
return imgs | [
"def",
"transform",
"(",
"self",
",",
"X",
",",
"y",
"=",
"None",
")",
":",
"insuffix",
"=",
"X",
".",
"_libsuffix",
"multires_fn",
"=",
"utils",
".",
"get_lib_fn",
"(",
"'multiResolutionAntsImage%s'",
"%",
"(",
"insuffix",
")",
")",
"casted_ptrs",
"=",
"multires_fn",
"(",
"X",
".",
"pointer",
",",
"self",
".",
"levels",
")",
"imgs",
"=",
"[",
"]",
"for",
"casted_ptr",
"in",
"casted_ptrs",
":",
"img",
"=",
"iio",
".",
"ANTsImage",
"(",
"pixeltype",
"=",
"X",
".",
"pixeltype",
",",
"dimension",
"=",
"X",
".",
"dimension",
",",
"components",
"=",
"X",
".",
"components",
",",
"pointer",
"=",
"casted_ptr",
")",
"if",
"self",
".",
"keep_shape",
":",
"img",
"=",
"img",
".",
"resample_image_to_target",
"(",
"X",
")",
"imgs",
".",
"append",
"(",
"img",
")",
"return",
"imgs"
] | Generate a set of multi-resolution ANTsImage types
Arguments
---------
X : ANTsImage
image to transform
y : ANTsImage (optional)
another image to transform
Example
-------
>>> import ants
>>> multires = ants.contrib.MultiResolutionImage(levels=4)
>>> img = ants.image_read(ants.get_data('r16'))
>>> imgs = multires.transform(img) | [
"Generate",
"a",
"set",
"of",
"multi",
"-",
"resolution",
"ANTsImage",
"types"
] | 638020af2cdfc5ff4bdb9809ffe67aa505727a3b | https://github.com/ANTsX/ANTsPy/blob/638020af2cdfc5ff4bdb9809ffe67aa505727a3b/ants/contrib/sampling/transforms.py#L60-L91 |
Subsets and Splits