source_code
stringlengths
45
7.24k
function_name
stringlengths
1
111
messages
list
def x() ->List[ServiceState]: keys = DB.get_keys('states*') LOG.debug('Loading list of known services.') services = [] for key in keys: values = key.split(':') if len(values) == 4: services.append(ServiceState(*values[1:4])) return services
get_service_state_list
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x() ->List[ServiceState]:\n keys = DB.get_keys('states*')\n LOG.debug('Loading list of known services.')\n services = []\n for key in keys:\n values = key.split(':')\n if len(values) == 4:\n services.append(ServiceState(*values[1:4]))\n return services\n", "role": "user" }, { "content": "get_service_state_list", "role": "assistant" } ]
def x() ->List[tuple]: keys = DB.get_keys('states*') services = [] for key in keys: values = key.split(':') if len(values) == 4: services.append(':'.join(values[1:])) return services
get_service_id_list
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x() ->List[tuple]:\n keys = DB.get_keys('states*')\n services = []\n for key in keys:\n values = key.split(':')\n if len(values) == 4:\n services.append(':'.join(values[1:]))\n return services\n", "role": "user" }, { "content": "get_service_id_list", "role": "assistant" } ]
def x(): LOG.debug('GET Processing Block list') _url = get_root_url() block_ids = sorted(DB.get_processing_block_ids()) LOG.debug('Processing Block IDs: %s', block_ids) response = dict(num_processing_blocks=len(block_ids), processing_blocks =list()) for block in DB.get_block_details(block_ids): block_id = block['id'] LOG.debug('Creating PB summary for %s', block_id) block['links'] = dict(detail='{}/processing-block/{}'.format(_url, block_id), scheduling_block='{}/scheduling-block/{}'.format( _url, block_id.split(':')[0])) response['processing_blocks'].append(block) response['links'] = {'self': '{}'.format(request.url), 'home': '{}'. format(_url)} return response, HTTPStatus.OK
get
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n LOG.debug('GET Processing Block list')\n _url = get_root_url()\n block_ids = sorted(DB.get_processing_block_ids())\n LOG.debug('Processing Block IDs: %s', block_ids)\n response = dict(num_processing_blocks=len(block_ids), processing_blocks\n =list())\n for block in DB.get_block_details(block_ids):\n block_id = block['id']\n LOG.debug('Creating PB summary for %s', block_id)\n block['links'] = dict(detail='{}/processing-block/{}'.format(_url,\n block_id), scheduling_block='{}/scheduling-block/{}'.format(\n _url, block_id.split(':')[0]))\n response['processing_blocks'].append(block)\n response['links'] = {'self': '{}'.format(request.url), 'home': '{}'.\n format(_url)}\n return response, HTTPStatus.OK\n", "role": "user" }, { "content": "get", "role": "assistant" } ]
def x(self) ->dict: self._check_object_exists() config_dict = DB.get_hash_dict(self.key) for _, value in config_dict.items(): for char in ['[', '{']: if char in value: value = ast.literal_eval(value) return config_dict
config
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self) ->dict:\n self._check_object_exists()\n config_dict = DB.get_hash_dict(self.key)\n for _, value in config_dict.items():\n for char in ['[', '{']:\n if char in value:\n value = ast.literal_eval(value)\n return config_dict\n", "role": "user" }, { "content": "config", "role": "assistant" } ]
def x(self, property_key: str) ->str: self._check_object_exists() return DB.get_hash_value(self.key, property_key)
get_property
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, property_key: str) ->str:\n self._check_object_exists()\n return DB.get_hash_value(self.key, property_key)\n", "role": "user" }, { "content": "get_property", "role": "assistant" } ]
def x(self, value): self._check_object_exists() DB.set_hash_value(self.key, 'status', value) self.publish('status_changed', event_data=dict(status=value))
set_status
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, value):\n self._check_object_exists()\n DB.set_hash_value(self.key, 'status', value)\n self.publish('status_changed', event_data=dict(status=value))\n", "role": "user" }, { "content": "set_status", "role": "assistant" } ]
def x(self, event_type: str, event_data: dict=None): import inspect import os.path _stack = inspect.stack() _origin = os.path.basename(_stack[3][1]) + '::' + _stack[3][3 ] + '::L{}'.format(_stack[3][2]) publish(event_type=event_type, event_data=event_data, object_type=self. _type, object_id=self._id, object_key=self._key, origin=_origin)
publish
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, event_type: str, event_data: dict=None):\n import inspect\n import os.path\n _stack = inspect.stack()\n _origin = os.path.basename(_stack[3][1]) + '::' + _stack[3][3\n ] + '::L{}'.format(_stack[3][2])\n publish(event_type=event_type, event_data=event_data, object_type=self.\n _type, object_id=self._id, object_key=self._key, origin=_origin)\n", "role": "user" }, { "content": "publish", "role": "assistant" } ]
def x(self) ->List[Event]: LOG.debug('Getting events for %s', self.key) return get_events(self.key)
get_events
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self) ->List[Event]:\n LOG.debug('Getting events for %s', self.key)\n return get_events(self.key)\n", "role": "user" }, { "content": "get_events", "role": "assistant" } ]
def x(self): if not DB.get_keys(self.key): raise KeyError("Object with key '{}' not exist".format(self.key))
check_object_exists
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if not DB.get_keys(self.key):\n raise KeyError(\"Object with key '{}' not exist\".format(self.key))\n", "role": "user" }, { "content": "check_object_exists", "role": "assistant" } ]
def x(subsystem: str, name: str, version: str) ->str: return '{}:{}:{}'.format(subsystem, name, version)
get_service_state_object_id
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(subsystem: str, name: str, version: str) ->str:\n return '{}:{}:{}'.format(subsystem, name, version)\n", "role": "user" }, { "content": "get_service_state_object_id", "role": "assistant" } ]
def x(self, compose_str: str) ->list: if not self._manager: raise RuntimeError('Services can only be run on swarm manager nodes') services_ids = [] try: service_config = yaml.load(compose_str) service_list = copy.deepcopy(service_config) service_config.pop('version') service_config.pop('services') for service_name in service_list['services']: service_exist = self._client.services.list(filters={'name': service_name}) if not service_exist: service_config['name'] = service_name service_spec = self._parse_services(service_config, service_name, service_list) created_service = self._client.services.create(**service_spec) service_id = created_service.short_id LOG.debug('Service created: %s', service_id) services_ids.append(service_id) else: LOG.debug('Services already exists') except yaml.YAMLError as exc: print(exc) return services_ids
create_services
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, compose_str: str) ->list:\n if not self._manager:\n raise RuntimeError('Services can only be run on swarm manager nodes')\n services_ids = []\n try:\n service_config = yaml.load(compose_str)\n service_list = copy.deepcopy(service_config)\n service_config.pop('version')\n service_config.pop('services')\n for service_name in service_list['services']:\n service_exist = self._client.services.list(filters={'name':\n service_name})\n if not service_exist:\n service_config['name'] = service_name\n service_spec = self._parse_services(service_config,\n service_name, service_list)\n created_service = self._client.services.create(**service_spec)\n service_id = created_service.short_id\n LOG.debug('Service created: %s', service_id)\n services_ids.append(service_id)\n else:\n LOG.debug('Services already exists')\n except yaml.YAMLError as exc:\n print(exc)\n return services_ids\n", "role": "user" }, { "content": "create_services", "role": "assistant" } ]
def x(self, volume_name: str, driver_spec: str=None): if driver_spec: driver = driver_spec else: driver = 'local' if not self._manager: raise RuntimeError( 'Services can only be deleted on swarm manager nodes') self._client.volumes.create(name=volume_name, driver=driver)
create_volume
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, volume_name: str, driver_spec: str=None):\n if driver_spec:\n driver = driver_spec\n else:\n driver = 'local'\n if not self._manager:\n raise RuntimeError(\n 'Services can only be deleted on swarm manager nodes')\n self._client.volumes.create(name=volume_name, driver=driver)\n", "role": "user" }, { "content": "create_volume", "role": "assistant" } ]
def x(self, service: str): if not self._manager: raise RuntimeError( 'Services can only be deleted on swarm manager nodes') self._api_client.remove_service(service)
delete_service
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, service: str):\n if not self._manager:\n raise RuntimeError(\n 'Services can only be deleted on swarm manager nodes')\n self._api_client.remove_service(service)\n", "role": "user" }, { "content": "delete_service", "role": "assistant" } ]
def x(self): if not self._manager: raise RuntimeError( 'Services can only be deleted on swarm manager nodes') service_list = self.get_service_list() for services in service_list: self._api_client.remove_service(services)
delete_all_services
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if not self._manager:\n raise RuntimeError(\n 'Services can only be deleted on swarm manager nodes')\n service_list = self.get_service_list()\n for services in service_list:\n self._api_client.remove_service(services)\n", "role": "user" }, { "content": "delete_all_services", "role": "assistant" } ]
def x(self, volume_name: str): if not self._manager: raise RuntimeError('Volumes can only be deleted on swarm manager nodes' ) self._api_client.remove_volume(volume_name)
delete_volume
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, volume_name: str):\n if not self._manager:\n raise RuntimeError('Volumes can only be deleted on swarm manager nodes'\n )\n self._api_client.remove_volume(volume_name)\n", "role": "user" }, { "content": "delete_volume", "role": "assistant" } ]
def x(self): if not self._manager: raise RuntimeError('Volumes can only be deleted on swarm manager nodes' ) volume_list = self.get_volume_list() for volumes in volume_list: self._api_client.remove_volume(volumes, force=True)
delete_all_volumes
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if not self._manager:\n raise RuntimeError('Volumes can only be deleted on swarm manager nodes'\n )\n volume_list = self.get_volume_list()\n for volumes in volume_list:\n self._api_client.remove_volume(volumes, force=True)\n", "role": "user" }, { "content": "delete_all_volumes", "role": "assistant" } ]
def x(self) ->list: services = [] if not self._manager: raise RuntimeError( 'Only the Swarm manager node can retrieve all the services.') service_list = self._client.services.list() for s_list in service_list: services.append(s_list.short_id) return services
get_service_list
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self) ->list:\n services = []\n if not self._manager:\n raise RuntimeError(\n 'Only the Swarm manager node can retrieve all the services.')\n service_list = self._client.services.list()\n for s_list in service_list:\n services.append(s_list.short_id)\n return services\n", "role": "user" }, { "content": "get_service_list", "role": "assistant" } ]
def x(self, service_id: str) ->str: if not self._manager: raise RuntimeError( 'Only the Swarm manager node can retrieve all the services details.' ) service = self._client.services.get(service_id) return service.name
get_service_name
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, service_id: str) ->str:\n if not self._manager:\n raise RuntimeError(\n 'Only the Swarm manager node can retrieve all the services details.'\n )\n service = self._client.services.get(service_id)\n return service.name\n", "role": "user" }, { "content": "get_service_name", "role": "assistant" } ]
def x(self, service_id: str) ->dict: if not self._manager: raise RuntimeError( 'Only the Swarm manager node can retrieve all the services details.' ) service = self._client.services.get(service_id) return service.attrs
get_service_details
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, service_id: str) ->dict:\n if not self._manager:\n raise RuntimeError(\n 'Only the Swarm manager node can retrieve all the services details.'\n )\n service = self._client.services.get(service_id)\n return service.attrs\n", "role": "user" }, { "content": "get_service_details", "role": "assistant" } ]
def x(self, service_id: str) ->str: service = self._client.services.get(service_id) for service_task in service.tasks(): service_state = service_task['DesiredState'] return service_state
get_service_state
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, service_id: str) ->str:\n service = self._client.services.get(service_id)\n for service_task in service.tasks():\n service_state = service_task['DesiredState']\n return service_state\n", "role": "user" }, { "content": "get_service_state", "role": "assistant" } ]
def x(self) ->list: nodes = [] if not self._manager: raise RuntimeError( 'Only the Swarm manager node can retrieve all the nodes.') node_list = self._client.nodes.list() for n_list in node_list: nodes.append(n_list.id) return nodes
get_node_list
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self) ->list:\n nodes = []\n if not self._manager:\n raise RuntimeError(\n 'Only the Swarm manager node can retrieve all the nodes.')\n node_list = self._client.nodes.list()\n for n_list in node_list:\n nodes.append(n_list.id)\n return nodes\n", "role": "user" }, { "content": "get_node_list", "role": "assistant" } ]
def x(self, node_id: list) ->dict: if not self._manager: raise RuntimeError( 'Only the Swarm manager node can retrieve node details.') node = self._client.nodes.get(node_id) return node.attrs
get_node_details
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, node_id: list) ->dict:\n if not self._manager:\n raise RuntimeError(\n 'Only the Swarm manager node can retrieve node details.')\n node = self._client.nodes.get(node_id)\n return node.attrs\n", "role": "user" }, { "content": "get_node_details", "role": "assistant" } ]
def x(self) ->list: containers = [] containers_list = self._client.containers.list() for c_list in containers_list: containers.append(c_list.short_id) return containers
get_container_list
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self) ->list:\n containers = []\n containers_list = self._client.containers.list()\n for c_list in containers_list:\n containers.append(c_list.short_id)\n return containers\n", "role": "user" }, { "content": "get_container_list", "role": "assistant" } ]
def x(self, container_id_or_name: str) ->dict: container = self._client.containers.get(container_id_or_name) return container.attrs
get_container_details
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, container_id_or_name: str) ->dict:\n container = self._client.containers.get(container_id_or_name)\n return container.attrs\n", "role": "user" }, { "content": "get_container_details", "role": "assistant" } ]
def x(self) ->list: volumes = [] if not self._manager: raise RuntimeError( 'Only the Swarm manager node can retrieve all the services.') volume_list = self._client.volumes.list() for v_list in volume_list: volumes.append(v_list.name) return volumes
get_volume_list
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self) ->list:\n volumes = []\n if not self._manager:\n raise RuntimeError(\n 'Only the Swarm manager node can retrieve all the services.')\n volume_list = self._client.volumes.list()\n for v_list in volume_list:\n volumes.append(v_list.name)\n return volumes\n", "role": "user" }, { "content": "get_volume_list", "role": "assistant" } ]
def x(self, volume_name: str) ->dict: if volume_name not in self.volumes: raise RuntimeError('No such volume found: ', volume_name) volume = self._client.volumes.get(volume_name) return volume.attrs
get_volume_details
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, volume_name: str) ->dict:\n if volume_name not in self.volumes:\n raise RuntimeError('No such volume found: ', volume_name)\n volume = self._client.volumes.get(volume_name)\n return volume.attrs\n", "role": "user" }, { "content": "get_volume_details", "role": "assistant" } ]
def x(self, service_id: str) ->str: if not self._manager: raise RuntimeError( 'Only the Swarm manager node can retrieve replication level of the service' ) service_details = self.get_service_details(service_id) actual_replica = service_details['Spec']['Mode']['Replicated']['Replicas'] return actual_replica
get_actual_replica
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, service_id: str) ->str:\n if not self._manager:\n raise RuntimeError(\n 'Only the Swarm manager node can retrieve replication level of the service'\n )\n service_details = self.get_service_details(service_id)\n actual_replica = service_details['Spec']['Mode']['Replicated']['Replicas']\n return actual_replica\n", "role": "user" }, { "content": "get_actual_replica", "role": "assistant" } ]
def x(self, service_id: str) ->str: replicas = [] if not self._manager: raise RuntimeError( 'Only the Swarm manager node can retrieve replication level of the service' ) service_tasks = self._client.services.get(service_id).tasks() for task in service_tasks: if task['Status']['State'] == 'running': replicas.append(task) return len(replicas)
get_replicas
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, service_id: str) ->str:\n replicas = []\n if not self._manager:\n raise RuntimeError(\n 'Only the Swarm manager node can retrieve replication level of the service'\n )\n service_tasks = self._client.services.get(service_id).tasks()\n for task in service_tasks:\n if task['Status']['State'] == 'running':\n replicas.append(task)\n return len(replicas)\n", "role": "user" }, { "content": "get_replicas", "role": "assistant" } ]
def x(self, node_name: str, labels: dict): if not self._manager: raise RuntimeError( 'Only the Swarm manager node can update node details.') node_spec = {'Availability': 'active', 'Name': node_name, 'Role': 'manager', 'Labels': labels} node = self._client.nodes.get(node_name) node.update(node_spec)
update_labels
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, node_name: str, labels: dict):\n if not self._manager:\n raise RuntimeError(\n 'Only the Swarm manager node can update node details.')\n node_spec = {'Availability': 'active', 'Name': node_name, 'Role':\n 'manager', 'Labels': labels}\n node = self._client.nodes.get(node_name)\n node.update(node_spec)\n", "role": "user" }, { "content": "update_labels", "role": "assistant" } ]
def x(self, service_config: dict, service_name: str, service_list: dict ) ->dict: for key, value in service_list['services'][service_name].items(): service_config[key] = value if 'command' in key: key = 'args' service_config['args'] = value service_config.pop('command') if 'ports' in key: endpoint_spec = self._parse_ports(value) service_config['endpoint_spec'] = endpoint_spec service_config.pop('ports') if 'volumes' in key: volume_spec = self._parse_volumes(value) service_config['mounts'] = volume_spec service_config.pop('volumes') if 'deploy' in key: self._parse_deploy(value, service_config) service_config.pop('deploy') if 'networks' in key: network_spec = self._parse_networks(service_list) service_config['networks'] = network_spec if 'logging' in key: self._parse_logging(value, service_config) service_config.pop('logging') if 'environment' in key: service_config['env'] = value service_config.pop('environment') return service_config
parse_services
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, service_config: dict, service_name: str, service_list: dict\n ) ->dict:\n for key, value in service_list['services'][service_name].items():\n service_config[key] = value\n if 'command' in key:\n key = 'args'\n service_config['args'] = value\n service_config.pop('command')\n if 'ports' in key:\n endpoint_spec = self._parse_ports(value)\n service_config['endpoint_spec'] = endpoint_spec\n service_config.pop('ports')\n if 'volumes' in key:\n volume_spec = self._parse_volumes(value)\n service_config['mounts'] = volume_spec\n service_config.pop('volumes')\n if 'deploy' in key:\n self._parse_deploy(value, service_config)\n service_config.pop('deploy')\n if 'networks' in key:\n network_spec = self._parse_networks(service_list)\n service_config['networks'] = network_spec\n if 'logging' in key:\n self._parse_logging(value, service_config)\n service_config.pop('logging')\n if 'environment' in key:\n service_config['env'] = value\n service_config.pop('environment')\n return service_config\n", "role": "user" }, { "content": "parse_services", "role": "assistant" } ]
def x(self, deploy_values: dict, service_config: dict): mode = {} for d_value in deploy_values: if 'restart_policy' in d_value: restart_spec = docker.types.RestartPolicy(**deploy_values[d_value]) service_config['restart_policy'] = restart_spec if 'placement' in d_value: for constraints_key, constraints_value in deploy_values[d_value ].items(): service_config[constraints_key] = constraints_value if 'mode' in d_value: mode[d_value] = deploy_values[d_value] if 'replicas' in d_value: mode[d_value] = deploy_values[d_value] if 'resources' in d_value: resource_spec = self._parse_resources(deploy_values, d_value) service_config['resources'] = resource_spec mode_spec = docker.types.ServiceMode(**mode) service_config['mode'] = mode_spec
parse_deploy
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, deploy_values: dict, service_config: dict):\n mode = {}\n for d_value in deploy_values:\n if 'restart_policy' in d_value:\n restart_spec = docker.types.RestartPolicy(**deploy_values[d_value])\n service_config['restart_policy'] = restart_spec\n if 'placement' in d_value:\n for constraints_key, constraints_value in deploy_values[d_value\n ].items():\n service_config[constraints_key] = constraints_value\n if 'mode' in d_value:\n mode[d_value] = deploy_values[d_value]\n if 'replicas' in d_value:\n mode[d_value] = deploy_values[d_value]\n if 'resources' in d_value:\n resource_spec = self._parse_resources(deploy_values, d_value)\n service_config['resources'] = resource_spec\n mode_spec = docker.types.ServiceMode(**mode)\n service_config['mode'] = mode_spec\n", "role": "user" }, { "content": "parse_deploy", "role": "assistant" } ]
def x(port_values: dict) ->dict: endpoints = {} for port_element in port_values: target_port = port_element.split(':') for port in target_port: endpoints[int(port)] = int(port) endpoint_spec = docker.types.EndpointSpec(ports=endpoints) return endpoint_spec
parse_ports
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(port_values: dict) ->dict:\n endpoints = {}\n for port_element in port_values:\n target_port = port_element.split(':')\n for port in target_port:\n endpoints[int(port)] = int(port)\n endpoint_spec = docker.types.EndpointSpec(ports=endpoints)\n return endpoint_spec\n", "role": "user" }, { "content": "parse_ports", "role": "assistant" } ]
def x(volume_values: dict) ->str: for v_values in volume_values: for v_key, v_value in v_values.items(): if v_key == 'source': if v_value == '.': source = os.path.dirname(os.path.abspath(__file__)) else: source = v_value if v_key == 'target': target = v_value volume_spec = [source + ':' + target] return volume_spec
parse_volumes
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(volume_values: dict) ->str:\n for v_values in volume_values:\n for v_key, v_value in v_values.items():\n if v_key == 'source':\n if v_value == '.':\n source = os.path.dirname(os.path.abspath(__file__))\n else:\n source = v_value\n if v_key == 'target':\n target = v_value\n volume_spec = [source + ':' + target]\n return volume_spec\n", "role": "user" }, { "content": "parse_volumes", "role": "assistant" } ]
def x(resource_values: dict, resource_name: str) ->dict: resources = {} for r_values in resource_values[resource_name]: if 'limits' in r_values: for r_key, r_value in resource_values[resource_name][r_values ].items(): if 'cpu' in r_key: cpu_value = float(r_value) * 10 ** 9 cpu_key = r_key[:3] + '_limit' resources[cpu_key] = int(cpu_value) if 'mem' in r_key: mem_value = re.sub('M', '', r_value) mem_key = r_key[:3] + '_limit' resources[mem_key] = int(mem_value) * 1048576 resources_spec = docker.types.Resources(**resources) return resources_spec
parse_resources
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(resource_values: dict, resource_name: str) ->dict:\n resources = {}\n for r_values in resource_values[resource_name]:\n if 'limits' in r_values:\n for r_key, r_value in resource_values[resource_name][r_values\n ].items():\n if 'cpu' in r_key:\n cpu_value = float(r_value) * 10 ** 9\n cpu_key = r_key[:3] + '_limit'\n resources[cpu_key] = int(cpu_value)\n if 'mem' in r_key:\n mem_value = re.sub('M', '', r_value)\n mem_key = r_key[:3] + '_limit'\n resources[mem_key] = int(mem_value) * 1048576\n resources_spec = docker.types.Resources(**resources)\n return resources_spec\n", "role": "user" }, { "content": "parse_resources", "role": "assistant" } ]
def x(service_list: dict) ->list: networks = [] for n_values in service_list['networks'].values(): for n_key, n_value in n_values.items(): if 'name' in n_key: networks.append(n_value) return networks
parse_networks
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(service_list: dict) ->list:\n networks = []\n for n_values in service_list['networks'].values():\n for n_key, n_value in n_values.items():\n if 'name' in n_key:\n networks.append(n_value)\n return networks\n", "role": "user" }, { "content": "parse_networks", "role": "assistant" } ]
def x(log_values: dict, service_config: dict): for log_key, log_value in log_values.items(): if 'driver' in log_key: service_config['log_driver'] = log_value if 'options' in log_key: service_config['log_driver_options'] = log_value
parse_logging
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(log_values: dict, service_config: dict):\n for log_key, log_value in log_values.items():\n if 'driver' in log_key:\n service_config['log_driver'] = log_value\n if 'options' in log_key:\n service_config['log_driver_options'] = log_value\n", "role": "user" }, { "content": "parse_logging", "role": "assistant" } ]
def x(): LOG.info('Initialising Processing Block queue.') queue = ProcessingBlockQueue() active_pb_ids = ProcessingBlockList().active LOG.info('Initialising PC PB queue: %s', active_pb_ids) for pb_id in active_pb_ids: pb = ProcessingBlock(pb_id) queue.put(pb.id, pb.priority, pb.type) return queue
init_queue
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n LOG.info('Initialising Processing Block queue.')\n queue = ProcessingBlockQueue()\n active_pb_ids = ProcessingBlockList().active\n LOG.info('Initialising PC PB queue: %s', active_pb_ids)\n for pb_id in active_pb_ids:\n pb = ProcessingBlock(pb_id)\n queue.put(pb.id, pb.priority, pb.type)\n return queue\n", "role": "user" }, { "content": "init_queue", "role": "assistant" } ]
def x(self): LOG.info('Starting to monitor PB events') check_counter = 0 while True: if check_counter == 50: check_counter = 0 LOG.debug('Checking for PB events...') published_events = self._pb_events.get_published_events() for event in published_events: if event.type == 'status_changed': LOG.info('PB status changed event: %s', event.data['status']) if event.data['status'] == 'created': LOG.info( 'Acknowledged PB created event (%s) for %s, [timestamp: %s]' , event.id, event.object_id, event.timestamp) pb = ProcessingBlock(event.object_id) self._queue.put(event.object_id, pb.priority, pb.type) if event.data['status'] == 'completed': LOG.info( 'Acknowledged PB completed event (%s) for %s, [timestamp: %s]' , event.id, event.object_id, event.timestamp) self._num_pbcs -= 1 if self._num_pbcs < 0: self._num_pbcs = 0 time.sleep(0.1) check_counter += 1
monitor_events
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n LOG.info('Starting to monitor PB events')\n check_counter = 0\n while True:\n if check_counter == 50:\n check_counter = 0\n LOG.debug('Checking for PB events...')\n published_events = self._pb_events.get_published_events()\n for event in published_events:\n if event.type == 'status_changed':\n LOG.info('PB status changed event: %s', event.data['status'])\n if event.data['status'] == 'created':\n LOG.info(\n 'Acknowledged PB created event (%s) for %s, [timestamp: %s]'\n , event.id, event.object_id, event.timestamp)\n pb = ProcessingBlock(event.object_id)\n self._queue.put(event.object_id, pb.priority, pb.type)\n if event.data['status'] == 'completed':\n LOG.info(\n 'Acknowledged PB completed event (%s) for %s, [timestamp: %s]'\n , event.id, event.object_id, event.timestamp)\n self._num_pbcs -= 1\n if self._num_pbcs < 0:\n self._num_pbcs = 0\n time.sleep(0.1)\n check_counter += 1\n", "role": "user" }, { "content": "monitor_events", "role": "assistant" } ]
def x(self): LOG.info('Starting Processing Block queue reporter.') while True: LOG.info('PB queue length = %d', len(self._queue)) time.sleep(self._report_interval) if active_count() != 5: LOG.critical( 'Processing Controller not running correctly! (%d/%d threads active)' , active_count(), 5)
processing_controller_status
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n LOG.info('Starting Processing Block queue reporter.')\n while True:\n LOG.info('PB queue length = %d', len(self._queue))\n time.sleep(self._report_interval)\n if active_count() != 5:\n LOG.critical(\n 'Processing Controller not running correctly! (%d/%d threads active)'\n , active_count(), 5)\n", "role": "user" }, { "content": "processing_controller_status", "role": "assistant" } ]
def x(self): LOG.info('Starting to Schedule Processing Blocks.') while True: time.sleep(0.5) if not self._queue: continue if self._num_pbcs >= self._max_pbcs: LOG.warning('Resource limit reached!') continue _inspect = Inspect(app=APP) if self._queue and _inspect.active() is not None: next_pb = self._queue[-1] LOG.info('Considering %s for execution...', next_pb[2]) utc_now = datetime.datetime.utcnow() time_in_queue = utc_now - datetime_from_isoformat(next_pb[4]) if time_in_queue.total_seconds() >= 10: item = self._queue.get() LOG.info('------------------------------------') LOG.info('>>> Executing %s! <<<', item) LOG.info('------------------------------------') execute_processing_block.delay(item) self._num_pbcs += 1 else: LOG.info('Waiting for resources for %s', next_pb[2])
schedule_processing_blocks
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n LOG.info('Starting to Schedule Processing Blocks.')\n while True:\n time.sleep(0.5)\n if not self._queue:\n continue\n if self._num_pbcs >= self._max_pbcs:\n LOG.warning('Resource limit reached!')\n continue\n _inspect = Inspect(app=APP)\n if self._queue and _inspect.active() is not None:\n next_pb = self._queue[-1]\n LOG.info('Considering %s for execution...', next_pb[2])\n utc_now = datetime.datetime.utcnow()\n time_in_queue = utc_now - datetime_from_isoformat(next_pb[4])\n if time_in_queue.total_seconds() >= 10:\n item = self._queue.get()\n LOG.info('------------------------------------')\n LOG.info('>>> Executing %s! <<<', item)\n LOG.info('------------------------------------')\n execute_processing_block.delay(item)\n self._num_pbcs += 1\n else:\n LOG.info('Waiting for resources for %s', next_pb[2])\n", "role": "user" }, { "content": "schedule_processing_blocks", "role": "assistant" } ]
def x(self): LOG.info('Starting to Monitor PBC status.') inspect = celery.current_app.control.inspect() workers = inspect.ping() start_time = time.time() while workers is None: time.sleep(0.1) elapsed = time.time() - start_time if elapsed > 20.0: LOG.warning('PBC not found!') break if workers is not None: for worker in workers: _tasks = inspect.registered_tasks()[worker] LOG.info('Worker: %s tasks:', worker) for task_index, task_name in enumerate(_tasks): LOG.info(' %02d : %s', task_index, task_name) while True: LOG.info('Checking PBC status (%d/%d)', self._num_pbcs, self._max_pbcs) celery_app = celery.current_app inspect = celery_app.control.inspect() workers = inspect.ping() if workers is None: LOG.warning('PBC service not found!') else: LOG.info('PBC state: %s', celery_app.events.State()) _active = inspect.active() _scheduled = inspect.scheduled() for worker in workers: LOG.info(' Worker %s: scheduled: %s, active: %s', worker, _active[worker], _scheduled[worker]) time.sleep(self._report_interval)
monitor_pbc_status
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n LOG.info('Starting to Monitor PBC status.')\n inspect = celery.current_app.control.inspect()\n workers = inspect.ping()\n start_time = time.time()\n while workers is None:\n time.sleep(0.1)\n elapsed = time.time() - start_time\n if elapsed > 20.0:\n LOG.warning('PBC not found!')\n break\n if workers is not None:\n for worker in workers:\n _tasks = inspect.registered_tasks()[worker]\n LOG.info('Worker: %s tasks:', worker)\n for task_index, task_name in enumerate(_tasks):\n LOG.info(' %02d : %s', task_index, task_name)\n while True:\n LOG.info('Checking PBC status (%d/%d)', self._num_pbcs, self._max_pbcs)\n celery_app = celery.current_app\n inspect = celery_app.control.inspect()\n workers = inspect.ping()\n if workers is None:\n LOG.warning('PBC service not found!')\n else:\n LOG.info('PBC state: %s', celery_app.events.State())\n _active = inspect.active()\n _scheduled = inspect.scheduled()\n for worker in workers:\n LOG.info(' Worker %s: scheduled: %s, active: %s', worker,\n _active[worker], _scheduled[worker])\n time.sleep(self._report_interval)\n", "role": "user" }, { "content": "monitor_pbc_status", "role": "assistant" } ]
def x(self): scheduler_threads = [Thread(target=self._monitor_events, daemon=True), Thread(target=self._processing_controller_status, daemon=True), Thread(target=self._schedule_processing_blocks, daemon=True), Thread(target=self._monitor_pbc_status, daemon=True)] for thread in scheduler_threads: thread.start() try: for thread in scheduler_threads: thread.join() except KeyboardInterrupt: LOG.info('Keyboard interrupt!') sys.exit(0) finally: LOG.info('Finally!')
start
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n scheduler_threads = [Thread(target=self._monitor_events, daemon=True),\n Thread(target=self._processing_controller_status, daemon=True),\n Thread(target=self._schedule_processing_blocks, daemon=True),\n Thread(target=self._monitor_pbc_status, daemon=True)]\n for thread in scheduler_threads:\n thread.start()\n try:\n for thread in scheduler_threads:\n thread.join()\n except KeyboardInterrupt:\n LOG.info('Keyboard interrupt!')\n sys.exit(0)\n finally:\n LOG.info('Finally!')\n", "role": "user" }, { "content": "start", "role": "assistant" } ]
def x(): if request.method == 'POST': response = {'message': 'POST Accepted'} logging.info('alarm POSTED!') data = request.data logging.info(data) string = json.dumps(data) producer.send('SIP-alarms', string.encode()) return response return ''
alarm
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n if request.method == 'POST':\n response = {'message': 'POST Accepted'}\n logging.info('alarm POSTED!')\n data = request.data\n logging.info(data)\n string = json.dumps(data)\n producer.send('SIP-alarms', string.encode())\n return response\n return ''\n", "role": "user" }, { "content": "alarm", "role": "assistant" } ]
def x(service: ServiceState): LOG.debug('Setting current state from target state for %s', service.id) service.update_current_state(service.target_state)
update_service_current_state
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(service: ServiceState):\n LOG.debug('Setting current state from target state for %s', service.id)\n service.update_current_state(service.target_state)\n", "role": "user" }, { "content": "update_service_current_state", "role": "assistant" } ]
def x(sdp_target_state: str): service_states = get_service_state_list() for service in service_states: if service.current_state != sdp_target_state: LOG.debug('Setting the current state of %s to be %s', service. id, sdp_target_state) service.update_current_state(sdp_target_state)
update_services_instant_gratification
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(sdp_target_state: str):\n service_states = get_service_state_list()\n for service in service_states:\n if service.current_state != sdp_target_state:\n LOG.debug('Setting the current state of %s to be %s', service.\n id, sdp_target_state)\n service.update_current_state(sdp_target_state)\n", "role": "user" }, { "content": "update_services_instant_gratification", "role": "assistant" } ]
def x(sdp_target_state: str): service_states = get_service_state_list() for service in service_states: if service.current_state != sdp_target_state: LOG.debug('Setting the target state of %s to be %s', service.id, sdp_target_state) service.update_target_state(sdp_target_state)
update_services_target_state
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(sdp_target_state: str):\n service_states = get_service_state_list()\n for service in service_states:\n if service.current_state != sdp_target_state:\n LOG.debug('Setting the target state of %s to be %s', service.id,\n sdp_target_state)\n service.update_target_state(sdp_target_state)\n", "role": "user" }, { "content": "update_services_target_state", "role": "assistant" } ]
def x(sdp_state: SDPState): LOG.info('Handling SDP target state updated event...') LOG.info('SDP target state: %s', sdp_state.target_state) if sdp_state.target_state == 'off': _update_services_target_state('off') sdp_state.update_current_state(sdp_state.target_state)
handle_sdp_target_state_updated
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(sdp_state: SDPState):\n LOG.info('Handling SDP target state updated event...')\n LOG.info('SDP target state: %s', sdp_state.target_state)\n if sdp_state.target_state == 'off':\n _update_services_target_state('off')\n sdp_state.update_current_state(sdp_state.target_state)\n", "role": "user" }, { "content": "handle_sdp_target_state_updated", "role": "assistant" } ]
def x(): parser = argparse.ArgumentParser(description='{} service.'.format( __service_id__)) parser.add_argument('--random_errors', action='store_true', help= 'Enable random errors') parser.add_argument('-v', action='store_true', help= 'Verbose mode (enable debug printing)') parser.add_argument('-vv', action='store_true', help='Extra verbose mode') args = parser.parse_args() if args.vv: init_logger(log_level='DEBUG', show_log_origin=True) elif args.v: init_logger(logger_name='sip.ec.master_controller', log_level='DEBUG') else: init_logger(log_level='INFO') return args
parse_args
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n parser = argparse.ArgumentParser(description='{} service.'.format(\n __service_id__))\n parser.add_argument('--random_errors', action='store_true', help=\n 'Enable random errors')\n parser.add_argument('-v', action='store_true', help=\n 'Verbose mode (enable debug printing)')\n parser.add_argument('-vv', action='store_true', help='Extra verbose mode')\n args = parser.parse_args()\n if args.vv:\n init_logger(log_level='DEBUG', show_log_origin=True)\n elif args.v:\n init_logger(logger_name='sip.ec.master_controller', log_level='DEBUG')\n else:\n init_logger(log_level='INFO')\n return args\n", "role": "user" }, { "content": "parse_args", "role": "assistant" } ]
def x(sdp_state: SDPState): LOG.info('Initialising: %s', __service_id__) _services = ['ExecutionControl:AlarmReceiver:1.0.0', 'ExecutionControl:AlertManager:1.0.0', 'ExecutionControl:ConfigurationDatabase:5.0.1', 'ExecutionControl:MasterController:1.3.0', 'ExecutionControl:ProcessingController:1.2.6', 'ExecutionControl:ProcessingBlockController:1.3.0', 'TangoControl:Database:1.0.4', 'TangoControl:MySQL:1.0.3', 'TangoControl:SDPMaster:1.2.1', 'TangoControl:Subarrays:1.2.0', 'TangoControl:ProcessingBlocks:1.2.0', 'Platform:Kafka:2.1.1', 'Platform:Prometheus:1.0.0', 'Platform:PrometheusPushGateway:0.7.0', 'Platform:RedisCommander:210.0.0', 'Platform:Zookeeper:3.4.13'] for service_id in _services: subsystem, name, version = service_id.split(':') ServiceState(subsystem, name, version) if sdp_state.current_state in ['unknown', 'off']: try: LOG.info("Setting the SDPState to 'init'") sdp_state.update_current_state('init', force=True) except ValueError as error: LOG.critical('Unable to set the State of SDP to init! %s', str( error)) LOG.info('Updating Service States') service_state_list = get_service_state_list() for service_state in service_state_list: if service_state.current_state in ['unknown', 'off']: service_state.update_current_state('init', force=True) for service_state in service_state_list: if service_state.current_state == 'init': time.sleep(random.uniform(0, 0.2)) service_state.update_current_state('on') services_on = [(service.current_state == 'on') for service in service_state_list] if all(services_on): LOG.info('All Services are online!.') sdp_state.update_current_state('standby') else: LOG.critical('Master Controller failed to initialise.') return service_state_list
init
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(sdp_state: SDPState):\n LOG.info('Initialising: %s', __service_id__)\n _services = ['ExecutionControl:AlarmReceiver:1.0.0',\n 'ExecutionControl:AlertManager:1.0.0',\n 'ExecutionControl:ConfigurationDatabase:5.0.1',\n 'ExecutionControl:MasterController:1.3.0',\n 'ExecutionControl:ProcessingController:1.2.6',\n 'ExecutionControl:ProcessingBlockController:1.3.0',\n 'TangoControl:Database:1.0.4', 'TangoControl:MySQL:1.0.3',\n 'TangoControl:SDPMaster:1.2.1', 'TangoControl:Subarrays:1.2.0',\n 'TangoControl:ProcessingBlocks:1.2.0', 'Platform:Kafka:2.1.1',\n 'Platform:Prometheus:1.0.0', 'Platform:PrometheusPushGateway:0.7.0',\n 'Platform:RedisCommander:210.0.0', 'Platform:Zookeeper:3.4.13']\n for service_id in _services:\n subsystem, name, version = service_id.split(':')\n ServiceState(subsystem, name, version)\n if sdp_state.current_state in ['unknown', 'off']:\n try:\n LOG.info(\"Setting the SDPState to 'init'\")\n sdp_state.update_current_state('init', force=True)\n except ValueError as error:\n LOG.critical('Unable to set the State of SDP to init! %s', str(\n error))\n LOG.info('Updating Service States')\n service_state_list = get_service_state_list()\n for service_state in service_state_list:\n if service_state.current_state in ['unknown', 'off']:\n service_state.update_current_state('init', force=True)\n for service_state in service_state_list:\n if service_state.current_state == 'init':\n time.sleep(random.uniform(0, 0.2))\n service_state.update_current_state('on')\n services_on = [(service.current_state == 'on') for service in\n service_state_list]\n if all(services_on):\n LOG.info('All Services are online!.')\n sdp_state.update_current_state('standby')\n else:\n LOG.critical('Master Controller failed to initialise.')\n return service_state_list\n", "role": "user" }, { "content": "init", "role": "assistant" } ]
def x(event: Event, sdp_state: SDPState, service_states: List[ServiceState]): LOG.debug('Event detected! (id : "%s", type: "%s", data: "%s")', event. object_id, event.type, event.data) if event.object_id == 'SDP' and event.type == 'current_state_updated': LOG.info('SDP current state updated, no action required!') if event.object_id == 'SDP' and event.type == 'target_state_updated': LOG.info("SDP target state changed to '%s'", sdp_state.target_state) if sdp_state.target_state == sdp_state.current_state: LOG.warning('SDP already in %s state', sdp_state.current_state) return if not sdp_state.is_target_state_allowed(sdp_state.target_state): LOG.error('Transition to %s is not allowed when in state %s', sdp_state.target_state, sdp_state.current_state) sdp_state.target_state = sdp_state.current_state return _update_services_target_state(sdp_state.target_state) if sdp_state.target_state == 'off': LOG.info('Turning off services!') for service_state in service_states: service_state.update_target_state('off') service_state.update_current_state('off') LOG.info('Processing target state change request ...') time.sleep(0.1) LOG.info('Done processing target state change request!') sdp_state.update_current_state(sdp_state.target_state) if sdp_state.current_state == 'alarm': LOG.debug('raising SDP state alarm') SIP_STATE_ALARM.set(1) else: SIP_STATE_ALARM.set(0) try: push_to_gateway('platform_pushgateway:9091', job='SIP', registry=COLLECTOR_REGISTRY) except urllib.error.URLError: LOG.warning('Unable to connect to the Alarms service!')
process_event
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(event: Event, sdp_state: SDPState, service_states: List[ServiceState]):\n LOG.debug('Event detected! (id : \"%s\", type: \"%s\", data: \"%s\")', event.\n object_id, event.type, event.data)\n if event.object_id == 'SDP' and event.type == 'current_state_updated':\n LOG.info('SDP current state updated, no action required!')\n if event.object_id == 'SDP' and event.type == 'target_state_updated':\n LOG.info(\"SDP target state changed to '%s'\", sdp_state.target_state)\n if sdp_state.target_state == sdp_state.current_state:\n LOG.warning('SDP already in %s state', sdp_state.current_state)\n return\n if not sdp_state.is_target_state_allowed(sdp_state.target_state):\n LOG.error('Transition to %s is not allowed when in state %s',\n sdp_state.target_state, sdp_state.current_state)\n sdp_state.target_state = sdp_state.current_state\n return\n _update_services_target_state(sdp_state.target_state)\n if sdp_state.target_state == 'off':\n LOG.info('Turning off services!')\n for service_state in service_states:\n service_state.update_target_state('off')\n service_state.update_current_state('off')\n LOG.info('Processing target state change request ...')\n time.sleep(0.1)\n LOG.info('Done processing target state change request!')\n sdp_state.update_current_state(sdp_state.target_state)\n if sdp_state.current_state == 'alarm':\n LOG.debug('raising SDP state alarm')\n SIP_STATE_ALARM.set(1)\n else:\n SIP_STATE_ALARM.set(0)\n try:\n push_to_gateway('platform_pushgateway:9091', job='SIP',\n registry=COLLECTOR_REGISTRY)\n except urllib.error.URLError:\n LOG.warning('Unable to connect to the Alarms service!')\n", "role": "user" }, { "content": "process_event", "role": "assistant" } ]
def x(): sdp_state = SDPState() service_states = get_service_state_list() state_events = sdp_state.get_event_queue(subscriber=__service_name__) state_is_off = sdp_state.current_state == 'off' counter = 0 while True: time.sleep(0.1) if not state_is_off: if counter % 1000 == 0: LOG.debug('Checking published events ... %d', counter / 1000) _published_events = state_events.get_published_events(process =True) for _state_event in _published_events: _process_event(_state_event, sdp_state, service_states) else: _state_event = state_events.get() if _state_event: _process_event(_state_event, sdp_state, service_states) state_is_off = sdp_state.current_state == 'off' counter += 1
process_state_change_events
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n sdp_state = SDPState()\n service_states = get_service_state_list()\n state_events = sdp_state.get_event_queue(subscriber=__service_name__)\n state_is_off = sdp_state.current_state == 'off'\n counter = 0\n while True:\n time.sleep(0.1)\n if not state_is_off:\n if counter % 1000 == 0:\n LOG.debug('Checking published events ... %d', counter / 1000)\n _published_events = state_events.get_published_events(process\n =True)\n for _state_event in _published_events:\n _process_event(_state_event, sdp_state, service_states)\n else:\n _state_event = state_events.get()\n if _state_event:\n _process_event(_state_event, sdp_state, service_states)\n state_is_off = sdp_state.current_state == 'off'\n counter += 1\n", "role": "user" }, { "content": "process_state_change_events", "role": "assistant" } ]
def x(): _parse_args() LOG.info('Starting: %s', __service_id__) LOG.info('Subscribing to state change events (subscriber = %s)', __service_name__) sdp_state = SDPState() _ = sdp_state.subscribe(subscriber=__service_name__) _ = _init(sdp_state) LOG.info('Finished initialising!') LOG.info('Responding to state change events ...') try: _process_state_change_events() except ValueError as error: LOG.critical('Value error: %s', str(error)) except KeyboardInterrupt as err: LOG.debug('Keyboard Interrupt %s', err) LOG.info('Exiting!')
main
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n _parse_args()\n LOG.info('Starting: %s', __service_id__)\n LOG.info('Subscribing to state change events (subscriber = %s)',\n __service_name__)\n sdp_state = SDPState()\n _ = sdp_state.subscribe(subscriber=__service_name__)\n _ = _init(sdp_state)\n LOG.info('Finished initialising!')\n LOG.info('Responding to state change events ...')\n try:\n _process_state_change_events()\n except ValueError as error:\n LOG.critical('Value error: %s', str(error))\n except KeyboardInterrupt as err:\n LOG.debug('Keyboard Interrupt %s', err)\n LOG.info('Exiting!')\n", "role": "user" }, { "content": "main", "role": "assistant" } ]
def x(): stream_config = spead2.send.StreamConfig(max_packet_size=16356, rate= 1000000000.0, burst_size=10, max_heaps=1) item_group = spead2.send.ItemGroup(flavour=spead2.Flavour(4, 64, 48, 0)) num_baselines = 512 * 513 // 2 dtype = [('TCI', 'i1'), ('FD', 'u1'), ('VIS', '<c8', 4)] item_group.add_item(id=24576, name='visibility_timestamp_count', description='', shape=tuple(), format=None, dtype='<u4') item_group.add_item(id=24577, name='visibility_timestamp_fraction', description='', shape=tuple(), format=None, dtype='<u4') item_group.add_item(id=24581, name='visibility_baseline_count', description='', shape=tuple(), format=None, dtype='<u4') item_group.add_item(id=24584, name='scan_id', description='', shape= tuple(), format=None, dtype='<u8') item_group.add_item(id=24586, name='correlator_output_data', description='', shape=(num_baselines,), dtype=dtype) streams = [] num_streams = 2 for i in range(num_streams): stream = spead2.send.UdpStream(thread_pool=spead2.ThreadPool( threads=1), hostname='127.0.0.1', port=41000 + i, config= stream_config) stream.send_heap(item_group.get_start()) streams.append(stream) vis = numpy.zeros(shape=(num_baselines,), dtype=dtype) num_heaps = 200 start_time = time.time() for stream in streams: item_group['visibility_timestamp_count'].value = 1 item_group['visibility_timestamp_fraction'].value = 0 item_group['visibility_baseline_count'].value = num_baselines item_group['scan_id'].value = 100000000 item_group['correlator_output_data'].value = vis for i in range(num_heaps): stream.send_heap(item_group.get_heap(descriptors='all', data='all') ) duration = time.time() - start_time data_size = num_streams * num_heaps * (vis.nbytes / 1000000.0) print('Sent %.3f MB in %.3f sec (%.3f MB/sec)' % (data_size, duration, data_size / duration)) for stream in streams: stream.send_heap(item_group.get_end())
main
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n stream_config = spead2.send.StreamConfig(max_packet_size=16356, rate=\n 1000000000.0, burst_size=10, max_heaps=1)\n item_group = spead2.send.ItemGroup(flavour=spead2.Flavour(4, 64, 48, 0))\n num_baselines = 512 * 513 // 2\n dtype = [('TCI', 'i1'), ('FD', 'u1'), ('VIS', '<c8', 4)]\n item_group.add_item(id=24576, name='visibility_timestamp_count',\n description='', shape=tuple(), format=None, dtype='<u4')\n item_group.add_item(id=24577, name='visibility_timestamp_fraction',\n description='', shape=tuple(), format=None, dtype='<u4')\n item_group.add_item(id=24581, name='visibility_baseline_count',\n description='', shape=tuple(), format=None, dtype='<u4')\n item_group.add_item(id=24584, name='scan_id', description='', shape=\n tuple(), format=None, dtype='<u8')\n item_group.add_item(id=24586, name='correlator_output_data',\n description='', shape=(num_baselines,), dtype=dtype)\n streams = []\n num_streams = 2\n for i in range(num_streams):\n stream = spead2.send.UdpStream(thread_pool=spead2.ThreadPool(\n threads=1), hostname='127.0.0.1', port=41000 + i, config=\n stream_config)\n stream.send_heap(item_group.get_start())\n streams.append(stream)\n vis = numpy.zeros(shape=(num_baselines,), dtype=dtype)\n num_heaps = 200\n start_time = time.time()\n for stream in streams:\n item_group['visibility_timestamp_count'].value = 1\n item_group['visibility_timestamp_fraction'].value = 0\n item_group['visibility_baseline_count'].value = num_baselines\n item_group['scan_id'].value = 100000000\n item_group['correlator_output_data'].value = vis\n for i in range(num_heaps):\n stream.send_heap(item_group.get_heap(descriptors='all', data='all')\n )\n duration = time.time() - start_time\n data_size = num_streams * num_heaps * (vis.nbytes / 1000000.0)\n print('Sent %.3f MB in %.3f sec (%.3f MB/sec)' % (data_size, duration, \n data_size / duration))\n for stream in streams:\n stream.send_heap(item_group.get_end())\n", "role": "user" }, { "content": "main", "role": "assistant" } ]
def x(cls, config: dict): timestamp = config.get('timestamp', None) return cls(config.get('id'), config.get('type'), config.get('data', dict()), config.get('origin', None), timestamp, config.get( 'object_type', None), config.get('object_id', None), config.get( 'object_key', None))
from_config
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, config: dict):\n timestamp = config.get('timestamp', None)\n return cls(config.get('id'), config.get('type'), config.get('data',\n dict()), config.get('origin', None), timestamp, config.get(\n 'object_type', None), config.get('object_id', None), config.get(\n 'object_key', None))\n", "role": "user" }, { "content": "from_config", "role": "assistant" } ]
def x(filename, imager, grid_data, grid_norm, grid_weights): ms = oskar.MeasurementSet.open(filename) block_start = 0 num_rows = ms.num_rows num_baselines = ms.num_stations * (ms.num_stations - 1) // 2 while block_start < num_rows: block_size = num_rows - block_start if block_size > num_baselines: block_size = num_baselines uvw = ms.read_column('UVW', block_start, block_size) vis_weights = ms.read_column('WEIGHT', block_start, block_size) if ms.num_pols == 4: vis_weights = 0.5 * (vis_weights[:, 0] + vis_weights[:, 3]) for j in range(ms.num_channels): coords = uvw * (ms.freq_start_hz + j * ms.freq_inc_hz ) / 299792458.0 vis_data = None if not imager.coords_only: vis_data = ms.read_vis(block_start, j, 1, block_size) if ms.num_pols == 4: vis_data = 0.5 * (vis_data[0, :, 0] + vis_data[0, :, 3]) grid_norm = imager.update_plane(coords[:, 0], coords[:, 1], coords[:, 2], vis_data, vis_weights, grid_data, grid_norm, grid_weights) block_start += block_size return grid_norm
process_input_data
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(filename, imager, grid_data, grid_norm, grid_weights):\n ms = oskar.MeasurementSet.open(filename)\n block_start = 0\n num_rows = ms.num_rows\n num_baselines = ms.num_stations * (ms.num_stations - 1) // 2\n while block_start < num_rows:\n block_size = num_rows - block_start\n if block_size > num_baselines:\n block_size = num_baselines\n uvw = ms.read_column('UVW', block_start, block_size)\n vis_weights = ms.read_column('WEIGHT', block_start, block_size)\n if ms.num_pols == 4:\n vis_weights = 0.5 * (vis_weights[:, 0] + vis_weights[:, 3])\n for j in range(ms.num_channels):\n coords = uvw * (ms.freq_start_hz + j * ms.freq_inc_hz\n ) / 299792458.0\n vis_data = None\n if not imager.coords_only:\n vis_data = ms.read_vis(block_start, j, 1, block_size)\n if ms.num_pols == 4:\n vis_data = 0.5 * (vis_data[0, :, 0] + vis_data[0, :, 3])\n grid_norm = imager.update_plane(coords[:, 0], coords[:, 1],\n coords[:, 2], vis_data, vis_weights, grid_data, grid_norm,\n grid_weights)\n block_start += block_size\n return grid_norm\n", "role": "user" }, { "content": "process_input_data", "role": "assistant" } ]
def x(): if len(sys.argv) < 2: raise RuntimeError( 'Usage: mpiexec -n <np> python mpi_imager_test.py <settings_file> <dir>' ) comm = MPI.COMM_WORLD settings = None inputs = None grid_weights = None log = logging.getLogger() log.setLevel(logging.DEBUG) if len(log.handlers) == 0: log.addHandler(logging.StreamHandler(sys.stdout)) if comm.Get_rank() == 0: with open(sys.argv[1]) as f: settings = json.load(f) data_dir = str(sys.argv[2]) inputs = glob(join(data_dir, '*.ms')) + glob(join(data_dir, '*.MS')) inputs = filter(None, inputs) log.info('Found input Measurement Sets: %s', ', '.join(inputs)) inputs = chunks(inputs, comm.Get_size()) comm.barrier() settings = comm.bcast(settings) inputs = comm.scatter(inputs) log.debug('Rank %d, processing [%s]', comm.Get_rank(), ', '.join(inputs)) precision = settings['precision'] imager = oskar.Imager(precision) for key, value in settings['imager'].items(): setattr(imager, key, value) grid_norm = 0.0 grid_dim = [imager.plane_size, imager.plane_size] grid_data = numpy.zeros(grid_dim, dtype='c8' if precision == 'single' else 'c16') if settings['combine']: if imager.weighting == 'Uniform' or imager.algorithm == 'W-projection': local_weights = None if imager.weighting == 'Uniform': grid_weights = numpy.zeros(grid_dim, dtype=precision) local_weights = numpy.zeros(grid_dim, dtype=precision) imager.coords_only = True for f in inputs: log.info('Reading coordinates from %s', f) process_input_data(f, imager, None, 0.0, local_weights) imager.coords_only = False num_w_planes = imager.num_w_planes num_w_planes = comm.allreduce(num_w_planes, op=MPI.MAX) imager.num_w_planes = num_w_planes if local_weights is not None: comm.Allreduce(local_weights, grid_weights, op=MPI.SUM) for f in inputs: log.info('Reading visibilities from %s', f) grid_norm = process_input_data(f, imager, grid_data, grid_norm, grid_weights) grid = numpy.zeros_like(grid_data) comm.Reduce(grid_data, grid, op=MPI.SUM) grid_norm = comm.reduce(grid_norm, op=MPI.SUM) if comm.Get_rank() == 0: save_image(imager, grid, grid_norm, settings['output_file']) log.info('Finished. Output file is %s', settings['output_file']) else: for f in inputs: grid_norm = 0.0 grid_data.fill(0) if imager.weighting == 'Uniform': grid_weights = numpy.zeros(grid_dim, dtype=precision) if (imager.weighting == 'Uniform' or imager.algorithm == 'W-projection'): imager.coords_only = True log.info('Reading coordinates from %s', f) process_input_data(f, imager, None, 0.0, grid_weights) imager.coords_only = False log.info('Reading visibilities from %s', f) grid_norm = process_input_data(f, imager, grid_data, grid_norm, grid_weights) output_file = splitext(f)[0] + '.fits' save_image(imager, grid_data, grid_norm, output_file) log.info('Finished. Output file is %s', output_file)
main
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n if len(sys.argv) < 2:\n raise RuntimeError(\n 'Usage: mpiexec -n <np> python mpi_imager_test.py <settings_file> <dir>'\n )\n comm = MPI.COMM_WORLD\n settings = None\n inputs = None\n grid_weights = None\n log = logging.getLogger()\n log.setLevel(logging.DEBUG)\n if len(log.handlers) == 0:\n log.addHandler(logging.StreamHandler(sys.stdout))\n if comm.Get_rank() == 0:\n with open(sys.argv[1]) as f:\n settings = json.load(f)\n data_dir = str(sys.argv[2])\n inputs = glob(join(data_dir, '*.ms')) + glob(join(data_dir, '*.MS'))\n inputs = filter(None, inputs)\n log.info('Found input Measurement Sets: %s', ', '.join(inputs))\n inputs = chunks(inputs, comm.Get_size())\n comm.barrier()\n settings = comm.bcast(settings)\n inputs = comm.scatter(inputs)\n log.debug('Rank %d, processing [%s]', comm.Get_rank(), ', '.join(inputs))\n precision = settings['precision']\n imager = oskar.Imager(precision)\n for key, value in settings['imager'].items():\n setattr(imager, key, value)\n grid_norm = 0.0\n grid_dim = [imager.plane_size, imager.plane_size]\n grid_data = numpy.zeros(grid_dim, dtype='c8' if precision == 'single' else\n 'c16')\n if settings['combine']:\n if imager.weighting == 'Uniform' or imager.algorithm == 'W-projection':\n local_weights = None\n if imager.weighting == 'Uniform':\n grid_weights = numpy.zeros(grid_dim, dtype=precision)\n local_weights = numpy.zeros(grid_dim, dtype=precision)\n imager.coords_only = True\n for f in inputs:\n log.info('Reading coordinates from %s', f)\n process_input_data(f, imager, None, 0.0, local_weights)\n imager.coords_only = False\n num_w_planes = imager.num_w_planes\n num_w_planes = comm.allreduce(num_w_planes, op=MPI.MAX)\n imager.num_w_planes = num_w_planes\n if local_weights is not None:\n comm.Allreduce(local_weights, grid_weights, op=MPI.SUM)\n for f in inputs:\n log.info('Reading visibilities from %s', f)\n grid_norm = process_input_data(f, imager, grid_data, grid_norm,\n grid_weights)\n grid = numpy.zeros_like(grid_data)\n comm.Reduce(grid_data, grid, op=MPI.SUM)\n grid_norm = comm.reduce(grid_norm, op=MPI.SUM)\n if comm.Get_rank() == 0:\n save_image(imager, grid, grid_norm, settings['output_file'])\n log.info('Finished. Output file is %s', settings['output_file'])\n else:\n for f in inputs:\n grid_norm = 0.0\n grid_data.fill(0)\n if imager.weighting == 'Uniform':\n grid_weights = numpy.zeros(grid_dim, dtype=precision)\n if (imager.weighting == 'Uniform' or imager.algorithm ==\n 'W-projection'):\n imager.coords_only = True\n log.info('Reading coordinates from %s', f)\n process_input_data(f, imager, None, 0.0, grid_weights)\n imager.coords_only = False\n log.info('Reading visibilities from %s', f)\n grid_norm = process_input_data(f, imager, grid_data, grid_norm,\n grid_weights)\n output_file = splitext(f)[0] + '.fits'\n save_image(imager, grid_data, grid_norm, output_file)\n log.info('Finished. Output file is %s', output_file)\n", "role": "user" }, { "content": "main", "role": "assistant" } ]
def x(): LOG.debug('GET list of SBIs.') _url = get_root_url() response = dict(scheduling_blocks=[], links=dict(home='{}'.format(_url))) block_ids = DB.get_sched_block_instance_ids() for block in DB.get_block_details(block_ids): block_id = block['id'] LOG.debug('Adding SBI %s to list', block_id) LOG.debug(block) block['num_processing_blocks'] = len(block['processing_block_ids']) temp = ['OK'] * 10 + ['WAITING'] * 4 + ['FAILED'] * 2 block['status'] = choice(temp) try: del block['processing_block_ids'] except KeyError: pass block['links'] = {'detail': '{}/scheduling-block/{}'.format(_url, block_id)} response['scheduling_blocks'].append(block) return response, HTTPStatus.OK
get
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n LOG.debug('GET list of SBIs.')\n _url = get_root_url()\n response = dict(scheduling_blocks=[], links=dict(home='{}'.format(_url)))\n block_ids = DB.get_sched_block_instance_ids()\n for block in DB.get_block_details(block_ids):\n block_id = block['id']\n LOG.debug('Adding SBI %s to list', block_id)\n LOG.debug(block)\n block['num_processing_blocks'] = len(block['processing_block_ids'])\n temp = ['OK'] * 10 + ['WAITING'] * 4 + ['FAILED'] * 2\n block['status'] = choice(temp)\n try:\n del block['processing_block_ids']\n except KeyError:\n pass\n block['links'] = {'detail': '{}/scheduling-block/{}'.format(_url,\n block_id)}\n response['scheduling_blocks'].append(block)\n return response, HTTPStatus.OK\n", "role": "user" }, { "content": "get", "role": "assistant" } ]
def x(): response = dict(blocks=[]) block_ids = DB.get_sched_block_instance_ids() for index, block_id in enumerate(block_ids): block = DB.get_block_details([block_id]).__next__() info = [index, block['id'], block['sub_array_id'], len(block[ 'processing_blocks'])] response['blocks'].append(info) return response, HTTPStatus.OK
get_table
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n response = dict(blocks=[])\n block_ids = DB.get_sched_block_instance_ids()\n for index, block_id in enumerate(block_ids):\n block = DB.get_block_details([block_id]).__next__()\n info = [index, block['id'], block['sub_array_id'], len(block[\n 'processing_blocks'])]\n response['blocks'].append(info)\n return response, HTTPStatus.OK\n", "role": "user" }, { "content": "get_table", "role": "assistant" } ]
def x(self, key, field, value): self._db.hset(key, field, value)
set_value
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, key, field, value):\n self._db.hset(key, field, value)\n", "role": "user" }, { "content": "set_value", "role": "assistant" } ]
def x(self, event_name, event_type, block_id): self._db.rpush(event_name, dict(type=event_type, id=block_id))
push_event
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, event_name, event_type, block_id):\n self._db.rpush(event_name, dict(type=event_type, id=block_id))\n", "role": "user" }, { "content": "push_event", "role": "assistant" } ]
def x(): signal.signal(signal.SIGTERM, _sig_handler) with open(sys.argv[1]) as fh: config = json.load(fh) os.chdir(os.path.expanduser('~')) receiver = PulsarStart(config, logging.getLogger()) receiver.run()
main
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n signal.signal(signal.SIGTERM, _sig_handler)\n with open(sys.argv[1]) as fh:\n config = json.load(fh)\n os.chdir(os.path.expanduser('~'))\n receiver = PulsarStart(config, logging.getLogger())\n receiver.run()\n", "role": "user" }, { "content": "main", "role": "assistant" } ]
def x(func): @wraps(func) def x(*args, **kwargs): try: return func(*args, **kwargs) except redis.exceptions.ConnectionError: raise ConnectionError( 'Unable to connect to the Redis Configuration Database. host = {}, port = {}, id = {}.' .format(REDIS_HOST, REDIS_PORT, REDIS_DB_ID)) return with_exception_handling
check_connection
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(func):\n\n @wraps(func)\n def x(*args, **kwargs):\n try:\n return func(*args, **kwargs)\n except redis.exceptions.ConnectionError:\n raise ConnectionError(\n 'Unable to connect to the Redis Configuration Database. host = {}, port = {}, id = {}.'\n .format(REDIS_HOST, REDIS_PORT, REDIS_DB_ID))\n return with_exception_handling\n", "role": "user" }, { "content": "check_connection", "role": "assistant" } ]
def x(self, key: str, my_dict: dict, hierarchical: bool=False): for _key, _value in my_dict.items(): if isinstance(_value, dict): if not hierarchical: self._db.hmset(key, {_key: json.dumps(_value)}) else: self.save_dict(key + ':' + _key, _value, hierarchical) elif isinstance(_value, list): if not hierarchical: self._db.hmset(key, {_key: str(_value)}) else: print('saving list at ', key + ':' + _key) self._db.lpush(key + ':' + _key, *_value[::-1]) elif isinstance(_value, bool): self._db.hmset(key, {_key: str(_value)}) else: self._db.hmset(key, {_key: _value})
save_dict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, key: str, my_dict: dict, hierarchical: bool=False):\n for _key, _value in my_dict.items():\n if isinstance(_value, dict):\n if not hierarchical:\n self._db.hmset(key, {_key: json.dumps(_value)})\n else:\n self.save_dict(key + ':' + _key, _value, hierarchical)\n elif isinstance(_value, list):\n if not hierarchical:\n self._db.hmset(key, {_key: str(_value)})\n else:\n print('saving list at ', key + ':' + _key)\n self._db.lpush(key + ':' + _key, *_value[::-1])\n elif isinstance(_value, bool):\n self._db.hmset(key, {_key: str(_value)})\n else:\n self._db.hmset(key, {_key: _value})\n", "role": "user" }, { "content": "save_dict", "role": "assistant" } ]
def x(my_dict, keys, values): temp = my_dict for depth, key in enumerate(keys): if depth < len(keys) - 1: if key not in temp: temp[key] = dict() temp = temp[key] elif key not in temp: temp[key] = values else: temp[key] = {**temp[key], **values} return my_dict
build_dict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(my_dict, keys, values):\n temp = my_dict\n for depth, key in enumerate(keys):\n if depth < len(keys) - 1:\n if key not in temp:\n temp[key] = dict()\n temp = temp[key]\n elif key not in temp:\n temp[key] = values\n else:\n temp[key] = {**temp[key], **values}\n return my_dict\n", "role": "user" }, { "content": "build_dict", "role": "assistant" } ]
def x(self, db_key: str) ->dict: if self._db.type(db_key) == 'list': db_values = self._db.lrange(db_key, 0, -1) for i, value in enumerate(db_values): try: db_values[i] = ast.literal_eval(value) except SyntaxError: pass except ValueError: pass else: db_values = self._db.hgetall(db_key) for _key, _value in db_values.items(): try: db_values[_key] = ast.literal_eval(_value) except SyntaxError: pass except ValueError: pass return db_values
load_values
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, db_key: str) ->dict:\n if self._db.type(db_key) == 'list':\n db_values = self._db.lrange(db_key, 0, -1)\n for i, value in enumerate(db_values):\n try:\n db_values[i] = ast.literal_eval(value)\n except SyntaxError:\n pass\n except ValueError:\n pass\n else:\n db_values = self._db.hgetall(db_key)\n for _key, _value in db_values.items():\n try:\n db_values[_key] = ast.literal_eval(_value)\n except SyntaxError:\n pass\n except ValueError:\n pass\n return db_values\n", "role": "user" }, { "content": "load_values", "role": "assistant" } ]
def x(self, db_key: str) ->dict: db_keys = self._db.keys(pattern=db_key + '*') my_dict = {} for _db_key in db_keys: if self._db.type(_db_key) == 'list': db_values = self._db.lrange(_db_key, 0, -1) for i, value in enumerate(db_values): try: db_values[i] = ast.literal_eval(value) except SyntaxError: pass except ValueError: pass else: db_values = self._db.hgetall(_db_key) for _key, _value in db_values.items(): try: db_values[_key] = ast.literal_eval(_value) except SyntaxError: pass except ValueError: pass my_dict = self._build_dict(my_dict, _db_key.split(':'), db_values) return my_dict[db_key]
load_dict_hierarchical
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, db_key: str) ->dict:\n db_keys = self._db.keys(pattern=db_key + '*')\n my_dict = {}\n for _db_key in db_keys:\n if self._db.type(_db_key) == 'list':\n db_values = self._db.lrange(_db_key, 0, -1)\n for i, value in enumerate(db_values):\n try:\n db_values[i] = ast.literal_eval(value)\n except SyntaxError:\n pass\n except ValueError:\n pass\n else:\n db_values = self._db.hgetall(_db_key)\n for _key, _value in db_values.items():\n try:\n db_values[_key] = ast.literal_eval(_value)\n except SyntaxError:\n pass\n except ValueError:\n pass\n my_dict = self._build_dict(my_dict, _db_key.split(':'), db_values)\n return my_dict[db_key]\n", "role": "user" }, { "content": "load_dict_hierarchical", "role": "assistant" } ]
def x(self, db_key: str, hierarchical: bool=False) ->dict: if not hierarchical: db_values = self._db.hgetall(db_key) for _key, _value in db_values.items(): if isinstance(_value, str): db_values[_key] = ast.literal_eval(_value) my_dict = db_values else: my_dict = self._load_dict_hierarchical(db_key) return my_dict
load_dict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, db_key: str, hierarchical: bool=False) ->dict:\n if not hierarchical:\n db_values = self._db.hgetall(db_key)\n for _key, _value in db_values.items():\n if isinstance(_value, str):\n db_values[_key] = ast.literal_eval(_value)\n my_dict = db_values\n else:\n my_dict = self._load_dict_hierarchical(db_key)\n return my_dict\n", "role": "user" }, { "content": "load_dict", "role": "assistant" } ]
def x(self, db_key: str, dict_keys: List[str], hierarchical: bool=False ) ->List: result = [] if not hierarchical: _values = self._db.hmget(db_key, *dict_keys) result = [ast.literal_eval(_value) for _value in _values] else: db_keys = self._db.keys(pattern=db_key + '*') for _db_key in db_keys: for name in _db_key.split(':')[1:]: if name in dict_keys: _values = self._load_values(_db_key) result.append(_values) _values = self._db.hmget(_db_key, *dict_keys) for i, value in enumerate(_values): try: _values[i] = ast.literal_eval(value) except SyntaxError: pass except ValueError: pass result += [value for value in _values if value is not None] return result
load_dict_values
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, db_key: str, dict_keys: List[str], hierarchical: bool=False\n ) ->List:\n result = []\n if not hierarchical:\n _values = self._db.hmget(db_key, *dict_keys)\n result = [ast.literal_eval(_value) for _value in _values]\n else:\n db_keys = self._db.keys(pattern=db_key + '*')\n for _db_key in db_keys:\n for name in _db_key.split(':')[1:]:\n if name in dict_keys:\n _values = self._load_values(_db_key)\n result.append(_values)\n _values = self._db.hmget(_db_key, *dict_keys)\n for i, value in enumerate(_values):\n try:\n _values[i] = ast.literal_eval(value)\n except SyntaxError:\n pass\n except ValueError:\n pass\n result += [value for value in _values if value is not None]\n return result\n", "role": "user" }, { "content": "load_dict_values", "role": "assistant" } ]
def x(self, key, field, value, pipeline=False): if pipeline: self._pipeline.hset(key, field, str(value)) else: self._db.hset(key, field, str(value))
set_hash_value
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, key, field, value, pipeline=False):\n if pipeline:\n self._pipeline.hset(key, field, str(value))\n else:\n self._db.hset(key, field, str(value))\n", "role": "user" }, { "content": "set_hash_value", "role": "assistant" } ]
def x(self, key, *value, pipeline=False): if pipeline: self._pipeline.lpush(key, *value) else: self._db.lpush(key, *value)
prepend_to_list
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, key, *value, pipeline=False):\n if pipeline:\n self._pipeline.lpush(key, *value)\n else:\n self._db.lpush(key, *value)\n", "role": "user" }, { "content": "prepend_to_list", "role": "assistant" } ]
def x(self, key, *value, pipeline=False): if pipeline: self._pipeline.rpush(key, *value) else: self._db.rpush(key, *value)
append_to_list
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, key, *value, pipeline=False):\n if pipeline:\n self._pipeline.rpush(key, *value)\n else:\n self._db.rpush(key, *value)\n", "role": "user" }, { "content": "append_to_list", "role": "assistant" } ]
def x(self, key, pipeline=False): if pipeline: return self._pipeline.lrange(key, 0, -1) return self._db.lrange(key, 0, -1)
get_list
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, key, pipeline=False):\n if pipeline:\n return self._pipeline.lrange(key, 0, -1)\n return self._db.lrange(key, 0, -1)\n", "role": "user" }, { "content": "get_list", "role": "assistant" } ]
def x(self, *names: str, pipeline=False): if pipeline: self._pipeline.delete(*names) else: self._db.delete(*names)
delete
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, *names: str, pipeline=False):\n if pipeline:\n self._pipeline.delete(*names)\n else:\n self._db.delete(*names)\n", "role": "user" }, { "content": "delete", "role": "assistant" } ]
def x(self, event_name, event_history=None): if event_history is None: event_history = event_name + '_history' return self._db.rpoplpush(event_name, event_history)
get_event
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, event_name, event_history=None):\n if event_history is None:\n event_history = event_name + '_history'\n return self._db.rpoplpush(event_name, event_history)\n", "role": "user" }, { "content": "get_event", "role": "assistant" } ]
def x(self, key: str, value, count: int=0, pipeline: bool=False): if pipeline: if redis.__version__ == '2.10.6': self._pipeline.lrem(name=key, value=value, num=count) else: self._pipeline.lrem(key, count, value) elif self._db.exists(key): if redis.__version__ == '2.10.6': self._db.lrem(name=key, value=value, num=count) else: self._db.lrem(key, count, value)
remove_from_list
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, key: str, value, count: int=0, pipeline: bool=False):\n if pipeline:\n if redis.__version__ == '2.10.6':\n self._pipeline.lrem(name=key, value=value, num=count)\n else:\n self._pipeline.lrem(key, count, value)\n elif self._db.exists(key):\n if redis.__version__ == '2.10.6':\n self._db.lrem(name=key, value=value, num=count)\n else:\n self._db.lrem(key, count, value)\n", "role": "user" }, { "content": "remove_from_list", "role": "assistant" } ]
def x(self, key, pipeline=False): if pipeline: self._pipeline.watch(key) else: self._db.watch(key)
watch
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, key, pipeline=False):\n if pipeline:\n self._pipeline.watch(key)\n else:\n self._db.watch(key)\n", "role": "user" }, { "content": "watch", "role": "assistant" } ]
def x(self, channel, message, pipeline=False): if pipeline: self._pipeline.publish(channel, message) else: self._db.publish(channel, message)
publish
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, channel, message, pipeline=False):\n if pipeline:\n self._pipeline.publish(channel, message)\n else:\n self._db.publish(channel, message)\n", "role": "user" }, { "content": "publish", "role": "assistant" } ]
def x(self) ->dict: services_health = {} services_ids = self._get_services() for service_id in services_ids: service_name = DC.get_service_name(service_id) if DC.get_replicas(service_id) != DC.get_actual_replica(service_id): services_health[service_name] = 'Unhealthy' else: services_health[service_name] = 'Healthy' return services_health
get_services_health
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self) ->dict:\n services_health = {}\n services_ids = self._get_services()\n for service_id in services_ids:\n service_name = DC.get_service_name(service_id)\n if DC.get_replicas(service_id) != DC.get_actual_replica(service_id):\n services_health[service_name] = 'Unhealthy'\n else:\n services_health[service_name] = 'Healthy'\n return services_health\n", "role": "user" }, { "content": "get_services_health", "role": "assistant" } ]
def x(self) ->str: services_health_status = self.get_services_health() health_status = all(status == 'Healthy' for status in services_health_status.values()) if health_status: overall_status = 'Healthy' else: overall_status = 'Unhealthy' return overall_status
get_overall_services_health
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self) ->str:\n services_health_status = self.get_services_health()\n health_status = all(status == 'Healthy' for status in\n services_health_status.values())\n if health_status:\n overall_status = 'Healthy'\n else:\n overall_status = 'Unhealthy'\n return overall_status\n", "role": "user" }, { "content": "get_overall_services_health", "role": "assistant" } ]
def x(service_id: str) ->str: if DC.get_replicas(service_id) != DC.get_actual_replica(service_id): health_status = 'Unhealthy' else: health_status = 'Healthy' return health_status
get_service_health
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(service_id: str) ->str:\n if DC.get_replicas(service_id) != DC.get_actual_replica(service_id):\n health_status = 'Unhealthy'\n else:\n health_status = 'Healthy'\n return health_status\n", "role": "user" }, { "content": "get_service_health", "role": "assistant" } ]
def x(): uptime = time.time() - START_TIME response = dict(uptime=f'{uptime:.2f}s', links=dict(root='{}'.format( get_root_url()))) return response, HTTPStatus.OK
get
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n uptime = time.time() - START_TIME\n response = dict(uptime=f'{uptime:.2f}s', links=dict(root='{}'.format(\n get_root_url())))\n return response, HTTPStatus.OK\n", "role": "user" }, { "content": "get", "role": "assistant" } ]
def x(): if len(sys.argv) < 2: raise RuntimeError('Usage: python3 async_recv.py <json config>') sip_logging.init_logger(show_thread=True) spead_config = json.loads(sys.argv[1]) receiver = SpeadReceiver(spead_config) receiver.run()
main
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n if len(sys.argv) < 2:\n raise RuntimeError('Usage: python3 async_recv.py <json config>')\n sip_logging.init_logger(show_thread=True)\n spead_config = json.loads(sys.argv[1])\n receiver = SpeadReceiver(spead_config)\n receiver.run()\n", "role": "user" }, { "content": "main", "role": "assistant" } ]
def x(self, i_block, receive_buffer): self._log.info('Worker thread processing block %i', i_block) time_overall0 = time.time() time_unpack = 0.0 time_write = 0.0 for i_heap, heap in enumerate(receive_buffer.result()): if isinstance(heap, spead2.recv.IncompleteHeap): self._log.info('Dropped incomplete heap %i', heap.cnt + 1) continue items = self._item_group.update(heap) i_chan = i_heap // self._num_buffer_times i_time = i_heap % self._num_buffer_times if 'correlator_output_data' in items: vis_data = items['correlator_output_data'].value['VIS'] if self._block is None: num_baselines = vis_data.shape[0] num_pols = vis_data[0].shape[0] self._block = numpy.zeros((self._num_buffer_times, self. _num_streams, num_baselines), dtype=('c8', num_pols)) self._block[:, :, :] = 0 time_unpack0 = time.time() self._block[i_time, i_chan, :] = vis_data time_unpack += time.time() - time_unpack0 val = self._block[i_time, i_chan, -1][-1].real self._log.debug('Data: %.3f', val) if self._block is not None: if self._config['process_data']: pass if self._config['write_data']: time_write0 = time.time() with open(self._config['filename'], 'ab') as f: numpy.save(f, self._block, allow_pickle=False) time_write += time.time() - time_write0 time_overall = time.time() - time_overall0 self._log.info('Total processing time: %.1f ms', 1000 * time_overall) self._log.info('Unpack was %.1f %%', 100 * time_unpack / time_overall) self._log.info('Write was %.1f %%', 100 * time_write / time_overall) if time_unpack != 0.0: self._log.info('Memory speed %.1f MB/s', self._block.nbytes * 1e-06 / time_unpack) if time_write != 0.0: self._log.info('Write speed %.1f MB/s', self._block.nbytes * 1e-06 / time_write)
process_buffer
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, i_block, receive_buffer):\n self._log.info('Worker thread processing block %i', i_block)\n time_overall0 = time.time()\n time_unpack = 0.0\n time_write = 0.0\n for i_heap, heap in enumerate(receive_buffer.result()):\n if isinstance(heap, spead2.recv.IncompleteHeap):\n self._log.info('Dropped incomplete heap %i', heap.cnt + 1)\n continue\n items = self._item_group.update(heap)\n i_chan = i_heap // self._num_buffer_times\n i_time = i_heap % self._num_buffer_times\n if 'correlator_output_data' in items:\n vis_data = items['correlator_output_data'].value['VIS']\n if self._block is None:\n num_baselines = vis_data.shape[0]\n num_pols = vis_data[0].shape[0]\n self._block = numpy.zeros((self._num_buffer_times, self.\n _num_streams, num_baselines), dtype=('c8', num_pols))\n self._block[:, :, :] = 0\n time_unpack0 = time.time()\n self._block[i_time, i_chan, :] = vis_data\n time_unpack += time.time() - time_unpack0\n val = self._block[i_time, i_chan, -1][-1].real\n self._log.debug('Data: %.3f', val)\n if self._block is not None:\n if self._config['process_data']:\n pass\n if self._config['write_data']:\n time_write0 = time.time()\n with open(self._config['filename'], 'ab') as f:\n numpy.save(f, self._block, allow_pickle=False)\n time_write += time.time() - time_write0\n time_overall = time.time() - time_overall0\n self._log.info('Total processing time: %.1f ms', 1000 * time_overall)\n self._log.info('Unpack was %.1f %%', 100 * time_unpack / time_overall)\n self._log.info('Write was %.1f %%', 100 * time_write / time_overall)\n if time_unpack != 0.0:\n self._log.info('Memory speed %.1f MB/s', self._block.nbytes * 1e-06 /\n time_unpack)\n if time_write != 0.0:\n self._log.info('Write speed %.1f MB/s', self._block.nbytes * 1e-06 /\n time_write)\n", "role": "user" }, { "content": "process_buffer", "role": "assistant" } ]
async def _run_loop(self, executor): """Main loop.""" loop = asyncio.get_event_loop() self._log.info('Waiting for %d streams to start...', self._num_streams) for stream in self._streams: await stream.get(loop=loop) i_block = 0 receive_buffer = [None, None] while True: processing_tasks = None if i_block > 0: i_buffer_proc = (i_block - 1) % 2 processing_tasks = loop.run_in_executor(executor, self. process_buffer, i_block - 1, receive_buffer[i_buffer_proc]) i_buffer_recv = i_block % 2 receive_tasks = [] for stream in self._streams: for _ in range(self._num_buffer_times): receive_tasks.append(loop.create_task(stream.get(loop=loop))) receive_buffer[i_buffer_recv] = asyncio.gather(*receive_tasks) self._log.info('Receiving block %i', i_block) try: await receive_buffer[i_buffer_recv] self._log.info('Received block %i', i_block) except spead2.Stopped: self._log.info('Stream Stopped') break if processing_tasks: await processing_tasks i_block += 1
run_loop
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def _run_loop(self, executor):\n \"\"\"Main loop.\"\"\"\n loop = asyncio.get_event_loop()\n self._log.info('Waiting for %d streams to start...', self._num_streams)\n for stream in self._streams:\n await stream.get(loop=loop)\n i_block = 0\n receive_buffer = [None, None]\n while True:\n processing_tasks = None\n if i_block > 0:\n i_buffer_proc = (i_block - 1) % 2\n processing_tasks = loop.run_in_executor(executor, self.\n process_buffer, i_block - 1, receive_buffer[i_buffer_proc])\n i_buffer_recv = i_block % 2\n receive_tasks = []\n for stream in self._streams:\n for _ in range(self._num_buffer_times):\n receive_tasks.append(loop.create_task(stream.get(loop=loop)))\n receive_buffer[i_buffer_recv] = asyncio.gather(*receive_tasks)\n self._log.info('Receiving block %i', i_block)\n try:\n await receive_buffer[i_buffer_recv]\n self._log.info('Received block %i', i_block)\n except spead2.Stopped:\n self._log.info('Stream Stopped')\n break\n if processing_tasks:\n await processing_tasks\n i_block += 1\n", "role": "user" }, { "content": "run_loop", "role": "assistant" } ]
def x(self): executor = concurrent.futures.ThreadPoolExecutor(max_workers=1) loop = asyncio.get_event_loop() loop.run_until_complete(self._run_loop(executor)) self._log.info('Shutting down...') executor.shutdown()
run
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n executor = concurrent.futures.ThreadPoolExecutor(max_workers=1)\n loop = asyncio.get_event_loop()\n loop.run_until_complete(self._run_loop(executor))\n self._log.info('Shutting down...')\n executor.shutdown()\n", "role": "user" }, { "content": "run", "role": "assistant" } ]
def x(self): self.seek(0) while True: brace_count = 0 quoted = False json_string = '' while True: b = self.read(1) if b == b'': return c = b.decode() if c == '\\': json_string += c json_string += self.read(1).decode() continue if c == '"': if quoted: quoted = False else: quoted = True if c == '{': brace_count += 1 elif c == '}': brace_count -= 1 json_string += c if brace_count == 0: break meta = json.loads(json_string) n_channels = meta['data_cube']['n_channels'] n_sub_integrations = meta['data_cube']['n_sub_integrations'] f = open('{0}_{1}_{2}.json'.format(meta['metadata'][ 'observation_id'], meta['metadata']['beam_id'], meta['metadata' ]['name']), 'w') f.write(json.dumps(meta)) f.close() data = self.read(n_channels * n_sub_integrations) f = open('{0}_{1}_{2}.data'.format(meta['metadata'][ 'observation_id'], meta['metadata']['beam_id'], meta['metadata' ]['name']), 'wb') f.write(data) f.close()
close
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n self.seek(0)\n while True:\n brace_count = 0\n quoted = False\n json_string = ''\n while True:\n b = self.read(1)\n if b == b'':\n return\n c = b.decode()\n if c == '\\\\':\n json_string += c\n json_string += self.read(1).decode()\n continue\n if c == '\"':\n if quoted:\n quoted = False\n else:\n quoted = True\n if c == '{':\n brace_count += 1\n elif c == '}':\n brace_count -= 1\n json_string += c\n if brace_count == 0:\n break\n meta = json.loads(json_string)\n n_channels = meta['data_cube']['n_channels']\n n_sub_integrations = meta['data_cube']['n_sub_integrations']\n f = open('{0}_{1}_{2}.json'.format(meta['metadata'][\n 'observation_id'], meta['metadata']['beam_id'], meta['metadata'\n ]['name']), 'w')\n f.write(json.dumps(meta))\n f.close()\n data = self.read(n_channels * n_sub_integrations)\n f = open('{0}_{1}_{2}.data'.format(meta['metadata'][\n 'observation_id'], meta['metadata']['beam_id'], meta['metadata'\n ]['name']), 'wb')\n f.write(data)\n f.close()\n", "role": "user" }, { "content": "close", "role": "assistant" } ]
def x(self): self._log.info('Starting Pulsar Search Interface') authorizer = DummyAuthorizer() authorizer.add_user(self._config['login']['user'], self._config['login' ]['psswd'], '.', perm=self._config['login']['perm']) authorizer.add_anonymous(os.getcwd()) handler = FTPHandler handler.authorizer = authorizer handler.abstracted_fs = PulsarFileSystem handler.banner = 'SKA SDP pulsar search interface.' address = self._config['address']['listen'], self._config['address']['port' ] server = FTPServer(address, handler) server.max_cons = 256 server.max_cons_per_ip = 5 server.serve_forever()
run
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n self._log.info('Starting Pulsar Search Interface')\n authorizer = DummyAuthorizer()\n authorizer.add_user(self._config['login']['user'], self._config['login'\n ]['psswd'], '.', perm=self._config['login']['perm'])\n authorizer.add_anonymous(os.getcwd())\n handler = FTPHandler\n handler.authorizer = authorizer\n handler.abstracted_fs = PulsarFileSystem\n handler.banner = 'SKA SDP pulsar search interface.'\n address = self._config['address']['listen'], self._config['address']['port'\n ]\n server = FTPServer(address, handler)\n server.max_cons = 256\n server.max_cons_per_ip = 5\n server.serve_forever()\n", "role": "user" }, { "content": "run", "role": "assistant" } ]
def x(self) ->str: self._config = self._load_config() return self._config.get('status')
status
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self) ->str:\n self._config = self._load_config()\n return self._config.get('status')\n", "role": "user" }, { "content": "status", "role": "assistant" } ]
def x(self, value): pb_key = SchedulingObject.get_key(PB_KEY, self._pb_id) stages = DB.get_hash_value(pb_key, 'workflow_stages') stages = ast.literal_eval(stages) stages[self._index]['status'] = value DB.set_hash_value(pb_key, 'workflow_stages', stages)
status
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, value):\n pb_key = SchedulingObject.get_key(PB_KEY, self._pb_id)\n stages = DB.get_hash_value(pb_key, 'workflow_stages')\n stages = ast.literal_eval(stages)\n stages[self._index]['status'] = value\n DB.set_hash_value(pb_key, 'workflow_stages', stages)\n", "role": "user" }, { "content": "status", "role": "assistant" } ]
def x(self): pb_key = SchedulingObject.get_key(PB_KEY, self._pb_id) stages = DB.get_hash_value(pb_key, 'workflow_stages') stages = ast.literal_eval(stages) return stages[self._index]
load_config
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n pb_key = SchedulingObject.get_key(PB_KEY, self._pb_id)\n stages = DB.get_hash_value(pb_key, 'workflow_stages')\n stages = ast.literal_eval(stages)\n return stages[self._index]\n", "role": "user" }, { "content": "load_config", "role": "assistant" } ]
def x(): parser = argparse.ArgumentParser(description='Register PB devices.') parser.add_argument('num_pb', type=int, help= 'Number of PBs devices to register.') return parser.parse_args()
parse_command_args
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n parser = argparse.ArgumentParser(description='Register PB devices.')\n parser.add_argument('num_pb', type=int, help=\n 'Number of PBs devices to register.')\n return parser.parse_args()\n", "role": "user" }, { "content": "parse_command_args", "role": "assistant" } ]
def x(num_pbs: int=100): tango_db = Database() LOG.info('Registering PB devices:') dev_info = DbDevInfo() dev_info._class = 'ProcessingBlockDevice' dev_info.server = 'processing_block_ds/1' for index in range(num_pbs): dev_info.name = 'sip_sdp/pb/{:05d}'.format(index) LOG.info('\t%s', dev_info.name) tango_db.add_device(dev_info)
register_pb_devices
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(num_pbs: int=100):\n tango_db = Database()\n LOG.info('Registering PB devices:')\n dev_info = DbDevInfo()\n dev_info._class = 'ProcessingBlockDevice'\n dev_info.server = 'processing_block_ds/1'\n for index in range(num_pbs):\n dev_info.name = 'sip_sdp/pb/{:05d}'.format(index)\n LOG.info('\\t%s', dev_info.name)\n tango_db.add_device(dev_info)\n", "role": "user" }, { "content": "register_pb_devices", "role": "assistant" } ]
def x(cls, config_dict: dict, schema_path: str=None): if schema_path is None: schema_path = join(dirname(__file__), 'schema', 'configure_sbi.json') with open(schema_path, 'r') as file: schema = json.loads(file.read()) validate(config_dict, schema) config_dict['status'] = 'created' if 'subarray_id' not in config_dict: config_dict['subarray_id'] = 'None' timestamp = datetime.datetime.utcnow().isoformat() config_dict['created'] = timestamp config_dict['updated'] = timestamp pb_list = copy.deepcopy(config_dict['processing_blocks']) config_dict.pop('processing_blocks', None) config_dict['processing_block_ids'] = [] for pb in pb_list: config_dict['processing_block_ids'].append(pb['id']) key = SchedulingObject.get_key(SBI_KEY, config_dict['id']) DB.save_dict(key, config_dict, hierarchical=False) key = '{}:active'.format(SBI_KEY) DB.append_to_list(key, config_dict['id']) sbi = SchedulingObject(SBI_KEY, config_dict['id']) sbi.set_status('created') for pb in pb_list: pb['sbi_id'] = config_dict['id'] cls._add_pb(pb) return cls(config_dict['id'])
from_config
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, config_dict: dict, schema_path: str=None):\n if schema_path is None:\n schema_path = join(dirname(__file__), 'schema', 'configure_sbi.json')\n with open(schema_path, 'r') as file:\n schema = json.loads(file.read())\n validate(config_dict, schema)\n config_dict['status'] = 'created'\n if 'subarray_id' not in config_dict:\n config_dict['subarray_id'] = 'None'\n timestamp = datetime.datetime.utcnow().isoformat()\n config_dict['created'] = timestamp\n config_dict['updated'] = timestamp\n pb_list = copy.deepcopy(config_dict['processing_blocks'])\n config_dict.pop('processing_blocks', None)\n config_dict['processing_block_ids'] = []\n for pb in pb_list:\n config_dict['processing_block_ids'].append(pb['id'])\n key = SchedulingObject.get_key(SBI_KEY, config_dict['id'])\n DB.save_dict(key, config_dict, hierarchical=False)\n key = '{}:active'.format(SBI_KEY)\n DB.append_to_list(key, config_dict['id'])\n sbi = SchedulingObject(SBI_KEY, config_dict['id'])\n sbi.set_status('created')\n for pb in pb_list:\n pb['sbi_id'] = config_dict['id']\n cls._add_pb(pb)\n return cls(config_dict['id'])\n", "role": "user" }, { "content": "from_config", "role": "assistant" } ]
def x(self): self.set_status('aborted') DB.remove_from_list('{}:active'.format(self._type), self._id) DB.append_to_list('{}:aborted'.format(self._type), self._id) sbi_pb_ids = ast.literal_eval(DB.get_hash_value(self._key, 'processing_block_ids')) for pb_id in sbi_pb_ids: pb = ProcessingBlock(pb_id) pb.abort()
abort
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n self.set_status('aborted')\n DB.remove_from_list('{}:active'.format(self._type), self._id)\n DB.append_to_list('{}:aborted'.format(self._type), self._id)\n sbi_pb_ids = ast.literal_eval(DB.get_hash_value(self._key,\n 'processing_block_ids'))\n for pb_id in sbi_pb_ids:\n pb = ProcessingBlock(pb_id)\n pb.abort()\n", "role": "user" }, { "content": "abort", "role": "assistant" } ]
def x(self) ->List[str]: values = DB.get_hash_value(self._key, 'processing_block_ids') return ast.literal_eval(values)
get_pb_ids
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self) ->List[str]:\n values = DB.get_hash_value(self._key, 'processing_block_ids')\n return ast.literal_eval(values)\n", "role": "user" }, { "content": "get_pb_ids", "role": "assistant" } ]
def x(date=None, project: str='sip', instance_id: int=None) ->str: if date is None: date = datetime.datetime.utcnow() if isinstance(date, datetime.datetime): date = date.strftime('%Y%m%d') if instance_id is None: instance_id = randint(0, 9999) return 'SBI-{}-{}-{:04d}'.format(date, project, instance_id)
get_id
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(date=None, project: str='sip', instance_id: int=None) ->str:\n if date is None:\n date = datetime.datetime.utcnow()\n if isinstance(date, datetime.datetime):\n date = date.strftime('%Y%m%d')\n if instance_id is None:\n instance_id = randint(0, 9999)\n return 'SBI-{}-{}-{:04d}'.format(date, project, instance_id)\n", "role": "user" }, { "content": "get_id", "role": "assistant" } ]
def x(pb_config: dict): pb_config['status'] = 'created' timestamp = datetime.datetime.utcnow().isoformat() pb_config['created'] = timestamp pb_config['updated'] = timestamp if 'priority' not in pb_config: pb_config['priority'] = 0 SchedulingBlockInstance._update_workflow_definition(pb_config) keys = ['resources_required', 'resources_assigned', 'dependencies'] for key in keys: if key not in pb_config: pb_config[key] = [] for stage in pb_config['workflow_stages']: if key not in stage: stage[key] = [] key = SchedulingObject.get_key(PB_KEY, pb_config['id']) DB.save_dict(key, pb_config, hierarchical=False) key = '{}:active'.format(PB_KEY) DB.append_to_list(key, pb_config['id']) key = '{}:active:{}'.format(PB_KEY, pb_config['type']) DB.append_to_list(key, pb_config['id']) pb = SchedulingObject(PB_KEY, pb_config['id']) pb.set_status('created')
add_pb
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(pb_config: dict):\n pb_config['status'] = 'created'\n timestamp = datetime.datetime.utcnow().isoformat()\n pb_config['created'] = timestamp\n pb_config['updated'] = timestamp\n if 'priority' not in pb_config:\n pb_config['priority'] = 0\n SchedulingBlockInstance._update_workflow_definition(pb_config)\n keys = ['resources_required', 'resources_assigned', 'dependencies']\n for key in keys:\n if key not in pb_config:\n pb_config[key] = []\n for stage in pb_config['workflow_stages']:\n if key not in stage:\n stage[key] = []\n key = SchedulingObject.get_key(PB_KEY, pb_config['id'])\n DB.save_dict(key, pb_config, hierarchical=False)\n key = '{}:active'.format(PB_KEY)\n DB.append_to_list(key, pb_config['id'])\n key = '{}:active:{}'.format(PB_KEY, pb_config['type'])\n DB.append_to_list(key, pb_config['id'])\n pb = SchedulingObject(PB_KEY, pb_config['id'])\n pb.set_status('created')\n", "role": "user" }, { "content": "add_pb", "role": "assistant" } ]
def x(pb_config: dict): known_workflows = get_workflows() workflow_id = pb_config['workflow']['id'] workflow_version = pb_config['workflow']['version'] if (workflow_id not in known_workflows or workflow_version not in known_workflows[workflow_id]): raise RuntimeError('Unknown workflow definition: {}:{}'.format( workflow_id, workflow_version)) workflow = get_workflow(workflow_id, workflow_version) for stage in workflow['stages']: stage['status'] = 'none' pb_config['workflow_parameters'] = pb_config['workflow']['parameters'] pb_config['workflow_id'] = pb_config['workflow']['id'] pb_config['workflow_version'] = pb_config['workflow']['version'] pb_config['workflow_stages'] = workflow['stages'] pb_config.pop('workflow', None)
update_workflow_definition
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(pb_config: dict):\n known_workflows = get_workflows()\n workflow_id = pb_config['workflow']['id']\n workflow_version = pb_config['workflow']['version']\n if (workflow_id not in known_workflows or workflow_version not in\n known_workflows[workflow_id]):\n raise RuntimeError('Unknown workflow definition: {}:{}'.format(\n workflow_id, workflow_version))\n workflow = get_workflow(workflow_id, workflow_version)\n for stage in workflow['stages']:\n stage['status'] = 'none'\n pb_config['workflow_parameters'] = pb_config['workflow']['parameters']\n pb_config['workflow_id'] = pb_config['workflow']['id']\n pb_config['workflow_version'] = pb_config['workflow']['version']\n pb_config['workflow_stages'] = workflow['stages']\n pb_config.pop('workflow', None)\n", "role": "user" }, { "content": "update_workflow_definition", "role": "assistant" } ]
def x(): response = {'links': {'message': 'Welcome to the SIP Processing Controller Interface', 'items': [{ 'href': '{}health'.format(request.url)}, {'href': '{}subarrays'. format(request.url)}, {'href': '{}scheduling_blocks'.format(request .url)}, {'href': '{}processing_blocks'.format(request.url)}]}} return response, HTTPStatus.OK
root
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n response = {'links': {'message':\n 'Welcome to the SIP Processing Controller Interface', 'items': [{\n 'href': '{}health'.format(request.url)}, {'href': '{}subarrays'.\n format(request.url)}, {'href': '{}scheduling_blocks'.format(request\n .url)}, {'href': '{}processing_blocks'.format(request.url)}]}}\n return response, HTTPStatus.OK\n", "role": "user" }, { "content": "root", "role": "assistant" } ]
def x(num_blocks, project='test'): _date = strftime('%Y%m%d', gmtime()) _project = project for i in range(num_blocks): yield '{}-{}-sbi{:03d}'.format(_date, _project, i)
generate_scheduling_block_id
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(num_blocks, project='test'):\n _date = strftime('%Y%m%d', gmtime())\n _project = project\n for i in range(num_blocks):\n yield '{}-{}-sbi{:03d}'.format(_date, _project, i)\n", "role": "user" }, { "content": "generate_scheduling_block_id", "role": "assistant" } ]