source_code
stringlengths 45
7.24k
| function_name
stringlengths 1
111
| messages
list |
---|---|---|
def x() ->List[ServiceState]:
keys = DB.get_keys('states*')
LOG.debug('Loading list of known services.')
services = []
for key in keys:
values = key.split(':')
if len(values) == 4:
services.append(ServiceState(*values[1:4]))
return services
|
get_service_state_list
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x() ->List[ServiceState]:\n keys = DB.get_keys('states*')\n LOG.debug('Loading list of known services.')\n services = []\n for key in keys:\n values = key.split(':')\n if len(values) == 4:\n services.append(ServiceState(*values[1:4]))\n return services\n",
"role": "user"
},
{
"content": "get_service_state_list",
"role": "assistant"
}
] |
def x() ->List[tuple]:
keys = DB.get_keys('states*')
services = []
for key in keys:
values = key.split(':')
if len(values) == 4:
services.append(':'.join(values[1:]))
return services
|
get_service_id_list
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x() ->List[tuple]:\n keys = DB.get_keys('states*')\n services = []\n for key in keys:\n values = key.split(':')\n if len(values) == 4:\n services.append(':'.join(values[1:]))\n return services\n",
"role": "user"
},
{
"content": "get_service_id_list",
"role": "assistant"
}
] |
def x():
LOG.debug('GET Processing Block list')
_url = get_root_url()
block_ids = sorted(DB.get_processing_block_ids())
LOG.debug('Processing Block IDs: %s', block_ids)
response = dict(num_processing_blocks=len(block_ids), processing_blocks
=list())
for block in DB.get_block_details(block_ids):
block_id = block['id']
LOG.debug('Creating PB summary for %s', block_id)
block['links'] = dict(detail='{}/processing-block/{}'.format(_url,
block_id), scheduling_block='{}/scheduling-block/{}'.format(
_url, block_id.split(':')[0]))
response['processing_blocks'].append(block)
response['links'] = {'self': '{}'.format(request.url), 'home': '{}'.
format(_url)}
return response, HTTPStatus.OK
|
get
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n LOG.debug('GET Processing Block list')\n _url = get_root_url()\n block_ids = sorted(DB.get_processing_block_ids())\n LOG.debug('Processing Block IDs: %s', block_ids)\n response = dict(num_processing_blocks=len(block_ids), processing_blocks\n =list())\n for block in DB.get_block_details(block_ids):\n block_id = block['id']\n LOG.debug('Creating PB summary for %s', block_id)\n block['links'] = dict(detail='{}/processing-block/{}'.format(_url,\n block_id), scheduling_block='{}/scheduling-block/{}'.format(\n _url, block_id.split(':')[0]))\n response['processing_blocks'].append(block)\n response['links'] = {'self': '{}'.format(request.url), 'home': '{}'.\n format(_url)}\n return response, HTTPStatus.OK\n",
"role": "user"
},
{
"content": "get",
"role": "assistant"
}
] |
def x(self) ->dict:
self._check_object_exists()
config_dict = DB.get_hash_dict(self.key)
for _, value in config_dict.items():
for char in ['[', '{']:
if char in value:
value = ast.literal_eval(value)
return config_dict
|
config
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self) ->dict:\n self._check_object_exists()\n config_dict = DB.get_hash_dict(self.key)\n for _, value in config_dict.items():\n for char in ['[', '{']:\n if char in value:\n value = ast.literal_eval(value)\n return config_dict\n",
"role": "user"
},
{
"content": "config",
"role": "assistant"
}
] |
def x(self, property_key: str) ->str:
self._check_object_exists()
return DB.get_hash_value(self.key, property_key)
|
get_property
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, property_key: str) ->str:\n self._check_object_exists()\n return DB.get_hash_value(self.key, property_key)\n",
"role": "user"
},
{
"content": "get_property",
"role": "assistant"
}
] |
def x(self, value):
self._check_object_exists()
DB.set_hash_value(self.key, 'status', value)
self.publish('status_changed', event_data=dict(status=value))
|
set_status
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, value):\n self._check_object_exists()\n DB.set_hash_value(self.key, 'status', value)\n self.publish('status_changed', event_data=dict(status=value))\n",
"role": "user"
},
{
"content": "set_status",
"role": "assistant"
}
] |
def x(self, event_type: str, event_data: dict=None):
import inspect
import os.path
_stack = inspect.stack()
_origin = os.path.basename(_stack[3][1]) + '::' + _stack[3][3
] + '::L{}'.format(_stack[3][2])
publish(event_type=event_type, event_data=event_data, object_type=self.
_type, object_id=self._id, object_key=self._key, origin=_origin)
|
publish
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, event_type: str, event_data: dict=None):\n import inspect\n import os.path\n _stack = inspect.stack()\n _origin = os.path.basename(_stack[3][1]) + '::' + _stack[3][3\n ] + '::L{}'.format(_stack[3][2])\n publish(event_type=event_type, event_data=event_data, object_type=self.\n _type, object_id=self._id, object_key=self._key, origin=_origin)\n",
"role": "user"
},
{
"content": "publish",
"role": "assistant"
}
] |
def x(self) ->List[Event]:
LOG.debug('Getting events for %s', self.key)
return get_events(self.key)
|
get_events
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self) ->List[Event]:\n LOG.debug('Getting events for %s', self.key)\n return get_events(self.key)\n",
"role": "user"
},
{
"content": "get_events",
"role": "assistant"
}
] |
def x(self):
if not DB.get_keys(self.key):
raise KeyError("Object with key '{}' not exist".format(self.key))
|
check_object_exists
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if not DB.get_keys(self.key):\n raise KeyError(\"Object with key '{}' not exist\".format(self.key))\n",
"role": "user"
},
{
"content": "check_object_exists",
"role": "assistant"
}
] |
def x(subsystem: str, name: str, version: str) ->str:
return '{}:{}:{}'.format(subsystem, name, version)
|
get_service_state_object_id
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(subsystem: str, name: str, version: str) ->str:\n return '{}:{}:{}'.format(subsystem, name, version)\n",
"role": "user"
},
{
"content": "get_service_state_object_id",
"role": "assistant"
}
] |
def x(self, compose_str: str) ->list:
if not self._manager:
raise RuntimeError('Services can only be run on swarm manager nodes')
services_ids = []
try:
service_config = yaml.load(compose_str)
service_list = copy.deepcopy(service_config)
service_config.pop('version')
service_config.pop('services')
for service_name in service_list['services']:
service_exist = self._client.services.list(filters={'name':
service_name})
if not service_exist:
service_config['name'] = service_name
service_spec = self._parse_services(service_config,
service_name, service_list)
created_service = self._client.services.create(**service_spec)
service_id = created_service.short_id
LOG.debug('Service created: %s', service_id)
services_ids.append(service_id)
else:
LOG.debug('Services already exists')
except yaml.YAMLError as exc:
print(exc)
return services_ids
|
create_services
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, compose_str: str) ->list:\n if not self._manager:\n raise RuntimeError('Services can only be run on swarm manager nodes')\n services_ids = []\n try:\n service_config = yaml.load(compose_str)\n service_list = copy.deepcopy(service_config)\n service_config.pop('version')\n service_config.pop('services')\n for service_name in service_list['services']:\n service_exist = self._client.services.list(filters={'name':\n service_name})\n if not service_exist:\n service_config['name'] = service_name\n service_spec = self._parse_services(service_config,\n service_name, service_list)\n created_service = self._client.services.create(**service_spec)\n service_id = created_service.short_id\n LOG.debug('Service created: %s', service_id)\n services_ids.append(service_id)\n else:\n LOG.debug('Services already exists')\n except yaml.YAMLError as exc:\n print(exc)\n return services_ids\n",
"role": "user"
},
{
"content": "create_services",
"role": "assistant"
}
] |
def x(self, volume_name: str, driver_spec: str=None):
if driver_spec:
driver = driver_spec
else:
driver = 'local'
if not self._manager:
raise RuntimeError(
'Services can only be deleted on swarm manager nodes')
self._client.volumes.create(name=volume_name, driver=driver)
|
create_volume
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, volume_name: str, driver_spec: str=None):\n if driver_spec:\n driver = driver_spec\n else:\n driver = 'local'\n if not self._manager:\n raise RuntimeError(\n 'Services can only be deleted on swarm manager nodes')\n self._client.volumes.create(name=volume_name, driver=driver)\n",
"role": "user"
},
{
"content": "create_volume",
"role": "assistant"
}
] |
def x(self, service: str):
if not self._manager:
raise RuntimeError(
'Services can only be deleted on swarm manager nodes')
self._api_client.remove_service(service)
|
delete_service
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, service: str):\n if not self._manager:\n raise RuntimeError(\n 'Services can only be deleted on swarm manager nodes')\n self._api_client.remove_service(service)\n",
"role": "user"
},
{
"content": "delete_service",
"role": "assistant"
}
] |
def x(self):
if not self._manager:
raise RuntimeError(
'Services can only be deleted on swarm manager nodes')
service_list = self.get_service_list()
for services in service_list:
self._api_client.remove_service(services)
|
delete_all_services
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if not self._manager:\n raise RuntimeError(\n 'Services can only be deleted on swarm manager nodes')\n service_list = self.get_service_list()\n for services in service_list:\n self._api_client.remove_service(services)\n",
"role": "user"
},
{
"content": "delete_all_services",
"role": "assistant"
}
] |
def x(self, volume_name: str):
if not self._manager:
raise RuntimeError('Volumes can only be deleted on swarm manager nodes'
)
self._api_client.remove_volume(volume_name)
|
delete_volume
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, volume_name: str):\n if not self._manager:\n raise RuntimeError('Volumes can only be deleted on swarm manager nodes'\n )\n self._api_client.remove_volume(volume_name)\n",
"role": "user"
},
{
"content": "delete_volume",
"role": "assistant"
}
] |
def x(self):
if not self._manager:
raise RuntimeError('Volumes can only be deleted on swarm manager nodes'
)
volume_list = self.get_volume_list()
for volumes in volume_list:
self._api_client.remove_volume(volumes, force=True)
|
delete_all_volumes
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if not self._manager:\n raise RuntimeError('Volumes can only be deleted on swarm manager nodes'\n )\n volume_list = self.get_volume_list()\n for volumes in volume_list:\n self._api_client.remove_volume(volumes, force=True)\n",
"role": "user"
},
{
"content": "delete_all_volumes",
"role": "assistant"
}
] |
def x(self) ->list:
services = []
if not self._manager:
raise RuntimeError(
'Only the Swarm manager node can retrieve all the services.')
service_list = self._client.services.list()
for s_list in service_list:
services.append(s_list.short_id)
return services
|
get_service_list
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self) ->list:\n services = []\n if not self._manager:\n raise RuntimeError(\n 'Only the Swarm manager node can retrieve all the services.')\n service_list = self._client.services.list()\n for s_list in service_list:\n services.append(s_list.short_id)\n return services\n",
"role": "user"
},
{
"content": "get_service_list",
"role": "assistant"
}
] |
def x(self, service_id: str) ->str:
if not self._manager:
raise RuntimeError(
'Only the Swarm manager node can retrieve all the services details.'
)
service = self._client.services.get(service_id)
return service.name
|
get_service_name
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, service_id: str) ->str:\n if not self._manager:\n raise RuntimeError(\n 'Only the Swarm manager node can retrieve all the services details.'\n )\n service = self._client.services.get(service_id)\n return service.name\n",
"role": "user"
},
{
"content": "get_service_name",
"role": "assistant"
}
] |
def x(self, service_id: str) ->dict:
if not self._manager:
raise RuntimeError(
'Only the Swarm manager node can retrieve all the services details.'
)
service = self._client.services.get(service_id)
return service.attrs
|
get_service_details
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, service_id: str) ->dict:\n if not self._manager:\n raise RuntimeError(\n 'Only the Swarm manager node can retrieve all the services details.'\n )\n service = self._client.services.get(service_id)\n return service.attrs\n",
"role": "user"
},
{
"content": "get_service_details",
"role": "assistant"
}
] |
def x(self, service_id: str) ->str:
service = self._client.services.get(service_id)
for service_task in service.tasks():
service_state = service_task['DesiredState']
return service_state
|
get_service_state
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, service_id: str) ->str:\n service = self._client.services.get(service_id)\n for service_task in service.tasks():\n service_state = service_task['DesiredState']\n return service_state\n",
"role": "user"
},
{
"content": "get_service_state",
"role": "assistant"
}
] |
def x(self) ->list:
nodes = []
if not self._manager:
raise RuntimeError(
'Only the Swarm manager node can retrieve all the nodes.')
node_list = self._client.nodes.list()
for n_list in node_list:
nodes.append(n_list.id)
return nodes
|
get_node_list
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self) ->list:\n nodes = []\n if not self._manager:\n raise RuntimeError(\n 'Only the Swarm manager node can retrieve all the nodes.')\n node_list = self._client.nodes.list()\n for n_list in node_list:\n nodes.append(n_list.id)\n return nodes\n",
"role": "user"
},
{
"content": "get_node_list",
"role": "assistant"
}
] |
def x(self, node_id: list) ->dict:
if not self._manager:
raise RuntimeError(
'Only the Swarm manager node can retrieve node details.')
node = self._client.nodes.get(node_id)
return node.attrs
|
get_node_details
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, node_id: list) ->dict:\n if not self._manager:\n raise RuntimeError(\n 'Only the Swarm manager node can retrieve node details.')\n node = self._client.nodes.get(node_id)\n return node.attrs\n",
"role": "user"
},
{
"content": "get_node_details",
"role": "assistant"
}
] |
def x(self) ->list:
containers = []
containers_list = self._client.containers.list()
for c_list in containers_list:
containers.append(c_list.short_id)
return containers
|
get_container_list
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self) ->list:\n containers = []\n containers_list = self._client.containers.list()\n for c_list in containers_list:\n containers.append(c_list.short_id)\n return containers\n",
"role": "user"
},
{
"content": "get_container_list",
"role": "assistant"
}
] |
def x(self, container_id_or_name: str) ->dict:
container = self._client.containers.get(container_id_or_name)
return container.attrs
|
get_container_details
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, container_id_or_name: str) ->dict:\n container = self._client.containers.get(container_id_or_name)\n return container.attrs\n",
"role": "user"
},
{
"content": "get_container_details",
"role": "assistant"
}
] |
def x(self) ->list:
volumes = []
if not self._manager:
raise RuntimeError(
'Only the Swarm manager node can retrieve all the services.')
volume_list = self._client.volumes.list()
for v_list in volume_list:
volumes.append(v_list.name)
return volumes
|
get_volume_list
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self) ->list:\n volumes = []\n if not self._manager:\n raise RuntimeError(\n 'Only the Swarm manager node can retrieve all the services.')\n volume_list = self._client.volumes.list()\n for v_list in volume_list:\n volumes.append(v_list.name)\n return volumes\n",
"role": "user"
},
{
"content": "get_volume_list",
"role": "assistant"
}
] |
def x(self, volume_name: str) ->dict:
if volume_name not in self.volumes:
raise RuntimeError('No such volume found: ', volume_name)
volume = self._client.volumes.get(volume_name)
return volume.attrs
|
get_volume_details
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, volume_name: str) ->dict:\n if volume_name not in self.volumes:\n raise RuntimeError('No such volume found: ', volume_name)\n volume = self._client.volumes.get(volume_name)\n return volume.attrs\n",
"role": "user"
},
{
"content": "get_volume_details",
"role": "assistant"
}
] |
def x(self, service_id: str) ->str:
if not self._manager:
raise RuntimeError(
'Only the Swarm manager node can retrieve replication level of the service'
)
service_details = self.get_service_details(service_id)
actual_replica = service_details['Spec']['Mode']['Replicated']['Replicas']
return actual_replica
|
get_actual_replica
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, service_id: str) ->str:\n if not self._manager:\n raise RuntimeError(\n 'Only the Swarm manager node can retrieve replication level of the service'\n )\n service_details = self.get_service_details(service_id)\n actual_replica = service_details['Spec']['Mode']['Replicated']['Replicas']\n return actual_replica\n",
"role": "user"
},
{
"content": "get_actual_replica",
"role": "assistant"
}
] |
def x(self, service_id: str) ->str:
replicas = []
if not self._manager:
raise RuntimeError(
'Only the Swarm manager node can retrieve replication level of the service'
)
service_tasks = self._client.services.get(service_id).tasks()
for task in service_tasks:
if task['Status']['State'] == 'running':
replicas.append(task)
return len(replicas)
|
get_replicas
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, service_id: str) ->str:\n replicas = []\n if not self._manager:\n raise RuntimeError(\n 'Only the Swarm manager node can retrieve replication level of the service'\n )\n service_tasks = self._client.services.get(service_id).tasks()\n for task in service_tasks:\n if task['Status']['State'] == 'running':\n replicas.append(task)\n return len(replicas)\n",
"role": "user"
},
{
"content": "get_replicas",
"role": "assistant"
}
] |
def x(self, node_name: str, labels: dict):
if not self._manager:
raise RuntimeError(
'Only the Swarm manager node can update node details.')
node_spec = {'Availability': 'active', 'Name': node_name, 'Role':
'manager', 'Labels': labels}
node = self._client.nodes.get(node_name)
node.update(node_spec)
|
update_labels
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, node_name: str, labels: dict):\n if not self._manager:\n raise RuntimeError(\n 'Only the Swarm manager node can update node details.')\n node_spec = {'Availability': 'active', 'Name': node_name, 'Role':\n 'manager', 'Labels': labels}\n node = self._client.nodes.get(node_name)\n node.update(node_spec)\n",
"role": "user"
},
{
"content": "update_labels",
"role": "assistant"
}
] |
def x(self, service_config: dict, service_name: str, service_list: dict
) ->dict:
for key, value in service_list['services'][service_name].items():
service_config[key] = value
if 'command' in key:
key = 'args'
service_config['args'] = value
service_config.pop('command')
if 'ports' in key:
endpoint_spec = self._parse_ports(value)
service_config['endpoint_spec'] = endpoint_spec
service_config.pop('ports')
if 'volumes' in key:
volume_spec = self._parse_volumes(value)
service_config['mounts'] = volume_spec
service_config.pop('volumes')
if 'deploy' in key:
self._parse_deploy(value, service_config)
service_config.pop('deploy')
if 'networks' in key:
network_spec = self._parse_networks(service_list)
service_config['networks'] = network_spec
if 'logging' in key:
self._parse_logging(value, service_config)
service_config.pop('logging')
if 'environment' in key:
service_config['env'] = value
service_config.pop('environment')
return service_config
|
parse_services
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, service_config: dict, service_name: str, service_list: dict\n ) ->dict:\n for key, value in service_list['services'][service_name].items():\n service_config[key] = value\n if 'command' in key:\n key = 'args'\n service_config['args'] = value\n service_config.pop('command')\n if 'ports' in key:\n endpoint_spec = self._parse_ports(value)\n service_config['endpoint_spec'] = endpoint_spec\n service_config.pop('ports')\n if 'volumes' in key:\n volume_spec = self._parse_volumes(value)\n service_config['mounts'] = volume_spec\n service_config.pop('volumes')\n if 'deploy' in key:\n self._parse_deploy(value, service_config)\n service_config.pop('deploy')\n if 'networks' in key:\n network_spec = self._parse_networks(service_list)\n service_config['networks'] = network_spec\n if 'logging' in key:\n self._parse_logging(value, service_config)\n service_config.pop('logging')\n if 'environment' in key:\n service_config['env'] = value\n service_config.pop('environment')\n return service_config\n",
"role": "user"
},
{
"content": "parse_services",
"role": "assistant"
}
] |
def x(self, deploy_values: dict, service_config: dict):
mode = {}
for d_value in deploy_values:
if 'restart_policy' in d_value:
restart_spec = docker.types.RestartPolicy(**deploy_values[d_value])
service_config['restart_policy'] = restart_spec
if 'placement' in d_value:
for constraints_key, constraints_value in deploy_values[d_value
].items():
service_config[constraints_key] = constraints_value
if 'mode' in d_value:
mode[d_value] = deploy_values[d_value]
if 'replicas' in d_value:
mode[d_value] = deploy_values[d_value]
if 'resources' in d_value:
resource_spec = self._parse_resources(deploy_values, d_value)
service_config['resources'] = resource_spec
mode_spec = docker.types.ServiceMode(**mode)
service_config['mode'] = mode_spec
|
parse_deploy
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, deploy_values: dict, service_config: dict):\n mode = {}\n for d_value in deploy_values:\n if 'restart_policy' in d_value:\n restart_spec = docker.types.RestartPolicy(**deploy_values[d_value])\n service_config['restart_policy'] = restart_spec\n if 'placement' in d_value:\n for constraints_key, constraints_value in deploy_values[d_value\n ].items():\n service_config[constraints_key] = constraints_value\n if 'mode' in d_value:\n mode[d_value] = deploy_values[d_value]\n if 'replicas' in d_value:\n mode[d_value] = deploy_values[d_value]\n if 'resources' in d_value:\n resource_spec = self._parse_resources(deploy_values, d_value)\n service_config['resources'] = resource_spec\n mode_spec = docker.types.ServiceMode(**mode)\n service_config['mode'] = mode_spec\n",
"role": "user"
},
{
"content": "parse_deploy",
"role": "assistant"
}
] |
def x(port_values: dict) ->dict:
endpoints = {}
for port_element in port_values:
target_port = port_element.split(':')
for port in target_port:
endpoints[int(port)] = int(port)
endpoint_spec = docker.types.EndpointSpec(ports=endpoints)
return endpoint_spec
|
parse_ports
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(port_values: dict) ->dict:\n endpoints = {}\n for port_element in port_values:\n target_port = port_element.split(':')\n for port in target_port:\n endpoints[int(port)] = int(port)\n endpoint_spec = docker.types.EndpointSpec(ports=endpoints)\n return endpoint_spec\n",
"role": "user"
},
{
"content": "parse_ports",
"role": "assistant"
}
] |
def x(volume_values: dict) ->str:
for v_values in volume_values:
for v_key, v_value in v_values.items():
if v_key == 'source':
if v_value == '.':
source = os.path.dirname(os.path.abspath(__file__))
else:
source = v_value
if v_key == 'target':
target = v_value
volume_spec = [source + ':' + target]
return volume_spec
|
parse_volumes
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(volume_values: dict) ->str:\n for v_values in volume_values:\n for v_key, v_value in v_values.items():\n if v_key == 'source':\n if v_value == '.':\n source = os.path.dirname(os.path.abspath(__file__))\n else:\n source = v_value\n if v_key == 'target':\n target = v_value\n volume_spec = [source + ':' + target]\n return volume_spec\n",
"role": "user"
},
{
"content": "parse_volumes",
"role": "assistant"
}
] |
def x(resource_values: dict, resource_name: str) ->dict:
resources = {}
for r_values in resource_values[resource_name]:
if 'limits' in r_values:
for r_key, r_value in resource_values[resource_name][r_values
].items():
if 'cpu' in r_key:
cpu_value = float(r_value) * 10 ** 9
cpu_key = r_key[:3] + '_limit'
resources[cpu_key] = int(cpu_value)
if 'mem' in r_key:
mem_value = re.sub('M', '', r_value)
mem_key = r_key[:3] + '_limit'
resources[mem_key] = int(mem_value) * 1048576
resources_spec = docker.types.Resources(**resources)
return resources_spec
|
parse_resources
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(resource_values: dict, resource_name: str) ->dict:\n resources = {}\n for r_values in resource_values[resource_name]:\n if 'limits' in r_values:\n for r_key, r_value in resource_values[resource_name][r_values\n ].items():\n if 'cpu' in r_key:\n cpu_value = float(r_value) * 10 ** 9\n cpu_key = r_key[:3] + '_limit'\n resources[cpu_key] = int(cpu_value)\n if 'mem' in r_key:\n mem_value = re.sub('M', '', r_value)\n mem_key = r_key[:3] + '_limit'\n resources[mem_key] = int(mem_value) * 1048576\n resources_spec = docker.types.Resources(**resources)\n return resources_spec\n",
"role": "user"
},
{
"content": "parse_resources",
"role": "assistant"
}
] |
def x(service_list: dict) ->list:
networks = []
for n_values in service_list['networks'].values():
for n_key, n_value in n_values.items():
if 'name' in n_key:
networks.append(n_value)
return networks
|
parse_networks
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(service_list: dict) ->list:\n networks = []\n for n_values in service_list['networks'].values():\n for n_key, n_value in n_values.items():\n if 'name' in n_key:\n networks.append(n_value)\n return networks\n",
"role": "user"
},
{
"content": "parse_networks",
"role": "assistant"
}
] |
def x(log_values: dict, service_config: dict):
for log_key, log_value in log_values.items():
if 'driver' in log_key:
service_config['log_driver'] = log_value
if 'options' in log_key:
service_config['log_driver_options'] = log_value
|
parse_logging
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(log_values: dict, service_config: dict):\n for log_key, log_value in log_values.items():\n if 'driver' in log_key:\n service_config['log_driver'] = log_value\n if 'options' in log_key:\n service_config['log_driver_options'] = log_value\n",
"role": "user"
},
{
"content": "parse_logging",
"role": "assistant"
}
] |
def x():
LOG.info('Initialising Processing Block queue.')
queue = ProcessingBlockQueue()
active_pb_ids = ProcessingBlockList().active
LOG.info('Initialising PC PB queue: %s', active_pb_ids)
for pb_id in active_pb_ids:
pb = ProcessingBlock(pb_id)
queue.put(pb.id, pb.priority, pb.type)
return queue
|
init_queue
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n LOG.info('Initialising Processing Block queue.')\n queue = ProcessingBlockQueue()\n active_pb_ids = ProcessingBlockList().active\n LOG.info('Initialising PC PB queue: %s', active_pb_ids)\n for pb_id in active_pb_ids:\n pb = ProcessingBlock(pb_id)\n queue.put(pb.id, pb.priority, pb.type)\n return queue\n",
"role": "user"
},
{
"content": "init_queue",
"role": "assistant"
}
] |
def x(self):
LOG.info('Starting to monitor PB events')
check_counter = 0
while True:
if check_counter == 50:
check_counter = 0
LOG.debug('Checking for PB events...')
published_events = self._pb_events.get_published_events()
for event in published_events:
if event.type == 'status_changed':
LOG.info('PB status changed event: %s', event.data['status'])
if event.data['status'] == 'created':
LOG.info(
'Acknowledged PB created event (%s) for %s, [timestamp: %s]'
, event.id, event.object_id, event.timestamp)
pb = ProcessingBlock(event.object_id)
self._queue.put(event.object_id, pb.priority, pb.type)
if event.data['status'] == 'completed':
LOG.info(
'Acknowledged PB completed event (%s) for %s, [timestamp: %s]'
, event.id, event.object_id, event.timestamp)
self._num_pbcs -= 1
if self._num_pbcs < 0:
self._num_pbcs = 0
time.sleep(0.1)
check_counter += 1
|
monitor_events
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n LOG.info('Starting to monitor PB events')\n check_counter = 0\n while True:\n if check_counter == 50:\n check_counter = 0\n LOG.debug('Checking for PB events...')\n published_events = self._pb_events.get_published_events()\n for event in published_events:\n if event.type == 'status_changed':\n LOG.info('PB status changed event: %s', event.data['status'])\n if event.data['status'] == 'created':\n LOG.info(\n 'Acknowledged PB created event (%s) for %s, [timestamp: %s]'\n , event.id, event.object_id, event.timestamp)\n pb = ProcessingBlock(event.object_id)\n self._queue.put(event.object_id, pb.priority, pb.type)\n if event.data['status'] == 'completed':\n LOG.info(\n 'Acknowledged PB completed event (%s) for %s, [timestamp: %s]'\n , event.id, event.object_id, event.timestamp)\n self._num_pbcs -= 1\n if self._num_pbcs < 0:\n self._num_pbcs = 0\n time.sleep(0.1)\n check_counter += 1\n",
"role": "user"
},
{
"content": "monitor_events",
"role": "assistant"
}
] |
def x(self):
LOG.info('Starting Processing Block queue reporter.')
while True:
LOG.info('PB queue length = %d', len(self._queue))
time.sleep(self._report_interval)
if active_count() != 5:
LOG.critical(
'Processing Controller not running correctly! (%d/%d threads active)'
, active_count(), 5)
|
processing_controller_status
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n LOG.info('Starting Processing Block queue reporter.')\n while True:\n LOG.info('PB queue length = %d', len(self._queue))\n time.sleep(self._report_interval)\n if active_count() != 5:\n LOG.critical(\n 'Processing Controller not running correctly! (%d/%d threads active)'\n , active_count(), 5)\n",
"role": "user"
},
{
"content": "processing_controller_status",
"role": "assistant"
}
] |
def x(self):
LOG.info('Starting to Schedule Processing Blocks.')
while True:
time.sleep(0.5)
if not self._queue:
continue
if self._num_pbcs >= self._max_pbcs:
LOG.warning('Resource limit reached!')
continue
_inspect = Inspect(app=APP)
if self._queue and _inspect.active() is not None:
next_pb = self._queue[-1]
LOG.info('Considering %s for execution...', next_pb[2])
utc_now = datetime.datetime.utcnow()
time_in_queue = utc_now - datetime_from_isoformat(next_pb[4])
if time_in_queue.total_seconds() >= 10:
item = self._queue.get()
LOG.info('------------------------------------')
LOG.info('>>> Executing %s! <<<', item)
LOG.info('------------------------------------')
execute_processing_block.delay(item)
self._num_pbcs += 1
else:
LOG.info('Waiting for resources for %s', next_pb[2])
|
schedule_processing_blocks
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n LOG.info('Starting to Schedule Processing Blocks.')\n while True:\n time.sleep(0.5)\n if not self._queue:\n continue\n if self._num_pbcs >= self._max_pbcs:\n LOG.warning('Resource limit reached!')\n continue\n _inspect = Inspect(app=APP)\n if self._queue and _inspect.active() is not None:\n next_pb = self._queue[-1]\n LOG.info('Considering %s for execution...', next_pb[2])\n utc_now = datetime.datetime.utcnow()\n time_in_queue = utc_now - datetime_from_isoformat(next_pb[4])\n if time_in_queue.total_seconds() >= 10:\n item = self._queue.get()\n LOG.info('------------------------------------')\n LOG.info('>>> Executing %s! <<<', item)\n LOG.info('------------------------------------')\n execute_processing_block.delay(item)\n self._num_pbcs += 1\n else:\n LOG.info('Waiting for resources for %s', next_pb[2])\n",
"role": "user"
},
{
"content": "schedule_processing_blocks",
"role": "assistant"
}
] |
def x(self):
LOG.info('Starting to Monitor PBC status.')
inspect = celery.current_app.control.inspect()
workers = inspect.ping()
start_time = time.time()
while workers is None:
time.sleep(0.1)
elapsed = time.time() - start_time
if elapsed > 20.0:
LOG.warning('PBC not found!')
break
if workers is not None:
for worker in workers:
_tasks = inspect.registered_tasks()[worker]
LOG.info('Worker: %s tasks:', worker)
for task_index, task_name in enumerate(_tasks):
LOG.info(' %02d : %s', task_index, task_name)
while True:
LOG.info('Checking PBC status (%d/%d)', self._num_pbcs, self._max_pbcs)
celery_app = celery.current_app
inspect = celery_app.control.inspect()
workers = inspect.ping()
if workers is None:
LOG.warning('PBC service not found!')
else:
LOG.info('PBC state: %s', celery_app.events.State())
_active = inspect.active()
_scheduled = inspect.scheduled()
for worker in workers:
LOG.info(' Worker %s: scheduled: %s, active: %s', worker,
_active[worker], _scheduled[worker])
time.sleep(self._report_interval)
|
monitor_pbc_status
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n LOG.info('Starting to Monitor PBC status.')\n inspect = celery.current_app.control.inspect()\n workers = inspect.ping()\n start_time = time.time()\n while workers is None:\n time.sleep(0.1)\n elapsed = time.time() - start_time\n if elapsed > 20.0:\n LOG.warning('PBC not found!')\n break\n if workers is not None:\n for worker in workers:\n _tasks = inspect.registered_tasks()[worker]\n LOG.info('Worker: %s tasks:', worker)\n for task_index, task_name in enumerate(_tasks):\n LOG.info(' %02d : %s', task_index, task_name)\n while True:\n LOG.info('Checking PBC status (%d/%d)', self._num_pbcs, self._max_pbcs)\n celery_app = celery.current_app\n inspect = celery_app.control.inspect()\n workers = inspect.ping()\n if workers is None:\n LOG.warning('PBC service not found!')\n else:\n LOG.info('PBC state: %s', celery_app.events.State())\n _active = inspect.active()\n _scheduled = inspect.scheduled()\n for worker in workers:\n LOG.info(' Worker %s: scheduled: %s, active: %s', worker,\n _active[worker], _scheduled[worker])\n time.sleep(self._report_interval)\n",
"role": "user"
},
{
"content": "monitor_pbc_status",
"role": "assistant"
}
] |
def x(self):
scheduler_threads = [Thread(target=self._monitor_events, daemon=True),
Thread(target=self._processing_controller_status, daemon=True),
Thread(target=self._schedule_processing_blocks, daemon=True),
Thread(target=self._monitor_pbc_status, daemon=True)]
for thread in scheduler_threads:
thread.start()
try:
for thread in scheduler_threads:
thread.join()
except KeyboardInterrupt:
LOG.info('Keyboard interrupt!')
sys.exit(0)
finally:
LOG.info('Finally!')
|
start
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n scheduler_threads = [Thread(target=self._monitor_events, daemon=True),\n Thread(target=self._processing_controller_status, daemon=True),\n Thread(target=self._schedule_processing_blocks, daemon=True),\n Thread(target=self._monitor_pbc_status, daemon=True)]\n for thread in scheduler_threads:\n thread.start()\n try:\n for thread in scheduler_threads:\n thread.join()\n except KeyboardInterrupt:\n LOG.info('Keyboard interrupt!')\n sys.exit(0)\n finally:\n LOG.info('Finally!')\n",
"role": "user"
},
{
"content": "start",
"role": "assistant"
}
] |
def x():
if request.method == 'POST':
response = {'message': 'POST Accepted'}
logging.info('alarm POSTED!')
data = request.data
logging.info(data)
string = json.dumps(data)
producer.send('SIP-alarms', string.encode())
return response
return ''
|
alarm
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n if request.method == 'POST':\n response = {'message': 'POST Accepted'}\n logging.info('alarm POSTED!')\n data = request.data\n logging.info(data)\n string = json.dumps(data)\n producer.send('SIP-alarms', string.encode())\n return response\n return ''\n",
"role": "user"
},
{
"content": "alarm",
"role": "assistant"
}
] |
def x(service: ServiceState):
LOG.debug('Setting current state from target state for %s', service.id)
service.update_current_state(service.target_state)
|
update_service_current_state
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(service: ServiceState):\n LOG.debug('Setting current state from target state for %s', service.id)\n service.update_current_state(service.target_state)\n",
"role": "user"
},
{
"content": "update_service_current_state",
"role": "assistant"
}
] |
def x(sdp_target_state: str):
service_states = get_service_state_list()
for service in service_states:
if service.current_state != sdp_target_state:
LOG.debug('Setting the current state of %s to be %s', service.
id, sdp_target_state)
service.update_current_state(sdp_target_state)
|
update_services_instant_gratification
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(sdp_target_state: str):\n service_states = get_service_state_list()\n for service in service_states:\n if service.current_state != sdp_target_state:\n LOG.debug('Setting the current state of %s to be %s', service.\n id, sdp_target_state)\n service.update_current_state(sdp_target_state)\n",
"role": "user"
},
{
"content": "update_services_instant_gratification",
"role": "assistant"
}
] |
def x(sdp_target_state: str):
service_states = get_service_state_list()
for service in service_states:
if service.current_state != sdp_target_state:
LOG.debug('Setting the target state of %s to be %s', service.id,
sdp_target_state)
service.update_target_state(sdp_target_state)
|
update_services_target_state
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(sdp_target_state: str):\n service_states = get_service_state_list()\n for service in service_states:\n if service.current_state != sdp_target_state:\n LOG.debug('Setting the target state of %s to be %s', service.id,\n sdp_target_state)\n service.update_target_state(sdp_target_state)\n",
"role": "user"
},
{
"content": "update_services_target_state",
"role": "assistant"
}
] |
def x(sdp_state: SDPState):
LOG.info('Handling SDP target state updated event...')
LOG.info('SDP target state: %s', sdp_state.target_state)
if sdp_state.target_state == 'off':
_update_services_target_state('off')
sdp_state.update_current_state(sdp_state.target_state)
|
handle_sdp_target_state_updated
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(sdp_state: SDPState):\n LOG.info('Handling SDP target state updated event...')\n LOG.info('SDP target state: %s', sdp_state.target_state)\n if sdp_state.target_state == 'off':\n _update_services_target_state('off')\n sdp_state.update_current_state(sdp_state.target_state)\n",
"role": "user"
},
{
"content": "handle_sdp_target_state_updated",
"role": "assistant"
}
] |
def x():
parser = argparse.ArgumentParser(description='{} service.'.format(
__service_id__))
parser.add_argument('--random_errors', action='store_true', help=
'Enable random errors')
parser.add_argument('-v', action='store_true', help=
'Verbose mode (enable debug printing)')
parser.add_argument('-vv', action='store_true', help='Extra verbose mode')
args = parser.parse_args()
if args.vv:
init_logger(log_level='DEBUG', show_log_origin=True)
elif args.v:
init_logger(logger_name='sip.ec.master_controller', log_level='DEBUG')
else:
init_logger(log_level='INFO')
return args
|
parse_args
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n parser = argparse.ArgumentParser(description='{} service.'.format(\n __service_id__))\n parser.add_argument('--random_errors', action='store_true', help=\n 'Enable random errors')\n parser.add_argument('-v', action='store_true', help=\n 'Verbose mode (enable debug printing)')\n parser.add_argument('-vv', action='store_true', help='Extra verbose mode')\n args = parser.parse_args()\n if args.vv:\n init_logger(log_level='DEBUG', show_log_origin=True)\n elif args.v:\n init_logger(logger_name='sip.ec.master_controller', log_level='DEBUG')\n else:\n init_logger(log_level='INFO')\n return args\n",
"role": "user"
},
{
"content": "parse_args",
"role": "assistant"
}
] |
def x(sdp_state: SDPState):
LOG.info('Initialising: %s', __service_id__)
_services = ['ExecutionControl:AlarmReceiver:1.0.0',
'ExecutionControl:AlertManager:1.0.0',
'ExecutionControl:ConfigurationDatabase:5.0.1',
'ExecutionControl:MasterController:1.3.0',
'ExecutionControl:ProcessingController:1.2.6',
'ExecutionControl:ProcessingBlockController:1.3.0',
'TangoControl:Database:1.0.4', 'TangoControl:MySQL:1.0.3',
'TangoControl:SDPMaster:1.2.1', 'TangoControl:Subarrays:1.2.0',
'TangoControl:ProcessingBlocks:1.2.0', 'Platform:Kafka:2.1.1',
'Platform:Prometheus:1.0.0', 'Platform:PrometheusPushGateway:0.7.0',
'Platform:RedisCommander:210.0.0', 'Platform:Zookeeper:3.4.13']
for service_id in _services:
subsystem, name, version = service_id.split(':')
ServiceState(subsystem, name, version)
if sdp_state.current_state in ['unknown', 'off']:
try:
LOG.info("Setting the SDPState to 'init'")
sdp_state.update_current_state('init', force=True)
except ValueError as error:
LOG.critical('Unable to set the State of SDP to init! %s', str(
error))
LOG.info('Updating Service States')
service_state_list = get_service_state_list()
for service_state in service_state_list:
if service_state.current_state in ['unknown', 'off']:
service_state.update_current_state('init', force=True)
for service_state in service_state_list:
if service_state.current_state == 'init':
time.sleep(random.uniform(0, 0.2))
service_state.update_current_state('on')
services_on = [(service.current_state == 'on') for service in
service_state_list]
if all(services_on):
LOG.info('All Services are online!.')
sdp_state.update_current_state('standby')
else:
LOG.critical('Master Controller failed to initialise.')
return service_state_list
|
init
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(sdp_state: SDPState):\n LOG.info('Initialising: %s', __service_id__)\n _services = ['ExecutionControl:AlarmReceiver:1.0.0',\n 'ExecutionControl:AlertManager:1.0.0',\n 'ExecutionControl:ConfigurationDatabase:5.0.1',\n 'ExecutionControl:MasterController:1.3.0',\n 'ExecutionControl:ProcessingController:1.2.6',\n 'ExecutionControl:ProcessingBlockController:1.3.0',\n 'TangoControl:Database:1.0.4', 'TangoControl:MySQL:1.0.3',\n 'TangoControl:SDPMaster:1.2.1', 'TangoControl:Subarrays:1.2.0',\n 'TangoControl:ProcessingBlocks:1.2.0', 'Platform:Kafka:2.1.1',\n 'Platform:Prometheus:1.0.0', 'Platform:PrometheusPushGateway:0.7.0',\n 'Platform:RedisCommander:210.0.0', 'Platform:Zookeeper:3.4.13']\n for service_id in _services:\n subsystem, name, version = service_id.split(':')\n ServiceState(subsystem, name, version)\n if sdp_state.current_state in ['unknown', 'off']:\n try:\n LOG.info(\"Setting the SDPState to 'init'\")\n sdp_state.update_current_state('init', force=True)\n except ValueError as error:\n LOG.critical('Unable to set the State of SDP to init! %s', str(\n error))\n LOG.info('Updating Service States')\n service_state_list = get_service_state_list()\n for service_state in service_state_list:\n if service_state.current_state in ['unknown', 'off']:\n service_state.update_current_state('init', force=True)\n for service_state in service_state_list:\n if service_state.current_state == 'init':\n time.sleep(random.uniform(0, 0.2))\n service_state.update_current_state('on')\n services_on = [(service.current_state == 'on') for service in\n service_state_list]\n if all(services_on):\n LOG.info('All Services are online!.')\n sdp_state.update_current_state('standby')\n else:\n LOG.critical('Master Controller failed to initialise.')\n return service_state_list\n",
"role": "user"
},
{
"content": "init",
"role": "assistant"
}
] |
def x(event: Event, sdp_state: SDPState, service_states: List[ServiceState]):
LOG.debug('Event detected! (id : "%s", type: "%s", data: "%s")', event.
object_id, event.type, event.data)
if event.object_id == 'SDP' and event.type == 'current_state_updated':
LOG.info('SDP current state updated, no action required!')
if event.object_id == 'SDP' and event.type == 'target_state_updated':
LOG.info("SDP target state changed to '%s'", sdp_state.target_state)
if sdp_state.target_state == sdp_state.current_state:
LOG.warning('SDP already in %s state', sdp_state.current_state)
return
if not sdp_state.is_target_state_allowed(sdp_state.target_state):
LOG.error('Transition to %s is not allowed when in state %s',
sdp_state.target_state, sdp_state.current_state)
sdp_state.target_state = sdp_state.current_state
return
_update_services_target_state(sdp_state.target_state)
if sdp_state.target_state == 'off':
LOG.info('Turning off services!')
for service_state in service_states:
service_state.update_target_state('off')
service_state.update_current_state('off')
LOG.info('Processing target state change request ...')
time.sleep(0.1)
LOG.info('Done processing target state change request!')
sdp_state.update_current_state(sdp_state.target_state)
if sdp_state.current_state == 'alarm':
LOG.debug('raising SDP state alarm')
SIP_STATE_ALARM.set(1)
else:
SIP_STATE_ALARM.set(0)
try:
push_to_gateway('platform_pushgateway:9091', job='SIP',
registry=COLLECTOR_REGISTRY)
except urllib.error.URLError:
LOG.warning('Unable to connect to the Alarms service!')
|
process_event
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(event: Event, sdp_state: SDPState, service_states: List[ServiceState]):\n LOG.debug('Event detected! (id : \"%s\", type: \"%s\", data: \"%s\")', event.\n object_id, event.type, event.data)\n if event.object_id == 'SDP' and event.type == 'current_state_updated':\n LOG.info('SDP current state updated, no action required!')\n if event.object_id == 'SDP' and event.type == 'target_state_updated':\n LOG.info(\"SDP target state changed to '%s'\", sdp_state.target_state)\n if sdp_state.target_state == sdp_state.current_state:\n LOG.warning('SDP already in %s state', sdp_state.current_state)\n return\n if not sdp_state.is_target_state_allowed(sdp_state.target_state):\n LOG.error('Transition to %s is not allowed when in state %s',\n sdp_state.target_state, sdp_state.current_state)\n sdp_state.target_state = sdp_state.current_state\n return\n _update_services_target_state(sdp_state.target_state)\n if sdp_state.target_state == 'off':\n LOG.info('Turning off services!')\n for service_state in service_states:\n service_state.update_target_state('off')\n service_state.update_current_state('off')\n LOG.info('Processing target state change request ...')\n time.sleep(0.1)\n LOG.info('Done processing target state change request!')\n sdp_state.update_current_state(sdp_state.target_state)\n if sdp_state.current_state == 'alarm':\n LOG.debug('raising SDP state alarm')\n SIP_STATE_ALARM.set(1)\n else:\n SIP_STATE_ALARM.set(0)\n try:\n push_to_gateway('platform_pushgateway:9091', job='SIP',\n registry=COLLECTOR_REGISTRY)\n except urllib.error.URLError:\n LOG.warning('Unable to connect to the Alarms service!')\n",
"role": "user"
},
{
"content": "process_event",
"role": "assistant"
}
] |
def x():
sdp_state = SDPState()
service_states = get_service_state_list()
state_events = sdp_state.get_event_queue(subscriber=__service_name__)
state_is_off = sdp_state.current_state == 'off'
counter = 0
while True:
time.sleep(0.1)
if not state_is_off:
if counter % 1000 == 0:
LOG.debug('Checking published events ... %d', counter / 1000)
_published_events = state_events.get_published_events(process
=True)
for _state_event in _published_events:
_process_event(_state_event, sdp_state, service_states)
else:
_state_event = state_events.get()
if _state_event:
_process_event(_state_event, sdp_state, service_states)
state_is_off = sdp_state.current_state == 'off'
counter += 1
|
process_state_change_events
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n sdp_state = SDPState()\n service_states = get_service_state_list()\n state_events = sdp_state.get_event_queue(subscriber=__service_name__)\n state_is_off = sdp_state.current_state == 'off'\n counter = 0\n while True:\n time.sleep(0.1)\n if not state_is_off:\n if counter % 1000 == 0:\n LOG.debug('Checking published events ... %d', counter / 1000)\n _published_events = state_events.get_published_events(process\n =True)\n for _state_event in _published_events:\n _process_event(_state_event, sdp_state, service_states)\n else:\n _state_event = state_events.get()\n if _state_event:\n _process_event(_state_event, sdp_state, service_states)\n state_is_off = sdp_state.current_state == 'off'\n counter += 1\n",
"role": "user"
},
{
"content": "process_state_change_events",
"role": "assistant"
}
] |
def x():
_parse_args()
LOG.info('Starting: %s', __service_id__)
LOG.info('Subscribing to state change events (subscriber = %s)',
__service_name__)
sdp_state = SDPState()
_ = sdp_state.subscribe(subscriber=__service_name__)
_ = _init(sdp_state)
LOG.info('Finished initialising!')
LOG.info('Responding to state change events ...')
try:
_process_state_change_events()
except ValueError as error:
LOG.critical('Value error: %s', str(error))
except KeyboardInterrupt as err:
LOG.debug('Keyboard Interrupt %s', err)
LOG.info('Exiting!')
|
main
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n _parse_args()\n LOG.info('Starting: %s', __service_id__)\n LOG.info('Subscribing to state change events (subscriber = %s)',\n __service_name__)\n sdp_state = SDPState()\n _ = sdp_state.subscribe(subscriber=__service_name__)\n _ = _init(sdp_state)\n LOG.info('Finished initialising!')\n LOG.info('Responding to state change events ...')\n try:\n _process_state_change_events()\n except ValueError as error:\n LOG.critical('Value error: %s', str(error))\n except KeyboardInterrupt as err:\n LOG.debug('Keyboard Interrupt %s', err)\n LOG.info('Exiting!')\n",
"role": "user"
},
{
"content": "main",
"role": "assistant"
}
] |
def x():
stream_config = spead2.send.StreamConfig(max_packet_size=16356, rate=
1000000000.0, burst_size=10, max_heaps=1)
item_group = spead2.send.ItemGroup(flavour=spead2.Flavour(4, 64, 48, 0))
num_baselines = 512 * 513 // 2
dtype = [('TCI', 'i1'), ('FD', 'u1'), ('VIS', '<c8', 4)]
item_group.add_item(id=24576, name='visibility_timestamp_count',
description='', shape=tuple(), format=None, dtype='<u4')
item_group.add_item(id=24577, name='visibility_timestamp_fraction',
description='', shape=tuple(), format=None, dtype='<u4')
item_group.add_item(id=24581, name='visibility_baseline_count',
description='', shape=tuple(), format=None, dtype='<u4')
item_group.add_item(id=24584, name='scan_id', description='', shape=
tuple(), format=None, dtype='<u8')
item_group.add_item(id=24586, name='correlator_output_data',
description='', shape=(num_baselines,), dtype=dtype)
streams = []
num_streams = 2
for i in range(num_streams):
stream = spead2.send.UdpStream(thread_pool=spead2.ThreadPool(
threads=1), hostname='127.0.0.1', port=41000 + i, config=
stream_config)
stream.send_heap(item_group.get_start())
streams.append(stream)
vis = numpy.zeros(shape=(num_baselines,), dtype=dtype)
num_heaps = 200
start_time = time.time()
for stream in streams:
item_group['visibility_timestamp_count'].value = 1
item_group['visibility_timestamp_fraction'].value = 0
item_group['visibility_baseline_count'].value = num_baselines
item_group['scan_id'].value = 100000000
item_group['correlator_output_data'].value = vis
for i in range(num_heaps):
stream.send_heap(item_group.get_heap(descriptors='all', data='all')
)
duration = time.time() - start_time
data_size = num_streams * num_heaps * (vis.nbytes / 1000000.0)
print('Sent %.3f MB in %.3f sec (%.3f MB/sec)' % (data_size, duration,
data_size / duration))
for stream in streams:
stream.send_heap(item_group.get_end())
|
main
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n stream_config = spead2.send.StreamConfig(max_packet_size=16356, rate=\n 1000000000.0, burst_size=10, max_heaps=1)\n item_group = spead2.send.ItemGroup(flavour=spead2.Flavour(4, 64, 48, 0))\n num_baselines = 512 * 513 // 2\n dtype = [('TCI', 'i1'), ('FD', 'u1'), ('VIS', '<c8', 4)]\n item_group.add_item(id=24576, name='visibility_timestamp_count',\n description='', shape=tuple(), format=None, dtype='<u4')\n item_group.add_item(id=24577, name='visibility_timestamp_fraction',\n description='', shape=tuple(), format=None, dtype='<u4')\n item_group.add_item(id=24581, name='visibility_baseline_count',\n description='', shape=tuple(), format=None, dtype='<u4')\n item_group.add_item(id=24584, name='scan_id', description='', shape=\n tuple(), format=None, dtype='<u8')\n item_group.add_item(id=24586, name='correlator_output_data',\n description='', shape=(num_baselines,), dtype=dtype)\n streams = []\n num_streams = 2\n for i in range(num_streams):\n stream = spead2.send.UdpStream(thread_pool=spead2.ThreadPool(\n threads=1), hostname='127.0.0.1', port=41000 + i, config=\n stream_config)\n stream.send_heap(item_group.get_start())\n streams.append(stream)\n vis = numpy.zeros(shape=(num_baselines,), dtype=dtype)\n num_heaps = 200\n start_time = time.time()\n for stream in streams:\n item_group['visibility_timestamp_count'].value = 1\n item_group['visibility_timestamp_fraction'].value = 0\n item_group['visibility_baseline_count'].value = num_baselines\n item_group['scan_id'].value = 100000000\n item_group['correlator_output_data'].value = vis\n for i in range(num_heaps):\n stream.send_heap(item_group.get_heap(descriptors='all', data='all')\n )\n duration = time.time() - start_time\n data_size = num_streams * num_heaps * (vis.nbytes / 1000000.0)\n print('Sent %.3f MB in %.3f sec (%.3f MB/sec)' % (data_size, duration, \n data_size / duration))\n for stream in streams:\n stream.send_heap(item_group.get_end())\n",
"role": "user"
},
{
"content": "main",
"role": "assistant"
}
] |
def x(cls, config: dict):
timestamp = config.get('timestamp', None)
return cls(config.get('id'), config.get('type'), config.get('data',
dict()), config.get('origin', None), timestamp, config.get(
'object_type', None), config.get('object_id', None), config.get(
'object_key', None))
|
from_config
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, config: dict):\n timestamp = config.get('timestamp', None)\n return cls(config.get('id'), config.get('type'), config.get('data',\n dict()), config.get('origin', None), timestamp, config.get(\n 'object_type', None), config.get('object_id', None), config.get(\n 'object_key', None))\n",
"role": "user"
},
{
"content": "from_config",
"role": "assistant"
}
] |
def x(filename, imager, grid_data, grid_norm, grid_weights):
ms = oskar.MeasurementSet.open(filename)
block_start = 0
num_rows = ms.num_rows
num_baselines = ms.num_stations * (ms.num_stations - 1) // 2
while block_start < num_rows:
block_size = num_rows - block_start
if block_size > num_baselines:
block_size = num_baselines
uvw = ms.read_column('UVW', block_start, block_size)
vis_weights = ms.read_column('WEIGHT', block_start, block_size)
if ms.num_pols == 4:
vis_weights = 0.5 * (vis_weights[:, 0] + vis_weights[:, 3])
for j in range(ms.num_channels):
coords = uvw * (ms.freq_start_hz + j * ms.freq_inc_hz
) / 299792458.0
vis_data = None
if not imager.coords_only:
vis_data = ms.read_vis(block_start, j, 1, block_size)
if ms.num_pols == 4:
vis_data = 0.5 * (vis_data[0, :, 0] + vis_data[0, :, 3])
grid_norm = imager.update_plane(coords[:, 0], coords[:, 1],
coords[:, 2], vis_data, vis_weights, grid_data, grid_norm,
grid_weights)
block_start += block_size
return grid_norm
|
process_input_data
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(filename, imager, grid_data, grid_norm, grid_weights):\n ms = oskar.MeasurementSet.open(filename)\n block_start = 0\n num_rows = ms.num_rows\n num_baselines = ms.num_stations * (ms.num_stations - 1) // 2\n while block_start < num_rows:\n block_size = num_rows - block_start\n if block_size > num_baselines:\n block_size = num_baselines\n uvw = ms.read_column('UVW', block_start, block_size)\n vis_weights = ms.read_column('WEIGHT', block_start, block_size)\n if ms.num_pols == 4:\n vis_weights = 0.5 * (vis_weights[:, 0] + vis_weights[:, 3])\n for j in range(ms.num_channels):\n coords = uvw * (ms.freq_start_hz + j * ms.freq_inc_hz\n ) / 299792458.0\n vis_data = None\n if not imager.coords_only:\n vis_data = ms.read_vis(block_start, j, 1, block_size)\n if ms.num_pols == 4:\n vis_data = 0.5 * (vis_data[0, :, 0] + vis_data[0, :, 3])\n grid_norm = imager.update_plane(coords[:, 0], coords[:, 1],\n coords[:, 2], vis_data, vis_weights, grid_data, grid_norm,\n grid_weights)\n block_start += block_size\n return grid_norm\n",
"role": "user"
},
{
"content": "process_input_data",
"role": "assistant"
}
] |
def x():
if len(sys.argv) < 2:
raise RuntimeError(
'Usage: mpiexec -n <np> python mpi_imager_test.py <settings_file> <dir>'
)
comm = MPI.COMM_WORLD
settings = None
inputs = None
grid_weights = None
log = logging.getLogger()
log.setLevel(logging.DEBUG)
if len(log.handlers) == 0:
log.addHandler(logging.StreamHandler(sys.stdout))
if comm.Get_rank() == 0:
with open(sys.argv[1]) as f:
settings = json.load(f)
data_dir = str(sys.argv[2])
inputs = glob(join(data_dir, '*.ms')) + glob(join(data_dir, '*.MS'))
inputs = filter(None, inputs)
log.info('Found input Measurement Sets: %s', ', '.join(inputs))
inputs = chunks(inputs, comm.Get_size())
comm.barrier()
settings = comm.bcast(settings)
inputs = comm.scatter(inputs)
log.debug('Rank %d, processing [%s]', comm.Get_rank(), ', '.join(inputs))
precision = settings['precision']
imager = oskar.Imager(precision)
for key, value in settings['imager'].items():
setattr(imager, key, value)
grid_norm = 0.0
grid_dim = [imager.plane_size, imager.plane_size]
grid_data = numpy.zeros(grid_dim, dtype='c8' if precision == 'single' else
'c16')
if settings['combine']:
if imager.weighting == 'Uniform' or imager.algorithm == 'W-projection':
local_weights = None
if imager.weighting == 'Uniform':
grid_weights = numpy.zeros(grid_dim, dtype=precision)
local_weights = numpy.zeros(grid_dim, dtype=precision)
imager.coords_only = True
for f in inputs:
log.info('Reading coordinates from %s', f)
process_input_data(f, imager, None, 0.0, local_weights)
imager.coords_only = False
num_w_planes = imager.num_w_planes
num_w_planes = comm.allreduce(num_w_planes, op=MPI.MAX)
imager.num_w_planes = num_w_planes
if local_weights is not None:
comm.Allreduce(local_weights, grid_weights, op=MPI.SUM)
for f in inputs:
log.info('Reading visibilities from %s', f)
grid_norm = process_input_data(f, imager, grid_data, grid_norm,
grid_weights)
grid = numpy.zeros_like(grid_data)
comm.Reduce(grid_data, grid, op=MPI.SUM)
grid_norm = comm.reduce(grid_norm, op=MPI.SUM)
if comm.Get_rank() == 0:
save_image(imager, grid, grid_norm, settings['output_file'])
log.info('Finished. Output file is %s', settings['output_file'])
else:
for f in inputs:
grid_norm = 0.0
grid_data.fill(0)
if imager.weighting == 'Uniform':
grid_weights = numpy.zeros(grid_dim, dtype=precision)
if (imager.weighting == 'Uniform' or imager.algorithm ==
'W-projection'):
imager.coords_only = True
log.info('Reading coordinates from %s', f)
process_input_data(f, imager, None, 0.0, grid_weights)
imager.coords_only = False
log.info('Reading visibilities from %s', f)
grid_norm = process_input_data(f, imager, grid_data, grid_norm,
grid_weights)
output_file = splitext(f)[0] + '.fits'
save_image(imager, grid_data, grid_norm, output_file)
log.info('Finished. Output file is %s', output_file)
|
main
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n if len(sys.argv) < 2:\n raise RuntimeError(\n 'Usage: mpiexec -n <np> python mpi_imager_test.py <settings_file> <dir>'\n )\n comm = MPI.COMM_WORLD\n settings = None\n inputs = None\n grid_weights = None\n log = logging.getLogger()\n log.setLevel(logging.DEBUG)\n if len(log.handlers) == 0:\n log.addHandler(logging.StreamHandler(sys.stdout))\n if comm.Get_rank() == 0:\n with open(sys.argv[1]) as f:\n settings = json.load(f)\n data_dir = str(sys.argv[2])\n inputs = glob(join(data_dir, '*.ms')) + glob(join(data_dir, '*.MS'))\n inputs = filter(None, inputs)\n log.info('Found input Measurement Sets: %s', ', '.join(inputs))\n inputs = chunks(inputs, comm.Get_size())\n comm.barrier()\n settings = comm.bcast(settings)\n inputs = comm.scatter(inputs)\n log.debug('Rank %d, processing [%s]', comm.Get_rank(), ', '.join(inputs))\n precision = settings['precision']\n imager = oskar.Imager(precision)\n for key, value in settings['imager'].items():\n setattr(imager, key, value)\n grid_norm = 0.0\n grid_dim = [imager.plane_size, imager.plane_size]\n grid_data = numpy.zeros(grid_dim, dtype='c8' if precision == 'single' else\n 'c16')\n if settings['combine']:\n if imager.weighting == 'Uniform' or imager.algorithm == 'W-projection':\n local_weights = None\n if imager.weighting == 'Uniform':\n grid_weights = numpy.zeros(grid_dim, dtype=precision)\n local_weights = numpy.zeros(grid_dim, dtype=precision)\n imager.coords_only = True\n for f in inputs:\n log.info('Reading coordinates from %s', f)\n process_input_data(f, imager, None, 0.0, local_weights)\n imager.coords_only = False\n num_w_planes = imager.num_w_planes\n num_w_planes = comm.allreduce(num_w_planes, op=MPI.MAX)\n imager.num_w_planes = num_w_planes\n if local_weights is not None:\n comm.Allreduce(local_weights, grid_weights, op=MPI.SUM)\n for f in inputs:\n log.info('Reading visibilities from %s', f)\n grid_norm = process_input_data(f, imager, grid_data, grid_norm,\n grid_weights)\n grid = numpy.zeros_like(grid_data)\n comm.Reduce(grid_data, grid, op=MPI.SUM)\n grid_norm = comm.reduce(grid_norm, op=MPI.SUM)\n if comm.Get_rank() == 0:\n save_image(imager, grid, grid_norm, settings['output_file'])\n log.info('Finished. Output file is %s', settings['output_file'])\n else:\n for f in inputs:\n grid_norm = 0.0\n grid_data.fill(0)\n if imager.weighting == 'Uniform':\n grid_weights = numpy.zeros(grid_dim, dtype=precision)\n if (imager.weighting == 'Uniform' or imager.algorithm ==\n 'W-projection'):\n imager.coords_only = True\n log.info('Reading coordinates from %s', f)\n process_input_data(f, imager, None, 0.0, grid_weights)\n imager.coords_only = False\n log.info('Reading visibilities from %s', f)\n grid_norm = process_input_data(f, imager, grid_data, grid_norm,\n grid_weights)\n output_file = splitext(f)[0] + '.fits'\n save_image(imager, grid_data, grid_norm, output_file)\n log.info('Finished. Output file is %s', output_file)\n",
"role": "user"
},
{
"content": "main",
"role": "assistant"
}
] |
def x():
LOG.debug('GET list of SBIs.')
_url = get_root_url()
response = dict(scheduling_blocks=[], links=dict(home='{}'.format(_url)))
block_ids = DB.get_sched_block_instance_ids()
for block in DB.get_block_details(block_ids):
block_id = block['id']
LOG.debug('Adding SBI %s to list', block_id)
LOG.debug(block)
block['num_processing_blocks'] = len(block['processing_block_ids'])
temp = ['OK'] * 10 + ['WAITING'] * 4 + ['FAILED'] * 2
block['status'] = choice(temp)
try:
del block['processing_block_ids']
except KeyError:
pass
block['links'] = {'detail': '{}/scheduling-block/{}'.format(_url,
block_id)}
response['scheduling_blocks'].append(block)
return response, HTTPStatus.OK
|
get
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n LOG.debug('GET list of SBIs.')\n _url = get_root_url()\n response = dict(scheduling_blocks=[], links=dict(home='{}'.format(_url)))\n block_ids = DB.get_sched_block_instance_ids()\n for block in DB.get_block_details(block_ids):\n block_id = block['id']\n LOG.debug('Adding SBI %s to list', block_id)\n LOG.debug(block)\n block['num_processing_blocks'] = len(block['processing_block_ids'])\n temp = ['OK'] * 10 + ['WAITING'] * 4 + ['FAILED'] * 2\n block['status'] = choice(temp)\n try:\n del block['processing_block_ids']\n except KeyError:\n pass\n block['links'] = {'detail': '{}/scheduling-block/{}'.format(_url,\n block_id)}\n response['scheduling_blocks'].append(block)\n return response, HTTPStatus.OK\n",
"role": "user"
},
{
"content": "get",
"role": "assistant"
}
] |
def x():
response = dict(blocks=[])
block_ids = DB.get_sched_block_instance_ids()
for index, block_id in enumerate(block_ids):
block = DB.get_block_details([block_id]).__next__()
info = [index, block['id'], block['sub_array_id'], len(block[
'processing_blocks'])]
response['blocks'].append(info)
return response, HTTPStatus.OK
|
get_table
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n response = dict(blocks=[])\n block_ids = DB.get_sched_block_instance_ids()\n for index, block_id in enumerate(block_ids):\n block = DB.get_block_details([block_id]).__next__()\n info = [index, block['id'], block['sub_array_id'], len(block[\n 'processing_blocks'])]\n response['blocks'].append(info)\n return response, HTTPStatus.OK\n",
"role": "user"
},
{
"content": "get_table",
"role": "assistant"
}
] |
def x(self, key, field, value):
self._db.hset(key, field, value)
|
set_value
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, key, field, value):\n self._db.hset(key, field, value)\n",
"role": "user"
},
{
"content": "set_value",
"role": "assistant"
}
] |
def x(self, event_name, event_type, block_id):
self._db.rpush(event_name, dict(type=event_type, id=block_id))
|
push_event
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, event_name, event_type, block_id):\n self._db.rpush(event_name, dict(type=event_type, id=block_id))\n",
"role": "user"
},
{
"content": "push_event",
"role": "assistant"
}
] |
def x():
signal.signal(signal.SIGTERM, _sig_handler)
with open(sys.argv[1]) as fh:
config = json.load(fh)
os.chdir(os.path.expanduser('~'))
receiver = PulsarStart(config, logging.getLogger())
receiver.run()
|
main
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n signal.signal(signal.SIGTERM, _sig_handler)\n with open(sys.argv[1]) as fh:\n config = json.load(fh)\n os.chdir(os.path.expanduser('~'))\n receiver = PulsarStart(config, logging.getLogger())\n receiver.run()\n",
"role": "user"
},
{
"content": "main",
"role": "assistant"
}
] |
def x(func):
@wraps(func)
def x(*args, **kwargs):
try:
return func(*args, **kwargs)
except redis.exceptions.ConnectionError:
raise ConnectionError(
'Unable to connect to the Redis Configuration Database. host = {}, port = {}, id = {}.'
.format(REDIS_HOST, REDIS_PORT, REDIS_DB_ID))
return with_exception_handling
|
check_connection
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(func):\n\n @wraps(func)\n def x(*args, **kwargs):\n try:\n return func(*args, **kwargs)\n except redis.exceptions.ConnectionError:\n raise ConnectionError(\n 'Unable to connect to the Redis Configuration Database. host = {}, port = {}, id = {}.'\n .format(REDIS_HOST, REDIS_PORT, REDIS_DB_ID))\n return with_exception_handling\n",
"role": "user"
},
{
"content": "check_connection",
"role": "assistant"
}
] |
def x(self, key: str, my_dict: dict, hierarchical: bool=False):
for _key, _value in my_dict.items():
if isinstance(_value, dict):
if not hierarchical:
self._db.hmset(key, {_key: json.dumps(_value)})
else:
self.save_dict(key + ':' + _key, _value, hierarchical)
elif isinstance(_value, list):
if not hierarchical:
self._db.hmset(key, {_key: str(_value)})
else:
print('saving list at ', key + ':' + _key)
self._db.lpush(key + ':' + _key, *_value[::-1])
elif isinstance(_value, bool):
self._db.hmset(key, {_key: str(_value)})
else:
self._db.hmset(key, {_key: _value})
|
save_dict
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, key: str, my_dict: dict, hierarchical: bool=False):\n for _key, _value in my_dict.items():\n if isinstance(_value, dict):\n if not hierarchical:\n self._db.hmset(key, {_key: json.dumps(_value)})\n else:\n self.save_dict(key + ':' + _key, _value, hierarchical)\n elif isinstance(_value, list):\n if not hierarchical:\n self._db.hmset(key, {_key: str(_value)})\n else:\n print('saving list at ', key + ':' + _key)\n self._db.lpush(key + ':' + _key, *_value[::-1])\n elif isinstance(_value, bool):\n self._db.hmset(key, {_key: str(_value)})\n else:\n self._db.hmset(key, {_key: _value})\n",
"role": "user"
},
{
"content": "save_dict",
"role": "assistant"
}
] |
def x(my_dict, keys, values):
temp = my_dict
for depth, key in enumerate(keys):
if depth < len(keys) - 1:
if key not in temp:
temp[key] = dict()
temp = temp[key]
elif key not in temp:
temp[key] = values
else:
temp[key] = {**temp[key], **values}
return my_dict
|
build_dict
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(my_dict, keys, values):\n temp = my_dict\n for depth, key in enumerate(keys):\n if depth < len(keys) - 1:\n if key not in temp:\n temp[key] = dict()\n temp = temp[key]\n elif key not in temp:\n temp[key] = values\n else:\n temp[key] = {**temp[key], **values}\n return my_dict\n",
"role": "user"
},
{
"content": "build_dict",
"role": "assistant"
}
] |
def x(self, db_key: str) ->dict:
if self._db.type(db_key) == 'list':
db_values = self._db.lrange(db_key, 0, -1)
for i, value in enumerate(db_values):
try:
db_values[i] = ast.literal_eval(value)
except SyntaxError:
pass
except ValueError:
pass
else:
db_values = self._db.hgetall(db_key)
for _key, _value in db_values.items():
try:
db_values[_key] = ast.literal_eval(_value)
except SyntaxError:
pass
except ValueError:
pass
return db_values
|
load_values
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, db_key: str) ->dict:\n if self._db.type(db_key) == 'list':\n db_values = self._db.lrange(db_key, 0, -1)\n for i, value in enumerate(db_values):\n try:\n db_values[i] = ast.literal_eval(value)\n except SyntaxError:\n pass\n except ValueError:\n pass\n else:\n db_values = self._db.hgetall(db_key)\n for _key, _value in db_values.items():\n try:\n db_values[_key] = ast.literal_eval(_value)\n except SyntaxError:\n pass\n except ValueError:\n pass\n return db_values\n",
"role": "user"
},
{
"content": "load_values",
"role": "assistant"
}
] |
def x(self, db_key: str) ->dict:
db_keys = self._db.keys(pattern=db_key + '*')
my_dict = {}
for _db_key in db_keys:
if self._db.type(_db_key) == 'list':
db_values = self._db.lrange(_db_key, 0, -1)
for i, value in enumerate(db_values):
try:
db_values[i] = ast.literal_eval(value)
except SyntaxError:
pass
except ValueError:
pass
else:
db_values = self._db.hgetall(_db_key)
for _key, _value in db_values.items():
try:
db_values[_key] = ast.literal_eval(_value)
except SyntaxError:
pass
except ValueError:
pass
my_dict = self._build_dict(my_dict, _db_key.split(':'), db_values)
return my_dict[db_key]
|
load_dict_hierarchical
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, db_key: str) ->dict:\n db_keys = self._db.keys(pattern=db_key + '*')\n my_dict = {}\n for _db_key in db_keys:\n if self._db.type(_db_key) == 'list':\n db_values = self._db.lrange(_db_key, 0, -1)\n for i, value in enumerate(db_values):\n try:\n db_values[i] = ast.literal_eval(value)\n except SyntaxError:\n pass\n except ValueError:\n pass\n else:\n db_values = self._db.hgetall(_db_key)\n for _key, _value in db_values.items():\n try:\n db_values[_key] = ast.literal_eval(_value)\n except SyntaxError:\n pass\n except ValueError:\n pass\n my_dict = self._build_dict(my_dict, _db_key.split(':'), db_values)\n return my_dict[db_key]\n",
"role": "user"
},
{
"content": "load_dict_hierarchical",
"role": "assistant"
}
] |
def x(self, db_key: str, hierarchical: bool=False) ->dict:
if not hierarchical:
db_values = self._db.hgetall(db_key)
for _key, _value in db_values.items():
if isinstance(_value, str):
db_values[_key] = ast.literal_eval(_value)
my_dict = db_values
else:
my_dict = self._load_dict_hierarchical(db_key)
return my_dict
|
load_dict
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, db_key: str, hierarchical: bool=False) ->dict:\n if not hierarchical:\n db_values = self._db.hgetall(db_key)\n for _key, _value in db_values.items():\n if isinstance(_value, str):\n db_values[_key] = ast.literal_eval(_value)\n my_dict = db_values\n else:\n my_dict = self._load_dict_hierarchical(db_key)\n return my_dict\n",
"role": "user"
},
{
"content": "load_dict",
"role": "assistant"
}
] |
def x(self, db_key: str, dict_keys: List[str], hierarchical: bool=False
) ->List:
result = []
if not hierarchical:
_values = self._db.hmget(db_key, *dict_keys)
result = [ast.literal_eval(_value) for _value in _values]
else:
db_keys = self._db.keys(pattern=db_key + '*')
for _db_key in db_keys:
for name in _db_key.split(':')[1:]:
if name in dict_keys:
_values = self._load_values(_db_key)
result.append(_values)
_values = self._db.hmget(_db_key, *dict_keys)
for i, value in enumerate(_values):
try:
_values[i] = ast.literal_eval(value)
except SyntaxError:
pass
except ValueError:
pass
result += [value for value in _values if value is not None]
return result
|
load_dict_values
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, db_key: str, dict_keys: List[str], hierarchical: bool=False\n ) ->List:\n result = []\n if not hierarchical:\n _values = self._db.hmget(db_key, *dict_keys)\n result = [ast.literal_eval(_value) for _value in _values]\n else:\n db_keys = self._db.keys(pattern=db_key + '*')\n for _db_key in db_keys:\n for name in _db_key.split(':')[1:]:\n if name in dict_keys:\n _values = self._load_values(_db_key)\n result.append(_values)\n _values = self._db.hmget(_db_key, *dict_keys)\n for i, value in enumerate(_values):\n try:\n _values[i] = ast.literal_eval(value)\n except SyntaxError:\n pass\n except ValueError:\n pass\n result += [value for value in _values if value is not None]\n return result\n",
"role": "user"
},
{
"content": "load_dict_values",
"role": "assistant"
}
] |
def x(self, key, field, value, pipeline=False):
if pipeline:
self._pipeline.hset(key, field, str(value))
else:
self._db.hset(key, field, str(value))
|
set_hash_value
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, key, field, value, pipeline=False):\n if pipeline:\n self._pipeline.hset(key, field, str(value))\n else:\n self._db.hset(key, field, str(value))\n",
"role": "user"
},
{
"content": "set_hash_value",
"role": "assistant"
}
] |
def x(self, key, *value, pipeline=False):
if pipeline:
self._pipeline.lpush(key, *value)
else:
self._db.lpush(key, *value)
|
prepend_to_list
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, key, *value, pipeline=False):\n if pipeline:\n self._pipeline.lpush(key, *value)\n else:\n self._db.lpush(key, *value)\n",
"role": "user"
},
{
"content": "prepend_to_list",
"role": "assistant"
}
] |
def x(self, key, *value, pipeline=False):
if pipeline:
self._pipeline.rpush(key, *value)
else:
self._db.rpush(key, *value)
|
append_to_list
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, key, *value, pipeline=False):\n if pipeline:\n self._pipeline.rpush(key, *value)\n else:\n self._db.rpush(key, *value)\n",
"role": "user"
},
{
"content": "append_to_list",
"role": "assistant"
}
] |
def x(self, key, pipeline=False):
if pipeline:
return self._pipeline.lrange(key, 0, -1)
return self._db.lrange(key, 0, -1)
|
get_list
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, key, pipeline=False):\n if pipeline:\n return self._pipeline.lrange(key, 0, -1)\n return self._db.lrange(key, 0, -1)\n",
"role": "user"
},
{
"content": "get_list",
"role": "assistant"
}
] |
def x(self, *names: str, pipeline=False):
if pipeline:
self._pipeline.delete(*names)
else:
self._db.delete(*names)
|
delete
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, *names: str, pipeline=False):\n if pipeline:\n self._pipeline.delete(*names)\n else:\n self._db.delete(*names)\n",
"role": "user"
},
{
"content": "delete",
"role": "assistant"
}
] |
def x(self, event_name, event_history=None):
if event_history is None:
event_history = event_name + '_history'
return self._db.rpoplpush(event_name, event_history)
|
get_event
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, event_name, event_history=None):\n if event_history is None:\n event_history = event_name + '_history'\n return self._db.rpoplpush(event_name, event_history)\n",
"role": "user"
},
{
"content": "get_event",
"role": "assistant"
}
] |
def x(self, key: str, value, count: int=0, pipeline: bool=False):
if pipeline:
if redis.__version__ == '2.10.6':
self._pipeline.lrem(name=key, value=value, num=count)
else:
self._pipeline.lrem(key, count, value)
elif self._db.exists(key):
if redis.__version__ == '2.10.6':
self._db.lrem(name=key, value=value, num=count)
else:
self._db.lrem(key, count, value)
|
remove_from_list
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, key: str, value, count: int=0, pipeline: bool=False):\n if pipeline:\n if redis.__version__ == '2.10.6':\n self._pipeline.lrem(name=key, value=value, num=count)\n else:\n self._pipeline.lrem(key, count, value)\n elif self._db.exists(key):\n if redis.__version__ == '2.10.6':\n self._db.lrem(name=key, value=value, num=count)\n else:\n self._db.lrem(key, count, value)\n",
"role": "user"
},
{
"content": "remove_from_list",
"role": "assistant"
}
] |
def x(self, key, pipeline=False):
if pipeline:
self._pipeline.watch(key)
else:
self._db.watch(key)
|
watch
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, key, pipeline=False):\n if pipeline:\n self._pipeline.watch(key)\n else:\n self._db.watch(key)\n",
"role": "user"
},
{
"content": "watch",
"role": "assistant"
}
] |
def x(self, channel, message, pipeline=False):
if pipeline:
self._pipeline.publish(channel, message)
else:
self._db.publish(channel, message)
|
publish
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, channel, message, pipeline=False):\n if pipeline:\n self._pipeline.publish(channel, message)\n else:\n self._db.publish(channel, message)\n",
"role": "user"
},
{
"content": "publish",
"role": "assistant"
}
] |
def x(self) ->dict:
services_health = {}
services_ids = self._get_services()
for service_id in services_ids:
service_name = DC.get_service_name(service_id)
if DC.get_replicas(service_id) != DC.get_actual_replica(service_id):
services_health[service_name] = 'Unhealthy'
else:
services_health[service_name] = 'Healthy'
return services_health
|
get_services_health
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self) ->dict:\n services_health = {}\n services_ids = self._get_services()\n for service_id in services_ids:\n service_name = DC.get_service_name(service_id)\n if DC.get_replicas(service_id) != DC.get_actual_replica(service_id):\n services_health[service_name] = 'Unhealthy'\n else:\n services_health[service_name] = 'Healthy'\n return services_health\n",
"role": "user"
},
{
"content": "get_services_health",
"role": "assistant"
}
] |
def x(self) ->str:
services_health_status = self.get_services_health()
health_status = all(status == 'Healthy' for status in
services_health_status.values())
if health_status:
overall_status = 'Healthy'
else:
overall_status = 'Unhealthy'
return overall_status
|
get_overall_services_health
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self) ->str:\n services_health_status = self.get_services_health()\n health_status = all(status == 'Healthy' for status in\n services_health_status.values())\n if health_status:\n overall_status = 'Healthy'\n else:\n overall_status = 'Unhealthy'\n return overall_status\n",
"role": "user"
},
{
"content": "get_overall_services_health",
"role": "assistant"
}
] |
def x(service_id: str) ->str:
if DC.get_replicas(service_id) != DC.get_actual_replica(service_id):
health_status = 'Unhealthy'
else:
health_status = 'Healthy'
return health_status
|
get_service_health
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(service_id: str) ->str:\n if DC.get_replicas(service_id) != DC.get_actual_replica(service_id):\n health_status = 'Unhealthy'\n else:\n health_status = 'Healthy'\n return health_status\n",
"role": "user"
},
{
"content": "get_service_health",
"role": "assistant"
}
] |
def x():
uptime = time.time() - START_TIME
response = dict(uptime=f'{uptime:.2f}s', links=dict(root='{}'.format(
get_root_url())))
return response, HTTPStatus.OK
|
get
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n uptime = time.time() - START_TIME\n response = dict(uptime=f'{uptime:.2f}s', links=dict(root='{}'.format(\n get_root_url())))\n return response, HTTPStatus.OK\n",
"role": "user"
},
{
"content": "get",
"role": "assistant"
}
] |
def x():
if len(sys.argv) < 2:
raise RuntimeError('Usage: python3 async_recv.py <json config>')
sip_logging.init_logger(show_thread=True)
spead_config = json.loads(sys.argv[1])
receiver = SpeadReceiver(spead_config)
receiver.run()
|
main
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n if len(sys.argv) < 2:\n raise RuntimeError('Usage: python3 async_recv.py <json config>')\n sip_logging.init_logger(show_thread=True)\n spead_config = json.loads(sys.argv[1])\n receiver = SpeadReceiver(spead_config)\n receiver.run()\n",
"role": "user"
},
{
"content": "main",
"role": "assistant"
}
] |
def x(self, i_block, receive_buffer):
self._log.info('Worker thread processing block %i', i_block)
time_overall0 = time.time()
time_unpack = 0.0
time_write = 0.0
for i_heap, heap in enumerate(receive_buffer.result()):
if isinstance(heap, spead2.recv.IncompleteHeap):
self._log.info('Dropped incomplete heap %i', heap.cnt + 1)
continue
items = self._item_group.update(heap)
i_chan = i_heap // self._num_buffer_times
i_time = i_heap % self._num_buffer_times
if 'correlator_output_data' in items:
vis_data = items['correlator_output_data'].value['VIS']
if self._block is None:
num_baselines = vis_data.shape[0]
num_pols = vis_data[0].shape[0]
self._block = numpy.zeros((self._num_buffer_times, self.
_num_streams, num_baselines), dtype=('c8', num_pols))
self._block[:, :, :] = 0
time_unpack0 = time.time()
self._block[i_time, i_chan, :] = vis_data
time_unpack += time.time() - time_unpack0
val = self._block[i_time, i_chan, -1][-1].real
self._log.debug('Data: %.3f', val)
if self._block is not None:
if self._config['process_data']:
pass
if self._config['write_data']:
time_write0 = time.time()
with open(self._config['filename'], 'ab') as f:
numpy.save(f, self._block, allow_pickle=False)
time_write += time.time() - time_write0
time_overall = time.time() - time_overall0
self._log.info('Total processing time: %.1f ms', 1000 * time_overall)
self._log.info('Unpack was %.1f %%', 100 * time_unpack / time_overall)
self._log.info('Write was %.1f %%', 100 * time_write / time_overall)
if time_unpack != 0.0:
self._log.info('Memory speed %.1f MB/s', self._block.nbytes * 1e-06 /
time_unpack)
if time_write != 0.0:
self._log.info('Write speed %.1f MB/s', self._block.nbytes * 1e-06 /
time_write)
|
process_buffer
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, i_block, receive_buffer):\n self._log.info('Worker thread processing block %i', i_block)\n time_overall0 = time.time()\n time_unpack = 0.0\n time_write = 0.0\n for i_heap, heap in enumerate(receive_buffer.result()):\n if isinstance(heap, spead2.recv.IncompleteHeap):\n self._log.info('Dropped incomplete heap %i', heap.cnt + 1)\n continue\n items = self._item_group.update(heap)\n i_chan = i_heap // self._num_buffer_times\n i_time = i_heap % self._num_buffer_times\n if 'correlator_output_data' in items:\n vis_data = items['correlator_output_data'].value['VIS']\n if self._block is None:\n num_baselines = vis_data.shape[0]\n num_pols = vis_data[0].shape[0]\n self._block = numpy.zeros((self._num_buffer_times, self.\n _num_streams, num_baselines), dtype=('c8', num_pols))\n self._block[:, :, :] = 0\n time_unpack0 = time.time()\n self._block[i_time, i_chan, :] = vis_data\n time_unpack += time.time() - time_unpack0\n val = self._block[i_time, i_chan, -1][-1].real\n self._log.debug('Data: %.3f', val)\n if self._block is not None:\n if self._config['process_data']:\n pass\n if self._config['write_data']:\n time_write0 = time.time()\n with open(self._config['filename'], 'ab') as f:\n numpy.save(f, self._block, allow_pickle=False)\n time_write += time.time() - time_write0\n time_overall = time.time() - time_overall0\n self._log.info('Total processing time: %.1f ms', 1000 * time_overall)\n self._log.info('Unpack was %.1f %%', 100 * time_unpack / time_overall)\n self._log.info('Write was %.1f %%', 100 * time_write / time_overall)\n if time_unpack != 0.0:\n self._log.info('Memory speed %.1f MB/s', self._block.nbytes * 1e-06 /\n time_unpack)\n if time_write != 0.0:\n self._log.info('Write speed %.1f MB/s', self._block.nbytes * 1e-06 /\n time_write)\n",
"role": "user"
},
{
"content": "process_buffer",
"role": "assistant"
}
] |
async def _run_loop(self, executor):
"""Main loop."""
loop = asyncio.get_event_loop()
self._log.info('Waiting for %d streams to start...', self._num_streams)
for stream in self._streams:
await stream.get(loop=loop)
i_block = 0
receive_buffer = [None, None]
while True:
processing_tasks = None
if i_block > 0:
i_buffer_proc = (i_block - 1) % 2
processing_tasks = loop.run_in_executor(executor, self.
process_buffer, i_block - 1, receive_buffer[i_buffer_proc])
i_buffer_recv = i_block % 2
receive_tasks = []
for stream in self._streams:
for _ in range(self._num_buffer_times):
receive_tasks.append(loop.create_task(stream.get(loop=loop)))
receive_buffer[i_buffer_recv] = asyncio.gather(*receive_tasks)
self._log.info('Receiving block %i', i_block)
try:
await receive_buffer[i_buffer_recv]
self._log.info('Received block %i', i_block)
except spead2.Stopped:
self._log.info('Stream Stopped')
break
if processing_tasks:
await processing_tasks
i_block += 1
|
run_loop
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def _run_loop(self, executor):\n \"\"\"Main loop.\"\"\"\n loop = asyncio.get_event_loop()\n self._log.info('Waiting for %d streams to start...', self._num_streams)\n for stream in self._streams:\n await stream.get(loop=loop)\n i_block = 0\n receive_buffer = [None, None]\n while True:\n processing_tasks = None\n if i_block > 0:\n i_buffer_proc = (i_block - 1) % 2\n processing_tasks = loop.run_in_executor(executor, self.\n process_buffer, i_block - 1, receive_buffer[i_buffer_proc])\n i_buffer_recv = i_block % 2\n receive_tasks = []\n for stream in self._streams:\n for _ in range(self._num_buffer_times):\n receive_tasks.append(loop.create_task(stream.get(loop=loop)))\n receive_buffer[i_buffer_recv] = asyncio.gather(*receive_tasks)\n self._log.info('Receiving block %i', i_block)\n try:\n await receive_buffer[i_buffer_recv]\n self._log.info('Received block %i', i_block)\n except spead2.Stopped:\n self._log.info('Stream Stopped')\n break\n if processing_tasks:\n await processing_tasks\n i_block += 1\n",
"role": "user"
},
{
"content": "run_loop",
"role": "assistant"
}
] |
def x(self):
executor = concurrent.futures.ThreadPoolExecutor(max_workers=1)
loop = asyncio.get_event_loop()
loop.run_until_complete(self._run_loop(executor))
self._log.info('Shutting down...')
executor.shutdown()
|
run
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n executor = concurrent.futures.ThreadPoolExecutor(max_workers=1)\n loop = asyncio.get_event_loop()\n loop.run_until_complete(self._run_loop(executor))\n self._log.info('Shutting down...')\n executor.shutdown()\n",
"role": "user"
},
{
"content": "run",
"role": "assistant"
}
] |
def x(self):
self.seek(0)
while True:
brace_count = 0
quoted = False
json_string = ''
while True:
b = self.read(1)
if b == b'':
return
c = b.decode()
if c == '\\':
json_string += c
json_string += self.read(1).decode()
continue
if c == '"':
if quoted:
quoted = False
else:
quoted = True
if c == '{':
brace_count += 1
elif c == '}':
brace_count -= 1
json_string += c
if brace_count == 0:
break
meta = json.loads(json_string)
n_channels = meta['data_cube']['n_channels']
n_sub_integrations = meta['data_cube']['n_sub_integrations']
f = open('{0}_{1}_{2}.json'.format(meta['metadata'][
'observation_id'], meta['metadata']['beam_id'], meta['metadata'
]['name']), 'w')
f.write(json.dumps(meta))
f.close()
data = self.read(n_channels * n_sub_integrations)
f = open('{0}_{1}_{2}.data'.format(meta['metadata'][
'observation_id'], meta['metadata']['beam_id'], meta['metadata'
]['name']), 'wb')
f.write(data)
f.close()
|
close
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n self.seek(0)\n while True:\n brace_count = 0\n quoted = False\n json_string = ''\n while True:\n b = self.read(1)\n if b == b'':\n return\n c = b.decode()\n if c == '\\\\':\n json_string += c\n json_string += self.read(1).decode()\n continue\n if c == '\"':\n if quoted:\n quoted = False\n else:\n quoted = True\n if c == '{':\n brace_count += 1\n elif c == '}':\n brace_count -= 1\n json_string += c\n if brace_count == 0:\n break\n meta = json.loads(json_string)\n n_channels = meta['data_cube']['n_channels']\n n_sub_integrations = meta['data_cube']['n_sub_integrations']\n f = open('{0}_{1}_{2}.json'.format(meta['metadata'][\n 'observation_id'], meta['metadata']['beam_id'], meta['metadata'\n ]['name']), 'w')\n f.write(json.dumps(meta))\n f.close()\n data = self.read(n_channels * n_sub_integrations)\n f = open('{0}_{1}_{2}.data'.format(meta['metadata'][\n 'observation_id'], meta['metadata']['beam_id'], meta['metadata'\n ]['name']), 'wb')\n f.write(data)\n f.close()\n",
"role": "user"
},
{
"content": "close",
"role": "assistant"
}
] |
def x(self):
self._log.info('Starting Pulsar Search Interface')
authorizer = DummyAuthorizer()
authorizer.add_user(self._config['login']['user'], self._config['login'
]['psswd'], '.', perm=self._config['login']['perm'])
authorizer.add_anonymous(os.getcwd())
handler = FTPHandler
handler.authorizer = authorizer
handler.abstracted_fs = PulsarFileSystem
handler.banner = 'SKA SDP pulsar search interface.'
address = self._config['address']['listen'], self._config['address']['port'
]
server = FTPServer(address, handler)
server.max_cons = 256
server.max_cons_per_ip = 5
server.serve_forever()
|
run
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n self._log.info('Starting Pulsar Search Interface')\n authorizer = DummyAuthorizer()\n authorizer.add_user(self._config['login']['user'], self._config['login'\n ]['psswd'], '.', perm=self._config['login']['perm'])\n authorizer.add_anonymous(os.getcwd())\n handler = FTPHandler\n handler.authorizer = authorizer\n handler.abstracted_fs = PulsarFileSystem\n handler.banner = 'SKA SDP pulsar search interface.'\n address = self._config['address']['listen'], self._config['address']['port'\n ]\n server = FTPServer(address, handler)\n server.max_cons = 256\n server.max_cons_per_ip = 5\n server.serve_forever()\n",
"role": "user"
},
{
"content": "run",
"role": "assistant"
}
] |
def x(self) ->str:
self._config = self._load_config()
return self._config.get('status')
|
status
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self) ->str:\n self._config = self._load_config()\n return self._config.get('status')\n",
"role": "user"
},
{
"content": "status",
"role": "assistant"
}
] |
def x(self, value):
pb_key = SchedulingObject.get_key(PB_KEY, self._pb_id)
stages = DB.get_hash_value(pb_key, 'workflow_stages')
stages = ast.literal_eval(stages)
stages[self._index]['status'] = value
DB.set_hash_value(pb_key, 'workflow_stages', stages)
|
status
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, value):\n pb_key = SchedulingObject.get_key(PB_KEY, self._pb_id)\n stages = DB.get_hash_value(pb_key, 'workflow_stages')\n stages = ast.literal_eval(stages)\n stages[self._index]['status'] = value\n DB.set_hash_value(pb_key, 'workflow_stages', stages)\n",
"role": "user"
},
{
"content": "status",
"role": "assistant"
}
] |
def x(self):
pb_key = SchedulingObject.get_key(PB_KEY, self._pb_id)
stages = DB.get_hash_value(pb_key, 'workflow_stages')
stages = ast.literal_eval(stages)
return stages[self._index]
|
load_config
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n pb_key = SchedulingObject.get_key(PB_KEY, self._pb_id)\n stages = DB.get_hash_value(pb_key, 'workflow_stages')\n stages = ast.literal_eval(stages)\n return stages[self._index]\n",
"role": "user"
},
{
"content": "load_config",
"role": "assistant"
}
] |
def x():
parser = argparse.ArgumentParser(description='Register PB devices.')
parser.add_argument('num_pb', type=int, help=
'Number of PBs devices to register.')
return parser.parse_args()
|
parse_command_args
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n parser = argparse.ArgumentParser(description='Register PB devices.')\n parser.add_argument('num_pb', type=int, help=\n 'Number of PBs devices to register.')\n return parser.parse_args()\n",
"role": "user"
},
{
"content": "parse_command_args",
"role": "assistant"
}
] |
def x(num_pbs: int=100):
tango_db = Database()
LOG.info('Registering PB devices:')
dev_info = DbDevInfo()
dev_info._class = 'ProcessingBlockDevice'
dev_info.server = 'processing_block_ds/1'
for index in range(num_pbs):
dev_info.name = 'sip_sdp/pb/{:05d}'.format(index)
LOG.info('\t%s', dev_info.name)
tango_db.add_device(dev_info)
|
register_pb_devices
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(num_pbs: int=100):\n tango_db = Database()\n LOG.info('Registering PB devices:')\n dev_info = DbDevInfo()\n dev_info._class = 'ProcessingBlockDevice'\n dev_info.server = 'processing_block_ds/1'\n for index in range(num_pbs):\n dev_info.name = 'sip_sdp/pb/{:05d}'.format(index)\n LOG.info('\\t%s', dev_info.name)\n tango_db.add_device(dev_info)\n",
"role": "user"
},
{
"content": "register_pb_devices",
"role": "assistant"
}
] |
def x(cls, config_dict: dict, schema_path: str=None):
if schema_path is None:
schema_path = join(dirname(__file__), 'schema', 'configure_sbi.json')
with open(schema_path, 'r') as file:
schema = json.loads(file.read())
validate(config_dict, schema)
config_dict['status'] = 'created'
if 'subarray_id' not in config_dict:
config_dict['subarray_id'] = 'None'
timestamp = datetime.datetime.utcnow().isoformat()
config_dict['created'] = timestamp
config_dict['updated'] = timestamp
pb_list = copy.deepcopy(config_dict['processing_blocks'])
config_dict.pop('processing_blocks', None)
config_dict['processing_block_ids'] = []
for pb in pb_list:
config_dict['processing_block_ids'].append(pb['id'])
key = SchedulingObject.get_key(SBI_KEY, config_dict['id'])
DB.save_dict(key, config_dict, hierarchical=False)
key = '{}:active'.format(SBI_KEY)
DB.append_to_list(key, config_dict['id'])
sbi = SchedulingObject(SBI_KEY, config_dict['id'])
sbi.set_status('created')
for pb in pb_list:
pb['sbi_id'] = config_dict['id']
cls._add_pb(pb)
return cls(config_dict['id'])
|
from_config
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, config_dict: dict, schema_path: str=None):\n if schema_path is None:\n schema_path = join(dirname(__file__), 'schema', 'configure_sbi.json')\n with open(schema_path, 'r') as file:\n schema = json.loads(file.read())\n validate(config_dict, schema)\n config_dict['status'] = 'created'\n if 'subarray_id' not in config_dict:\n config_dict['subarray_id'] = 'None'\n timestamp = datetime.datetime.utcnow().isoformat()\n config_dict['created'] = timestamp\n config_dict['updated'] = timestamp\n pb_list = copy.deepcopy(config_dict['processing_blocks'])\n config_dict.pop('processing_blocks', None)\n config_dict['processing_block_ids'] = []\n for pb in pb_list:\n config_dict['processing_block_ids'].append(pb['id'])\n key = SchedulingObject.get_key(SBI_KEY, config_dict['id'])\n DB.save_dict(key, config_dict, hierarchical=False)\n key = '{}:active'.format(SBI_KEY)\n DB.append_to_list(key, config_dict['id'])\n sbi = SchedulingObject(SBI_KEY, config_dict['id'])\n sbi.set_status('created')\n for pb in pb_list:\n pb['sbi_id'] = config_dict['id']\n cls._add_pb(pb)\n return cls(config_dict['id'])\n",
"role": "user"
},
{
"content": "from_config",
"role": "assistant"
}
] |
def x(self):
self.set_status('aborted')
DB.remove_from_list('{}:active'.format(self._type), self._id)
DB.append_to_list('{}:aborted'.format(self._type), self._id)
sbi_pb_ids = ast.literal_eval(DB.get_hash_value(self._key,
'processing_block_ids'))
for pb_id in sbi_pb_ids:
pb = ProcessingBlock(pb_id)
pb.abort()
|
abort
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n self.set_status('aborted')\n DB.remove_from_list('{}:active'.format(self._type), self._id)\n DB.append_to_list('{}:aborted'.format(self._type), self._id)\n sbi_pb_ids = ast.literal_eval(DB.get_hash_value(self._key,\n 'processing_block_ids'))\n for pb_id in sbi_pb_ids:\n pb = ProcessingBlock(pb_id)\n pb.abort()\n",
"role": "user"
},
{
"content": "abort",
"role": "assistant"
}
] |
def x(self) ->List[str]:
values = DB.get_hash_value(self._key, 'processing_block_ids')
return ast.literal_eval(values)
|
get_pb_ids
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self) ->List[str]:\n values = DB.get_hash_value(self._key, 'processing_block_ids')\n return ast.literal_eval(values)\n",
"role": "user"
},
{
"content": "get_pb_ids",
"role": "assistant"
}
] |
def x(date=None, project: str='sip', instance_id: int=None) ->str:
if date is None:
date = datetime.datetime.utcnow()
if isinstance(date, datetime.datetime):
date = date.strftime('%Y%m%d')
if instance_id is None:
instance_id = randint(0, 9999)
return 'SBI-{}-{}-{:04d}'.format(date, project, instance_id)
|
get_id
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(date=None, project: str='sip', instance_id: int=None) ->str:\n if date is None:\n date = datetime.datetime.utcnow()\n if isinstance(date, datetime.datetime):\n date = date.strftime('%Y%m%d')\n if instance_id is None:\n instance_id = randint(0, 9999)\n return 'SBI-{}-{}-{:04d}'.format(date, project, instance_id)\n",
"role": "user"
},
{
"content": "get_id",
"role": "assistant"
}
] |
def x(pb_config: dict):
pb_config['status'] = 'created'
timestamp = datetime.datetime.utcnow().isoformat()
pb_config['created'] = timestamp
pb_config['updated'] = timestamp
if 'priority' not in pb_config:
pb_config['priority'] = 0
SchedulingBlockInstance._update_workflow_definition(pb_config)
keys = ['resources_required', 'resources_assigned', 'dependencies']
for key in keys:
if key not in pb_config:
pb_config[key] = []
for stage in pb_config['workflow_stages']:
if key not in stage:
stage[key] = []
key = SchedulingObject.get_key(PB_KEY, pb_config['id'])
DB.save_dict(key, pb_config, hierarchical=False)
key = '{}:active'.format(PB_KEY)
DB.append_to_list(key, pb_config['id'])
key = '{}:active:{}'.format(PB_KEY, pb_config['type'])
DB.append_to_list(key, pb_config['id'])
pb = SchedulingObject(PB_KEY, pb_config['id'])
pb.set_status('created')
|
add_pb
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(pb_config: dict):\n pb_config['status'] = 'created'\n timestamp = datetime.datetime.utcnow().isoformat()\n pb_config['created'] = timestamp\n pb_config['updated'] = timestamp\n if 'priority' not in pb_config:\n pb_config['priority'] = 0\n SchedulingBlockInstance._update_workflow_definition(pb_config)\n keys = ['resources_required', 'resources_assigned', 'dependencies']\n for key in keys:\n if key not in pb_config:\n pb_config[key] = []\n for stage in pb_config['workflow_stages']:\n if key not in stage:\n stage[key] = []\n key = SchedulingObject.get_key(PB_KEY, pb_config['id'])\n DB.save_dict(key, pb_config, hierarchical=False)\n key = '{}:active'.format(PB_KEY)\n DB.append_to_list(key, pb_config['id'])\n key = '{}:active:{}'.format(PB_KEY, pb_config['type'])\n DB.append_to_list(key, pb_config['id'])\n pb = SchedulingObject(PB_KEY, pb_config['id'])\n pb.set_status('created')\n",
"role": "user"
},
{
"content": "add_pb",
"role": "assistant"
}
] |
def x(pb_config: dict):
known_workflows = get_workflows()
workflow_id = pb_config['workflow']['id']
workflow_version = pb_config['workflow']['version']
if (workflow_id not in known_workflows or workflow_version not in
known_workflows[workflow_id]):
raise RuntimeError('Unknown workflow definition: {}:{}'.format(
workflow_id, workflow_version))
workflow = get_workflow(workflow_id, workflow_version)
for stage in workflow['stages']:
stage['status'] = 'none'
pb_config['workflow_parameters'] = pb_config['workflow']['parameters']
pb_config['workflow_id'] = pb_config['workflow']['id']
pb_config['workflow_version'] = pb_config['workflow']['version']
pb_config['workflow_stages'] = workflow['stages']
pb_config.pop('workflow', None)
|
update_workflow_definition
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(pb_config: dict):\n known_workflows = get_workflows()\n workflow_id = pb_config['workflow']['id']\n workflow_version = pb_config['workflow']['version']\n if (workflow_id not in known_workflows or workflow_version not in\n known_workflows[workflow_id]):\n raise RuntimeError('Unknown workflow definition: {}:{}'.format(\n workflow_id, workflow_version))\n workflow = get_workflow(workflow_id, workflow_version)\n for stage in workflow['stages']:\n stage['status'] = 'none'\n pb_config['workflow_parameters'] = pb_config['workflow']['parameters']\n pb_config['workflow_id'] = pb_config['workflow']['id']\n pb_config['workflow_version'] = pb_config['workflow']['version']\n pb_config['workflow_stages'] = workflow['stages']\n pb_config.pop('workflow', None)\n",
"role": "user"
},
{
"content": "update_workflow_definition",
"role": "assistant"
}
] |
def x():
response = {'links': {'message':
'Welcome to the SIP Processing Controller Interface', 'items': [{
'href': '{}health'.format(request.url)}, {'href': '{}subarrays'.
format(request.url)}, {'href': '{}scheduling_blocks'.format(request
.url)}, {'href': '{}processing_blocks'.format(request.url)}]}}
return response, HTTPStatus.OK
|
root
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n response = {'links': {'message':\n 'Welcome to the SIP Processing Controller Interface', 'items': [{\n 'href': '{}health'.format(request.url)}, {'href': '{}subarrays'.\n format(request.url)}, {'href': '{}scheduling_blocks'.format(request\n .url)}, {'href': '{}processing_blocks'.format(request.url)}]}}\n return response, HTTPStatus.OK\n",
"role": "user"
},
{
"content": "root",
"role": "assistant"
}
] |
def x(num_blocks, project='test'):
_date = strftime('%Y%m%d', gmtime())
_project = project
for i in range(num_blocks):
yield '{}-{}-sbi{:03d}'.format(_date, _project, i)
|
generate_scheduling_block_id
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(num_blocks, project='test'):\n _date = strftime('%Y%m%d', gmtime())\n _project = project\n for i in range(num_blocks):\n yield '{}-{}-sbi{:03d}'.format(_date, _project, i)\n",
"role": "user"
},
{
"content": "generate_scheduling_block_id",
"role": "assistant"
}
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.