source_code
stringlengths
45
7.24k
function_name
stringlengths
1
111
messages
list
def x(self, delivery_tag=0, requeue=True): if not compatibility.is_integer(delivery_tag): raise AMQPInvalidArgument('delivery_tag should be an integer') elif not isinstance(requeue, bool): raise AMQPInvalidArgument('requeue should be a boolean') reject_frame = specification.Basic.Reject(delivery_tag=delivery_tag, requeue=requeue) self._channel.write_frame(reject_frame)
reject
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, delivery_tag=0, requeue=True):\n if not compatibility.is_integer(delivery_tag):\n raise AMQPInvalidArgument('delivery_tag should be an integer')\n elif not isinstance(requeue, bool):\n raise AMQPInvalidArgument('requeue should be a boolean')\n reject_frame = specification.Basic.Reject(delivery_tag=delivery_tag,\n requeue=requeue)\n self._channel.write_frame(reject_frame)\n", "role": "user" }, { "content": "reject", "role": "assistant" } ]
def x(self, consume_rpc_result): consumer_tag = consume_rpc_result['consumer_tag'] self._channel.add_consumer_tag(consumer_tag) return consumer_tag
consume_add_and_get_tag
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, consume_rpc_result):\n consumer_tag = consume_rpc_result['consumer_tag']\n self._channel.add_consumer_tag(consumer_tag)\n return consumer_tag\n", "role": "user" }, { "content": "consume_add_and_get_tag", "role": "assistant" } ]
def x(self, arguments, consumer_tag, exclusive, no_ack, no_local, queue): consume_frame = specification.Basic.Consume(queue=queue, consumer_tag= consumer_tag, exclusive=exclusive, no_local=no_local, no_ack=no_ack, arguments=arguments) return self._channel.rpc_request(consume_frame)
consume_rpc_request
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, arguments, consumer_tag, exclusive, no_ack, no_local, queue):\n consume_frame = specification.Basic.Consume(queue=queue, consumer_tag=\n consumer_tag, exclusive=exclusive, no_local=no_local, no_ack=no_ack,\n arguments=arguments)\n return self._channel.rpc_request(consume_frame)\n", "role": "user" }, { "content": "consume_rpc_request", "role": "assistant" } ]
def x(body, exchange, immediate, mandatory, properties, routing_key): if not compatibility.is_string(body): raise AMQPInvalidArgument('body should be a string') elif not compatibility.is_string(routing_key): raise AMQPInvalidArgument('routing_key should be a string') elif not compatibility.is_string(exchange): raise AMQPInvalidArgument('exchange should be a string') elif properties is not None and not isinstance(properties, dict): raise AMQPInvalidArgument('properties should be a dict or None') elif not isinstance(mandatory, bool): raise AMQPInvalidArgument('mandatory should be a boolean') elif not isinstance(immediate, bool): raise AMQPInvalidArgument('immediate should be a boolean')
validate_publish_parameters
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(body, exchange, immediate, mandatory, properties, routing_key):\n if not compatibility.is_string(body):\n raise AMQPInvalidArgument('body should be a string')\n elif not compatibility.is_string(routing_key):\n raise AMQPInvalidArgument('routing_key should be a string')\n elif not compatibility.is_string(exchange):\n raise AMQPInvalidArgument('exchange should be a string')\n elif properties is not None and not isinstance(properties, dict):\n raise AMQPInvalidArgument('properties should be a dict or None')\n elif not isinstance(mandatory, bool):\n raise AMQPInvalidArgument('mandatory should be a boolean')\n elif not isinstance(immediate, bool):\n raise AMQPInvalidArgument('immediate should be a boolean')\n", "role": "user" }, { "content": "validate_publish_parameters", "role": "assistant" } ]
def x(body, properties): if 'content_encoding' not in properties: properties['content_encoding'] = 'utf-8' encoding = properties['content_encoding'] if compatibility.is_unicode(body): body = body.encode(encoding) elif compatibility.PYTHON3 and isinstance(body, str): body = bytes(body, encoding=encoding) return body
handle_utf8_payload
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(body, properties):\n if 'content_encoding' not in properties:\n properties['content_encoding'] = 'utf-8'\n encoding = properties['content_encoding']\n if compatibility.is_unicode(body):\n body = body.encode(encoding)\n elif compatibility.PYTHON3 and isinstance(body, str):\n body = bytes(body, encoding=encoding)\n return body\n", "role": "user" }, { "content": "handle_utf8_payload", "role": "assistant" } ]
def x(self, get_frame, auto_decode): message_uuid = self._channel.rpc.register_request(get_frame. valid_responses + ['ContentHeader', 'ContentBody']) try: self._channel.write_frame(get_frame) get_ok_frame = self._channel.rpc.get_request(message_uuid, raw=True, multiple=True) if isinstance(get_ok_frame, specification.Basic.GetEmpty): return None content_header = self._channel.rpc.get_request(message_uuid, raw= True, multiple=True) body = self._get_content_body(message_uuid, content_header.body_size) finally: self._channel.rpc.remove(message_uuid) return Message(channel=self._channel, body=body, method=dict( get_ok_frame), properties=dict(content_header.properties), auto_decode=auto_decode)
get_message
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, get_frame, auto_decode):\n message_uuid = self._channel.rpc.register_request(get_frame.\n valid_responses + ['ContentHeader', 'ContentBody'])\n try:\n self._channel.write_frame(get_frame)\n get_ok_frame = self._channel.rpc.get_request(message_uuid, raw=True,\n multiple=True)\n if isinstance(get_ok_frame, specification.Basic.GetEmpty):\n return None\n content_header = self._channel.rpc.get_request(message_uuid, raw=\n True, multiple=True)\n body = self._get_content_body(message_uuid, content_header.body_size)\n finally:\n self._channel.rpc.remove(message_uuid)\n return Message(channel=self._channel, body=body, method=dict(\n get_ok_frame), properties=dict(content_header.properties),\n auto_decode=auto_decode)\n", "role": "user" }, { "content": "get_message", "role": "assistant" } ]
def x(self, frames_out): confirm_uuid = self._channel.rpc.register_request(['Basic.Ack', 'Basic.Nack']) self._channel.write_frames(frames_out) result = self._channel.rpc.get_request(confirm_uuid, raw=True) self._channel.check_for_errors() if isinstance(result, specification.Basic.Ack): return True return False
publish_confirm
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, frames_out):\n confirm_uuid = self._channel.rpc.register_request(['Basic.Ack',\n 'Basic.Nack'])\n self._channel.write_frames(frames_out)\n result = self._channel.rpc.get_request(confirm_uuid, raw=True)\n self._channel.check_for_errors()\n if isinstance(result, specification.Basic.Ack):\n return True\n return False\n", "role": "user" }, { "content": "publish_confirm", "role": "assistant" } ]
def x(self, body): frames = int(math.ceil(len(body) / float(self._max_frame_size))) for offset in compatibility.RANGE(0, frames): start_frame = self._max_frame_size * offset end_frame = start_frame + self._max_frame_size body_len = len(body) if end_frame > body_len: end_frame = body_len yield pamqp_body.ContentBody(body[start_frame:end_frame])
create_content_body
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, body):\n frames = int(math.ceil(len(body) / float(self._max_frame_size)))\n for offset in compatibility.RANGE(0, frames):\n start_frame = self._max_frame_size * offset\n end_frame = start_frame + self._max_frame_size\n body_len = len(body)\n if end_frame > body_len:\n end_frame = body_len\n yield pamqp_body.ContentBody(body[start_frame:end_frame])\n", "role": "user" }, { "content": "create_content_body", "role": "assistant" } ]
def x(self, message_uuid, body_size): body = bytes() while len(body) < body_size: body_piece = self._channel.rpc.get_request(message_uuid, raw=True, multiple=True) if not body_piece.value: break body += body_piece.value return body
get_content_body
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, message_uuid, body_size):\n body = bytes()\n while len(body) < body_size:\n body_piece = self._channel.rpc.get_request(message_uuid, raw=True,\n multiple=True)\n if not body_piece.value:\n break\n body += body_piece.value\n return body\n", "role": "user" }, { "content": "get_content_body", "role": "assistant" } ]
def x(self, queue='', passive=False, durable=False, exclusive=False, auto_delete=False, arguments=None): if not compatibility.is_string(queue): raise AMQPInvalidArgument('queue should be a string') elif not isinstance(passive, bool): raise AMQPInvalidArgument('passive should be a boolean') elif not isinstance(durable, bool): raise AMQPInvalidArgument('durable should be a boolean') elif not isinstance(exclusive, bool): raise AMQPInvalidArgument('exclusive should be a boolean') elif not isinstance(auto_delete, bool): raise AMQPInvalidArgument('auto_delete should be a boolean') elif arguments is not None and not isinstance(arguments, dict): raise AMQPInvalidArgument('arguments should be a dict or None') declare_frame = pamqp_queue.Declare(queue=queue, passive=passive, durable=durable, exclusive=exclusive, auto_delete=auto_delete, arguments=arguments) return self._channel.rpc_request(declare_frame)
declare
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, queue='', passive=False, durable=False, exclusive=False,\n auto_delete=False, arguments=None):\n if not compatibility.is_string(queue):\n raise AMQPInvalidArgument('queue should be a string')\n elif not isinstance(passive, bool):\n raise AMQPInvalidArgument('passive should be a boolean')\n elif not isinstance(durable, bool):\n raise AMQPInvalidArgument('durable should be a boolean')\n elif not isinstance(exclusive, bool):\n raise AMQPInvalidArgument('exclusive should be a boolean')\n elif not isinstance(auto_delete, bool):\n raise AMQPInvalidArgument('auto_delete should be a boolean')\n elif arguments is not None and not isinstance(arguments, dict):\n raise AMQPInvalidArgument('arguments should be a dict or None')\n declare_frame = pamqp_queue.Declare(queue=queue, passive=passive,\n durable=durable, exclusive=exclusive, auto_delete=auto_delete,\n arguments=arguments)\n return self._channel.rpc_request(declare_frame)\n", "role": "user" }, { "content": "declare", "role": "assistant" } ]
def x(self, queue='', if_unused=False, if_empty=False): if not compatibility.is_string(queue): raise AMQPInvalidArgument('queue should be a string') elif not isinstance(if_unused, bool): raise AMQPInvalidArgument('if_unused should be a boolean') elif not isinstance(if_empty, bool): raise AMQPInvalidArgument('if_empty should be a boolean') delete_frame = pamqp_queue.Delete(queue=queue, if_unused=if_unused, if_empty=if_empty) return self._channel.rpc_request(delete_frame)
delete
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, queue='', if_unused=False, if_empty=False):\n if not compatibility.is_string(queue):\n raise AMQPInvalidArgument('queue should be a string')\n elif not isinstance(if_unused, bool):\n raise AMQPInvalidArgument('if_unused should be a boolean')\n elif not isinstance(if_empty, bool):\n raise AMQPInvalidArgument('if_empty should be a boolean')\n delete_frame = pamqp_queue.Delete(queue=queue, if_unused=if_unused,\n if_empty=if_empty)\n return self._channel.rpc_request(delete_frame)\n", "role": "user" }, { "content": "delete", "role": "assistant" } ]
def x(self, queue): if not compatibility.is_string(queue): raise AMQPInvalidArgument('queue should be a string') purge_frame = pamqp_queue.Purge(queue=queue) return self._channel.rpc_request(purge_frame)
purge
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, queue):\n if not compatibility.is_string(queue):\n raise AMQPInvalidArgument('queue should be a string')\n purge_frame = pamqp_queue.Purge(queue=queue)\n return self._channel.rpc_request(purge_frame)\n", "role": "user" }, { "content": "purge", "role": "assistant" } ]
def x(self, queue='', exchange='', routing_key='', arguments=None): if not compatibility.is_string(queue): raise AMQPInvalidArgument('queue should be a string') elif not compatibility.is_string(exchange): raise AMQPInvalidArgument('exchange should be a string') elif not compatibility.is_string(routing_key): raise AMQPInvalidArgument('routing_key should be a string') elif arguments is not None and not isinstance(arguments, dict): raise AMQPInvalidArgument('arguments should be a dict or None') bind_frame = pamqp_queue.Bind(queue=queue, exchange=exchange, routing_key=routing_key, arguments=arguments) return self._channel.rpc_request(bind_frame)
bind
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, queue='', exchange='', routing_key='', arguments=None):\n if not compatibility.is_string(queue):\n raise AMQPInvalidArgument('queue should be a string')\n elif not compatibility.is_string(exchange):\n raise AMQPInvalidArgument('exchange should be a string')\n elif not compatibility.is_string(routing_key):\n raise AMQPInvalidArgument('routing_key should be a string')\n elif arguments is not None and not isinstance(arguments, dict):\n raise AMQPInvalidArgument('arguments should be a dict or None')\n bind_frame = pamqp_queue.Bind(queue=queue, exchange=exchange,\n routing_key=routing_key, arguments=arguments)\n return self._channel.rpc_request(bind_frame)\n", "role": "user" }, { "content": "bind", "role": "assistant" } ]
def x(self, queue='', exchange='', routing_key='', arguments=None): if not compatibility.is_string(queue): raise AMQPInvalidArgument('queue should be a string') elif not compatibility.is_string(exchange): raise AMQPInvalidArgument('exchange should be a string') elif not compatibility.is_string(routing_key): raise AMQPInvalidArgument('routing_key should be a string') elif arguments is not None and not isinstance(arguments, dict): raise AMQPInvalidArgument('arguments should be a dict or None') unbind_frame = pamqp_queue.Unbind(queue=queue, exchange=exchange, routing_key=routing_key, arguments=arguments) return self._channel.rpc_request(unbind_frame)
unbind
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, queue='', exchange='', routing_key='', arguments=None):\n if not compatibility.is_string(queue):\n raise AMQPInvalidArgument('queue should be a string')\n elif not compatibility.is_string(exchange):\n raise AMQPInvalidArgument('exchange should be a string')\n elif not compatibility.is_string(routing_key):\n raise AMQPInvalidArgument('routing_key should be a string')\n elif arguments is not None and not isinstance(arguments, dict):\n raise AMQPInvalidArgument('arguments should be a dict or None')\n unbind_frame = pamqp_queue.Unbind(queue=queue, exchange=exchange,\n routing_key=routing_key, arguments=arguments)\n return self._channel.rpc_request(unbind_frame)\n", "role": "user" }, { "content": "unbind", "role": "assistant" } ]
def x(self, frame_in): LOGGER.debug('Frame Received: %s', frame_in.name) if frame_in.name == 'Heartbeat': return elif frame_in.name == 'Connection.Close': self._close_connection(frame_in) elif frame_in.name == 'Connection.CloseOk': self._close_connection_ok() elif frame_in.name == 'Connection.Blocked': self._blocked_connection(frame_in) elif frame_in.name == 'Connection.Unblocked': self._unblocked_connection() elif frame_in.name == 'Connection.OpenOk': self._set_connection_state(Stateful.OPEN) elif frame_in.name == 'Connection.Start': self.server_properties = frame_in.server_properties self._send_start_ok(frame_in) elif frame_in.name == 'Connection.Tune': self._send_tune_ok(frame_in) self._send_open_connection() else: LOGGER.error('[Channel0] Unhandled Frame: %s', frame_in.name)
on_frame
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, frame_in):\n LOGGER.debug('Frame Received: %s', frame_in.name)\n if frame_in.name == 'Heartbeat':\n return\n elif frame_in.name == 'Connection.Close':\n self._close_connection(frame_in)\n elif frame_in.name == 'Connection.CloseOk':\n self._close_connection_ok()\n elif frame_in.name == 'Connection.Blocked':\n self._blocked_connection(frame_in)\n elif frame_in.name == 'Connection.Unblocked':\n self._unblocked_connection()\n elif frame_in.name == 'Connection.OpenOk':\n self._set_connection_state(Stateful.OPEN)\n elif frame_in.name == 'Connection.Start':\n self.server_properties = frame_in.server_properties\n self._send_start_ok(frame_in)\n elif frame_in.name == 'Connection.Tune':\n self._send_tune_ok(frame_in)\n self._send_open_connection()\n else:\n LOGGER.error('[Channel0] Unhandled Frame: %s', frame_in.name)\n", "role": "user" }, { "content": "on_frame", "role": "assistant" } ]
def x(self, frame_in): self._set_connection_state(Stateful.CLOSED) if frame_in.reply_code != 200: reply_text = try_utf8_decode(frame_in.reply_text) message = 'Connection was closed by remote server: %s' % reply_text exception = AMQPConnectionError(message, reply_code=frame_in.reply_code ) self._connection.exceptions.append(exception)
close_connection
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, frame_in):\n self._set_connection_state(Stateful.CLOSED)\n if frame_in.reply_code != 200:\n reply_text = try_utf8_decode(frame_in.reply_text)\n message = 'Connection was closed by remote server: %s' % reply_text\n exception = AMQPConnectionError(message, reply_code=frame_in.reply_code\n )\n self._connection.exceptions.append(exception)\n", "role": "user" }, { "content": "close_connection", "role": "assistant" } ]
def x(self, frame_in): self.is_blocked = True LOGGER.warning('Connection is blocked by remote server: %s', try_utf8_decode(frame_in.reason))
blocked_connection
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, frame_in):\n self.is_blocked = True\n LOGGER.warning('Connection is blocked by remote server: %s',\n try_utf8_decode(frame_in.reason))\n", "role": "user" }, { "content": "blocked_connection", "role": "assistant" } ]
def x(self, frame_in): mechanisms = try_utf8_decode(frame_in.mechanisms) if 'EXTERNAL' in mechanisms: mechanism = 'EXTERNAL' credentials = '\x00\x00' elif 'PLAIN' in mechanisms: mechanism = 'PLAIN' credentials = self._plain_credentials() else: exception = AMQPConnectionError( 'Unsupported Security Mechanism(s): %s' % frame_in.mechanisms) self._connection.exceptions.append(exception) return start_ok_frame = specification.Connection.StartOk(mechanism=mechanism, client_properties=self._client_properties(), response=credentials, locale=LOCALE) self._write_frame(start_ok_frame)
send_start_ok
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, frame_in):\n mechanisms = try_utf8_decode(frame_in.mechanisms)\n if 'EXTERNAL' in mechanisms:\n mechanism = 'EXTERNAL'\n credentials = '\\x00\\x00'\n elif 'PLAIN' in mechanisms:\n mechanism = 'PLAIN'\n credentials = self._plain_credentials()\n else:\n exception = AMQPConnectionError(\n 'Unsupported Security Mechanism(s): %s' % frame_in.mechanisms)\n self._connection.exceptions.append(exception)\n return\n start_ok_frame = specification.Connection.StartOk(mechanism=mechanism,\n client_properties=self._client_properties(), response=credentials,\n locale=LOCALE)\n self._write_frame(start_ok_frame)\n", "role": "user" }, { "content": "send_start_ok", "role": "assistant" } ]
def x(self, frame_in): self.max_allowed_channels = self._negotiate(frame_in.channel_max, MAX_CHANNELS) self.max_frame_size = self._negotiate(frame_in.frame_max, MAX_FRAME_SIZE) LOGGER.debug('Negotiated max frame size %d, max channels %d', self. max_frame_size, self.max_allowed_channels) tune_ok_frame = specification.Connection.TuneOk(channel_max=self. max_allowed_channels, frame_max=self.max_frame_size, heartbeat=self ._heartbeat) self._write_frame(tune_ok_frame)
send_tune_ok
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, frame_in):\n self.max_allowed_channels = self._negotiate(frame_in.channel_max,\n MAX_CHANNELS)\n self.max_frame_size = self._negotiate(frame_in.frame_max, MAX_FRAME_SIZE)\n LOGGER.debug('Negotiated max frame size %d, max channels %d', self.\n max_frame_size, self.max_allowed_channels)\n tune_ok_frame = specification.Connection.TuneOk(channel_max=self.\n max_allowed_channels, frame_max=self.max_frame_size, heartbeat=self\n ._heartbeat)\n self._write_frame(tune_ok_frame)\n", "role": "user" }, { "content": "send_tune_ok", "role": "assistant" } ]
def x(self): open_frame = specification.Connection.Open(virtual_host=self. _parameters['virtual_host']) self._write_frame(open_frame)
send_open_connection
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n open_frame = specification.Connection.Open(virtual_host=self.\n _parameters['virtual_host'])\n self._write_frame(open_frame)\n", "role": "user" }, { "content": "send_open_connection", "role": "assistant" } ]
def x(self, frame_out): self._connection.write_frame(0, frame_out) LOGGER.debug('Frame Sent: %s', frame_out.name)
write_frame
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, frame_out):\n self._connection.write_frame(0, frame_out)\n LOGGER.debug('Frame Sent: %s', frame_out.name)\n", "role": "user" }, { "content": "write_frame", "role": "assistant" } ]
def x(): return {'product': 'AMQPStorm', 'platform': 'Python %s (%s)' % ( platform.python_version(), platform.python_implementation()), 'capabilities': {'basic.nack': True, 'connection.blocked': True, 'publisher_confirms': True, 'consumer_cancel_notify': True, 'authentication_failure_close': True}, 'information': 'See https://github.com/eandersson/amqpstorm', 'version': __version__}
client_properties
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n return {'product': 'AMQPStorm', 'platform': 'Python %s (%s)' % (\n platform.python_version(), platform.python_implementation()),\n 'capabilities': {'basic.nack': True, 'connection.blocked': True,\n 'publisher_confirms': True, 'consumer_cancel_notify': True,\n 'authentication_failure_close': True}, 'information':\n 'See https://github.com/eandersson/amqpstorm', 'version': __version__}\n", "role": "user" }, { "content": "client_properties", "role": "assistant" } ]
def x(self, break_on_empty=False, to_tuple=False, auto_decode=True): self.check_for_errors() while not self.is_closed: message = self._build_message(auto_decode=auto_decode) if not message: self.check_for_errors() sleep(IDLE_WAIT) if break_on_empty and not self._inbound: break continue if to_tuple: yield message.to_tuple() continue yield message
build_inbound_messages
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, break_on_empty=False, to_tuple=False, auto_decode=True):\n self.check_for_errors()\n while not self.is_closed:\n message = self._build_message(auto_decode=auto_decode)\n if not message:\n self.check_for_errors()\n sleep(IDLE_WAIT)\n if break_on_empty and not self._inbound:\n break\n continue\n if to_tuple:\n yield message.to_tuple()\n continue\n yield message\n", "role": "user" }, { "content": "build_inbound_messages", "role": "assistant" } ]
def x(self, reply_code=200, reply_text=''): if not compatibility.is_integer(reply_code): raise AMQPInvalidArgument('reply_code should be an integer') elif not compatibility.is_string(reply_text): raise AMQPInvalidArgument('reply_text should be a string') try: if self._connection.is_closed or self.is_closed: self.stop_consuming() LOGGER.debug('Channel #%d forcefully Closed', self.channel_id) return self.set_state(self.CLOSING) LOGGER.debug('Channel #%d Closing', self.channel_id) try: self.stop_consuming() except AMQPChannelError: self.remove_consumer_tag() self.rpc_request(specification.Channel.Close(reply_code=reply_code, reply_text=reply_text), connection_adapter=self._connection) finally: if self._inbound: del self._inbound[:] self.set_state(self.CLOSED) if self._on_close_impl: self._on_close_impl(self.channel_id) LOGGER.debug('Channel #%d Closed', self.channel_id)
close
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, reply_code=200, reply_text=''):\n if not compatibility.is_integer(reply_code):\n raise AMQPInvalidArgument('reply_code should be an integer')\n elif not compatibility.is_string(reply_text):\n raise AMQPInvalidArgument('reply_text should be a string')\n try:\n if self._connection.is_closed or self.is_closed:\n self.stop_consuming()\n LOGGER.debug('Channel #%d forcefully Closed', self.channel_id)\n return\n self.set_state(self.CLOSING)\n LOGGER.debug('Channel #%d Closing', self.channel_id)\n try:\n self.stop_consuming()\n except AMQPChannelError:\n self.remove_consumer_tag()\n self.rpc_request(specification.Channel.Close(reply_code=reply_code,\n reply_text=reply_text), connection_adapter=self._connection)\n finally:\n if self._inbound:\n del self._inbound[:]\n self.set_state(self.CLOSED)\n if self._on_close_impl:\n self._on_close_impl(self.channel_id)\n LOGGER.debug('Channel #%d Closed', self.channel_id)\n", "role": "user" }, { "content": "close", "role": "assistant" } ]
def x(self): try: self._connection.check_for_errors() except AMQPConnectionError: self.set_state(self.CLOSED) raise if self.exceptions: exception = self.exceptions[0] if self.is_open: self.exceptions.pop(0) raise exception if self.is_closed: raise AMQPChannelError('channel was closed')
check_for_errors
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n try:\n self._connection.check_for_errors()\n except AMQPConnectionError:\n self.set_state(self.CLOSED)\n raise\n if self.exceptions:\n exception = self.exceptions[0]\n if self.is_open:\n self.exceptions.pop(0)\n raise exception\n if self.is_closed:\n raise AMQPChannelError('channel was closed')\n", "role": "user" }, { "content": "check_for_errors", "role": "assistant" } ]
def x(self): self._confirming_deliveries = True confirm_frame = specification.Confirm.Select() return self.rpc_request(confirm_frame)
confirm_deliveries
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n self._confirming_deliveries = True\n confirm_frame = specification.Confirm.Select()\n return self.rpc_request(confirm_frame)\n", "role": "user" }, { "content": "confirm_deliveries", "role": "assistant" } ]
def x(self, frame_in): if self.rpc.on_frame(frame_in): return if frame_in.name in CONTENT_FRAME: self._inbound.append(frame_in) elif frame_in.name == 'Basic.Cancel': self._basic_cancel(frame_in) elif frame_in.name == 'Basic.CancelOk': self.remove_consumer_tag(frame_in.consumer_tag) elif frame_in.name == 'Basic.ConsumeOk': self.add_consumer_tag(frame_in['consumer_tag']) elif frame_in.name == 'Basic.Return': self._basic_return(frame_in) elif frame_in.name == 'Channel.Close': self._close_channel(frame_in) elif frame_in.name == 'Channel.Flow': self.write_frame(specification.Channel.FlowOk(frame_in.active)) else: LOGGER.error('[Channel%d] Unhandled Frame: %s -- %s', self. channel_id, frame_in.name, dict(frame_in))
on_frame
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, frame_in):\n if self.rpc.on_frame(frame_in):\n return\n if frame_in.name in CONTENT_FRAME:\n self._inbound.append(frame_in)\n elif frame_in.name == 'Basic.Cancel':\n self._basic_cancel(frame_in)\n elif frame_in.name == 'Basic.CancelOk':\n self.remove_consumer_tag(frame_in.consumer_tag)\n elif frame_in.name == 'Basic.ConsumeOk':\n self.add_consumer_tag(frame_in['consumer_tag'])\n elif frame_in.name == 'Basic.Return':\n self._basic_return(frame_in)\n elif frame_in.name == 'Channel.Close':\n self._close_channel(frame_in)\n elif frame_in.name == 'Channel.Flow':\n self.write_frame(specification.Channel.FlowOk(frame_in.active))\n else:\n LOGGER.error('[Channel%d] Unhandled Frame: %s -- %s', self.\n channel_id, frame_in.name, dict(frame_in))\n", "role": "user" }, { "content": "on_frame", "role": "assistant" } ]
def x(self): self._inbound = [] self._exceptions = [] self.set_state(self.OPENING) self.rpc_request(specification.Channel.Open()) self.set_state(self.OPEN)
open
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n self._inbound = []\n self._exceptions = []\n self.set_state(self.OPENING)\n self.rpc_request(specification.Channel.Open())\n self.set_state(self.OPEN)\n", "role": "user" }, { "content": "open", "role": "assistant" } ]
def x(self, to_tuple=False, auto_decode=True): if not self._consumer_callbacks: raise AMQPChannelError('no consumer callback defined') for message in self.build_inbound_messages(break_on_empty=True, auto_decode=auto_decode): consumer_tag = message._method.get('consumer_tag') if to_tuple: self._consumer_callbacks[consumer_tag](*message.to_tuple()) continue self._consumer_callbacks[consumer_tag](message)
process_data_events
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, to_tuple=False, auto_decode=True):\n if not self._consumer_callbacks:\n raise AMQPChannelError('no consumer callback defined')\n for message in self.build_inbound_messages(break_on_empty=True,\n auto_decode=auto_decode):\n consumer_tag = message._method.get('consumer_tag')\n if to_tuple:\n self._consumer_callbacks[consumer_tag](*message.to_tuple())\n continue\n self._consumer_callbacks[consumer_tag](message)\n", "role": "user" }, { "content": "process_data_events", "role": "assistant" } ]
def x(self, frame_out, connection_adapter=None): with self.rpc.lock: uuid = self.rpc.register_request(frame_out.valid_responses) self._connection.write_frame(self.channel_id, frame_out) return self.rpc.get_request(uuid, connection_adapter=connection_adapter )
rpc_request
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, frame_out, connection_adapter=None):\n with self.rpc.lock:\n uuid = self.rpc.register_request(frame_out.valid_responses)\n self._connection.write_frame(self.channel_id, frame_out)\n return self.rpc.get_request(uuid, connection_adapter=connection_adapter\n )\n", "role": "user" }, { "content": "rpc_request", "role": "assistant" } ]
def x(self, to_tuple=False, auto_decode=True): while not self.is_closed: self.process_data_events(to_tuple=to_tuple, auto_decode=auto_decode) if self.consumer_tags: sleep(IDLE_WAIT) continue break
start_consuming
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, to_tuple=False, auto_decode=True):\n while not self.is_closed:\n self.process_data_events(to_tuple=to_tuple, auto_decode=auto_decode)\n if self.consumer_tags:\n sleep(IDLE_WAIT)\n continue\n break\n", "role": "user" }, { "content": "start_consuming", "role": "assistant" } ]
def x(self): if not self.consumer_tags: return if not self.is_closed: for tag in self.consumer_tags: self.basic.cancel(tag) self.remove_consumer_tag()
stop_consuming
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if not self.consumer_tags:\n return\n if not self.is_closed:\n for tag in self.consumer_tags:\n self.basic.cancel(tag)\n self.remove_consumer_tag()\n", "role": "user" }, { "content": "stop_consuming", "role": "assistant" } ]
def x(self, frame_out): self.check_for_errors() self._connection.write_frame(self.channel_id, frame_out)
write_frame
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, frame_out):\n self.check_for_errors()\n self._connection.write_frame(self.channel_id, frame_out)\n", "role": "user" }, { "content": "write_frame", "role": "assistant" } ]
def x(self, frames_out): self.check_for_errors() self._connection.write_frames(self.channel_id, frames_out)
write_frames
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, frames_out):\n self.check_for_errors()\n self._connection.write_frames(self.channel_id, frames_out)\n", "role": "user" }, { "content": "write_frames", "role": "assistant" } ]
def x(self, frame_in): LOGGER.warning('Received Basic.Cancel on consumer_tag: %s', try_utf8_decode(frame_in.consumer_tag)) self.remove_consumer_tag(frame_in.consumer_tag)
basic_cancel
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, frame_in):\n LOGGER.warning('Received Basic.Cancel on consumer_tag: %s',\n try_utf8_decode(frame_in.consumer_tag))\n self.remove_consumer_tag(frame_in.consumer_tag)\n", "role": "user" }, { "content": "basic_cancel", "role": "assistant" } ]
def x(self, frame_in): reply_text = try_utf8_decode(frame_in.reply_text) message = ( "Message not delivered: %s (%s) to queue '%s' from exchange '%s'" % (reply_text, frame_in.reply_code, frame_in.routing_key, frame_in. exchange)) exception = AMQPMessageError(message, reply_code=frame_in.reply_code) self.exceptions.append(exception)
basic_return
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, frame_in):\n reply_text = try_utf8_decode(frame_in.reply_text)\n message = (\n \"Message not delivered: %s (%s) to queue '%s' from exchange '%s'\" %\n (reply_text, frame_in.reply_code, frame_in.routing_key, frame_in.\n exchange))\n exception = AMQPMessageError(message, reply_code=frame_in.reply_code)\n self.exceptions.append(exception)\n", "role": "user" }, { "content": "basic_return", "role": "assistant" } ]
def x(self, auto_decode): with self.lock: if len(self._inbound) < 2: return None headers = self._build_message_headers() if not headers: return None basic_deliver, content_header = headers body = self._build_message_body(content_header.body_size) message = Message(channel=self, body=body, method=dict(basic_deliver), properties=dict(content_header.properties), auto_decode=auto_decode) return message
build_message
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, auto_decode):\n with self.lock:\n if len(self._inbound) < 2:\n return None\n headers = self._build_message_headers()\n if not headers:\n return None\n basic_deliver, content_header = headers\n body = self._build_message_body(content_header.body_size)\n message = Message(channel=self, body=body, method=dict(basic_deliver),\n properties=dict(content_header.properties), auto_decode=auto_decode)\n return message\n", "role": "user" }, { "content": "build_message", "role": "assistant" } ]
def x(self): basic_deliver = self._inbound.pop(0) if not isinstance(basic_deliver, specification.Basic.Deliver): LOGGER.warning( 'Received an out-of-order frame: %s was expecting a Basic.Deliver frame' , type(basic_deliver)) return None content_header = self._inbound.pop(0) if not isinstance(content_header, ContentHeader): LOGGER.warning( 'Received an out-of-order frame: %s was expecting a ContentHeader frame' , type(content_header)) return None return basic_deliver, content_header
build_message_headers
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n basic_deliver = self._inbound.pop(0)\n if not isinstance(basic_deliver, specification.Basic.Deliver):\n LOGGER.warning(\n 'Received an out-of-order frame: %s was expecting a Basic.Deliver frame'\n , type(basic_deliver))\n return None\n content_header = self._inbound.pop(0)\n if not isinstance(content_header, ContentHeader):\n LOGGER.warning(\n 'Received an out-of-order frame: %s was expecting a ContentHeader frame'\n , type(content_header))\n return None\n return basic_deliver, content_header\n", "role": "user" }, { "content": "build_message_headers", "role": "assistant" } ]
def x(self, body_size): body = bytes() while len(body) < body_size: if not self._inbound: self.check_for_errors() sleep(IDLE_WAIT) continue body_piece = self._inbound.pop(0) if not body_piece.value: break body += body_piece.value return body
build_message_body
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, body_size):\n body = bytes()\n while len(body) < body_size:\n if not self._inbound:\n self.check_for_errors()\n sleep(IDLE_WAIT)\n continue\n body_piece = self._inbound.pop(0)\n if not body_piece.value:\n break\n body += body_piece.value\n return body\n", "role": "user" }, { "content": "build_message_body", "role": "assistant" } ]
def x(self, frame_in): if frame_in.reply_code != 200: reply_text = try_utf8_decode(frame_in.reply_text) message = 'Channel %d was closed by remote server: %s' % (self. _channel_id, reply_text) exception = AMQPChannelError(message, reply_code=frame_in.reply_code) self.exceptions.append(exception) self.set_state(self.CLOSED) if self._connection.is_open: try: self._connection.write_frame(self.channel_id, specification. Channel.CloseOk()) except AMQPConnectionError: pass self.close()
close_channel
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, frame_in):\n if frame_in.reply_code != 200:\n reply_text = try_utf8_decode(frame_in.reply_text)\n message = 'Channel %d was closed by remote server: %s' % (self.\n _channel_id, reply_text)\n exception = AMQPChannelError(message, reply_code=frame_in.reply_code)\n self.exceptions.append(exception)\n self.set_state(self.CLOSED)\n if self._connection.is_open:\n try:\n self._connection.write_frame(self.channel_id, specification.\n Channel.CloseOk())\n except AMQPConnectionError:\n pass\n self.close()\n", "role": "user" }, { "content": "close_channel", "role": "assistant" } ]
def x(self, username, password, tags=''): user_payload = json.dumps({'password': password, 'tags': tags}) return self.http_client.put(API_USER % username, payload=user_payload)
create
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, username, password, tags=''):\n user_payload = json.dumps({'password': password, 'tags': tags})\n return self.http_client.put(API_USER % username, payload=user_payload)\n", "role": "user" }, { "content": "create", "role": "assistant" } ]
def x(self, username, virtual_host): virtual_host = quote(virtual_host, '') return self.http_client.get(API_USER_VIRTUAL_HOST_PERMISSIONS % ( virtual_host, username))
get_permission
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, username, virtual_host):\n virtual_host = quote(virtual_host, '')\n return self.http_client.get(API_USER_VIRTUAL_HOST_PERMISSIONS % (\n virtual_host, username))\n", "role": "user" }, { "content": "get_permission", "role": "assistant" } ]
def x(self, username, virtual_host, configure_regex='.*', write_regex='.*', read_regex='.*'): virtual_host = quote(virtual_host, '') permission_payload = json.dumps({'configure': configure_regex, 'read': read_regex, 'write': write_regex}) return self.http_client.put(API_USER_VIRTUAL_HOST_PERMISSIONS % ( virtual_host, username), payload=permission_payload)
set_permission
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, username, virtual_host, configure_regex='.*', write_regex='.*',\n read_regex='.*'):\n virtual_host = quote(virtual_host, '')\n permission_payload = json.dumps({'configure': configure_regex, 'read':\n read_regex, 'write': write_regex})\n return self.http_client.put(API_USER_VIRTUAL_HOST_PERMISSIONS % (\n virtual_host, username), payload=permission_payload)\n", "role": "user" }, { "content": "set_permission", "role": "assistant" } ]
def x(self, username, virtual_host): virtual_host = quote(virtual_host, '') return self.http_client.delete(API_USER_VIRTUAL_HOST_PERMISSIONS % ( virtual_host, username))
delete_permission
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, username, virtual_host):\n virtual_host = quote(virtual_host, '')\n return self.http_client.delete(API_USER_VIRTUAL_HOST_PERMISSIONS % (\n virtual_host, username))\n", "role": "user" }, { "content": "delete_permission", "role": "assistant" } ]
def x(self): self._stopped.clear() if not self._connection or self._connection.is_closed: self._create_connection() while not self._stopped.is_set(): try: self._connection.check_for_errors() self._update_consumers() except amqpstorm.AMQPError as why: LOGGER.warning(why) self._stop_consumers() self._create_connection() time.sleep(1)
start_server
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n self._stopped.clear()\n if not self._connection or self._connection.is_closed:\n self._create_connection()\n while not self._stopped.is_set():\n try:\n self._connection.check_for_errors()\n self._update_consumers()\n except amqpstorm.AMQPError as why:\n LOGGER.warning(why)\n self._stop_consumers()\n self._create_connection()\n time.sleep(1)\n", "role": "user" }, { "content": "start_server", "role": "assistant" } ]
def x(self): consumer_to_start = min(max(self.number_of_consumers - len(self. _consumers), 0), 2) for _ in range(consumer_to_start): consumer = Consumer(self.rpc_queue) self._start_consumer(consumer) self._consumers.append(consumer) for consumer in self._consumers: if consumer.active: continue self._start_consumer(consumer) break self._stop_consumers(self.number_of_consumers)
update_consumers
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n consumer_to_start = min(max(self.number_of_consumers - len(self.\n _consumers), 0), 2)\n for _ in range(consumer_to_start):\n consumer = Consumer(self.rpc_queue)\n self._start_consumer(consumer)\n self._consumers.append(consumer)\n for consumer in self._consumers:\n if consumer.active:\n continue\n self._start_consumer(consumer)\n break\n self._stop_consumers(self.number_of_consumers)\n", "role": "user" }, { "content": "update_consumers", "role": "assistant" } ]
def x(self, path, payload=None, headers=None): return self._request('get', path, payload, headers)
get
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, path, payload=None, headers=None):\n return self._request('get', path, payload, headers)\n", "role": "user" }, { "content": "get", "role": "assistant" } ]
def x(self, path, payload=None, headers=None): return self._request('post', path, payload, headers)
post
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, path, payload=None, headers=None):\n return self._request('post', path, payload, headers)\n", "role": "user" }, { "content": "post", "role": "assistant" } ]
def x(self, path, payload=None, headers=None): return self._request('delete', path, payload, headers)
delete
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, path, payload=None, headers=None):\n return self._request('delete', path, payload, headers)\n", "role": "user" }, { "content": "delete", "role": "assistant" } ]
def x(self, path, payload=None, headers=None): return self._request('put', path, payload, headers)
put
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, path, payload=None, headers=None):\n return self._request('put', path, payload, headers)\n", "role": "user" }, { "content": "put", "role": "assistant" } ]
def x(self, method, path, payload=None, headers=None): url = urlparse.urljoin(self._base_url, 'api/%s' % path) headers = headers or {} headers['content-type'] = 'application/json' try: response = requests.request(method, url, auth=self._auth, data= payload, headers=headers, cert=self._cert, verify=self._verify, timeout=self._timeout) except requests.RequestException as why: raise ApiConnectionError(str(why)) json_response = self._get_json_output(response) self._check_for_errors(response, json_response) return json_response
request
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, method, path, payload=None, headers=None):\n url = urlparse.urljoin(self._base_url, 'api/%s' % path)\n headers = headers or {}\n headers['content-type'] = 'application/json'\n try:\n response = requests.request(method, url, auth=self._auth, data=\n payload, headers=headers, cert=self._cert, verify=self._verify,\n timeout=self._timeout)\n except requests.RequestException as why:\n raise ApiConnectionError(str(why))\n json_response = self._get_json_output(response)\n self._check_for_errors(response, json_response)\n return json_response\n", "role": "user" }, { "content": "request", "role": "assistant" } ]
def x(response, json_response): status_code = response.status_code try: response.raise_for_status() except requests.HTTPError as why: raise ApiError(str(why), reply_code=status_code) if isinstance(json_response, dict) and 'error' in json_response: raise ApiError(json_response['error'], reply_code=status_code)
check_for_errors
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(response, json_response):\n status_code = response.status_code\n try:\n response.raise_for_status()\n except requests.HTTPError as why:\n raise ApiError(str(why), reply_code=status_code)\n if isinstance(json_response, dict) and 'error' in json_response:\n raise ApiError(json_response['error'], reply_code=status_code)\n", "role": "user" }, { "content": "check_for_errors", "role": "assistant" } ]
def x(self, node=None): if not node: return self.http_client.get(HEALTHCHECKS) return self.http_client.get(HEALTHCHECKS_NODE % node)
get
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, node=None):\n if not node:\n return self.http_client.get(HEALTHCHECKS)\n return self.http_client.get(HEALTHCHECKS_NODE % node)\n", "role": "user" }, { "content": "get", "role": "assistant" } ]
def x(ac, an=None, fill=np.nan): ac = asarray_ndim(ac, 2) if an is None: an = np.sum(ac, axis=1) else: an = asarray_ndim(an, 1) check_dim0_aligned(ac, an) n_pairs = an * (an - 1) / 2 n_same = np.sum(ac * (ac - 1) / 2, axis=1) n_diff = n_pairs - n_same with ignore_invalid(): mpd = np.where(n_pairs > 0, n_diff / n_pairs, fill) return mpd
mean_pairwise_difference
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(ac, an=None, fill=np.nan):\n ac = asarray_ndim(ac, 2)\n if an is None:\n an = np.sum(ac, axis=1)\n else:\n an = asarray_ndim(an, 1)\n check_dim0_aligned(ac, an)\n n_pairs = an * (an - 1) / 2\n n_same = np.sum(ac * (ac - 1) / 2, axis=1)\n n_diff = n_pairs - n_same\n with ignore_invalid():\n mpd = np.where(n_pairs > 0, n_diff / n_pairs, fill)\n return mpd\n", "role": "user" }, { "content": "mean_pairwise_difference", "role": "assistant" } ]
def x(ac1, ac2, an1=None, an2=None, fill=np.nan): ac1 = asarray_ndim(ac1, 2) ac2 = asarray_ndim(ac2, 2) check_dim0_aligned(ac1, ac2) ac1, ac2 = ensure_dim1_aligned(ac1, ac2) if an1 is None: an1 = np.sum(ac1, axis=1) else: an1 = asarray_ndim(an1, 1) check_dim0_aligned(ac1, an1) if an2 is None: an2 = np.sum(ac2, axis=1) else: an2 = asarray_ndim(an2, 1) check_dim0_aligned(ac2, an2) n_pairs = an1 * an2 n_same = np.sum(ac1 * ac2, axis=1) n_diff = n_pairs - n_same with ignore_invalid(): mpd = np.where(n_pairs > 0, n_diff / n_pairs, fill) return mpd
mean_pairwise_difference_between
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(ac1, ac2, an1=None, an2=None, fill=np.nan):\n ac1 = asarray_ndim(ac1, 2)\n ac2 = asarray_ndim(ac2, 2)\n check_dim0_aligned(ac1, ac2)\n ac1, ac2 = ensure_dim1_aligned(ac1, ac2)\n if an1 is None:\n an1 = np.sum(ac1, axis=1)\n else:\n an1 = asarray_ndim(an1, 1)\n check_dim0_aligned(ac1, an1)\n if an2 is None:\n an2 = np.sum(ac2, axis=1)\n else:\n an2 = asarray_ndim(an2, 1)\n check_dim0_aligned(ac2, an2)\n n_pairs = an1 * an2\n n_same = np.sum(ac1 * ac2, axis=1)\n n_diff = n_pairs - n_same\n with ignore_invalid():\n mpd = np.where(n_pairs > 0, n_diff / n_pairs, fill)\n return mpd\n", "role": "user" }, { "content": "mean_pairwise_difference_between", "role": "assistant" } ]
def x(pos, ac, start=None, stop=None, is_accessible=None): if not isinstance(pos, SortedIndex): pos = SortedIndex(pos, copy=False) ac = asarray_ndim(ac, 2) is_accessible = asarray_ndim(is_accessible, 1, allow_none=True) if start is not None or stop is not None: loc = pos.locate_range(start, stop) pos = pos[loc] ac = ac[loc] if start is None: start = pos[0] if stop is None: stop = pos[-1] mpd = mean_pairwise_difference(ac, fill=0) mpd_sum = np.sum(mpd) if is_accessible is None: n_bases = stop - start + 1 else: n_bases = np.count_nonzero(is_accessible[start - 1:stop]) pi = mpd_sum / n_bases return pi
sequence_diversity
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(pos, ac, start=None, stop=None, is_accessible=None):\n if not isinstance(pos, SortedIndex):\n pos = SortedIndex(pos, copy=False)\n ac = asarray_ndim(ac, 2)\n is_accessible = asarray_ndim(is_accessible, 1, allow_none=True)\n if start is not None or stop is not None:\n loc = pos.locate_range(start, stop)\n pos = pos[loc]\n ac = ac[loc]\n if start is None:\n start = pos[0]\n if stop is None:\n stop = pos[-1]\n mpd = mean_pairwise_difference(ac, fill=0)\n mpd_sum = np.sum(mpd)\n if is_accessible is None:\n n_bases = stop - start + 1\n else:\n n_bases = np.count_nonzero(is_accessible[start - 1:stop])\n pi = mpd_sum / n_bases\n return pi\n", "role": "user" }, { "content": "sequence_diversity", "role": "assistant" } ]
def x(pos, ac1, ac2, an1=None, an2=None, start=None, stop=None, is_accessible=None): if not isinstance(pos, SortedIndex): pos = SortedIndex(pos, copy=False) ac1 = asarray_ndim(ac1, 2) ac2 = asarray_ndim(ac2, 2) if an1 is not None: an1 = asarray_ndim(an1, 1) if an2 is not None: an2 = asarray_ndim(an2, 1) is_accessible = asarray_ndim(is_accessible, 1, allow_none=True) if start is not None or stop is not None: loc = pos.locate_range(start, stop) pos = pos[loc] ac1 = ac1[loc] ac2 = ac2[loc] if an1 is not None: an1 = an1[loc] if an2 is not None: an2 = an2[loc] if start is None: start = pos[0] if stop is None: stop = pos[-1] mpd = mean_pairwise_difference_between(ac1, ac2, an1=an1, an2=an2, fill=0) mpd_sum = np.sum(mpd) if is_accessible is None: n_bases = stop - start + 1 else: n_bases = np.count_nonzero(is_accessible[start - 1:stop]) dxy = mpd_sum / n_bases return dxy
sequence_divergence
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(pos, ac1, ac2, an1=None, an2=None, start=None, stop=None,\n is_accessible=None):\n if not isinstance(pos, SortedIndex):\n pos = SortedIndex(pos, copy=False)\n ac1 = asarray_ndim(ac1, 2)\n ac2 = asarray_ndim(ac2, 2)\n if an1 is not None:\n an1 = asarray_ndim(an1, 1)\n if an2 is not None:\n an2 = asarray_ndim(an2, 1)\n is_accessible = asarray_ndim(is_accessible, 1, allow_none=True)\n if start is not None or stop is not None:\n loc = pos.locate_range(start, stop)\n pos = pos[loc]\n ac1 = ac1[loc]\n ac2 = ac2[loc]\n if an1 is not None:\n an1 = an1[loc]\n if an2 is not None:\n an2 = an2[loc]\n if start is None:\n start = pos[0]\n if stop is None:\n stop = pos[-1]\n mpd = mean_pairwise_difference_between(ac1, ac2, an1=an1, an2=an2, fill=0)\n mpd_sum = np.sum(mpd)\n if is_accessible is None:\n n_bases = stop - start + 1\n else:\n n_bases = np.count_nonzero(is_accessible[start - 1:stop])\n dxy = mpd_sum / n_bases\n return dxy\n", "role": "user" }, { "content": "sequence_divergence", "role": "assistant" } ]
def x(pos, ac, size=None, start=None, stop=None, step=None, windows=None, is_accessible=None, fill=np.nan): if not isinstance(pos, SortedIndex): pos = SortedIndex(pos, copy=False) is_accessible = asarray_ndim(is_accessible, 1, allow_none=True) mpd = mean_pairwise_difference(ac, fill=0) mpd_sum, windows, counts = windowed_statistic(pos, values=mpd, statistic=np.sum, size=size, start=start, stop=stop, step=step, windows=windows, fill=0) pi, n_bases = per_base(mpd_sum, windows, is_accessible=is_accessible, fill=fill) return pi, windows, n_bases, counts
windowed_diversity
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(pos, ac, size=None, start=None, stop=None, step=None, windows=None,\n is_accessible=None, fill=np.nan):\n if not isinstance(pos, SortedIndex):\n pos = SortedIndex(pos, copy=False)\n is_accessible = asarray_ndim(is_accessible, 1, allow_none=True)\n mpd = mean_pairwise_difference(ac, fill=0)\n mpd_sum, windows, counts = windowed_statistic(pos, values=mpd,\n statistic=np.sum, size=size, start=start, stop=stop, step=step,\n windows=windows, fill=0)\n pi, n_bases = per_base(mpd_sum, windows, is_accessible=is_accessible,\n fill=fill)\n return pi, windows, n_bases, counts\n", "role": "user" }, { "content": "windowed_diversity", "role": "assistant" } ]
def x(pos, ac1, ac2, size=None, start=None, stop=None, step=None, windows= None, is_accessible=None, fill=np.nan): pos = SortedIndex(pos, copy=False) is_accessible = asarray_ndim(is_accessible, 1, allow_none=True) mpd = mean_pairwise_difference_between(ac1, ac2, fill=0) mpd_sum, windows, counts = windowed_statistic(pos, values=mpd, statistic=np.sum, size=size, start=start, stop=stop, step=step, windows=windows, fill=0) dxy, n_bases = per_base(mpd_sum, windows, is_accessible=is_accessible, fill=fill) return dxy, windows, n_bases, counts
windowed_divergence
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(pos, ac1, ac2, size=None, start=None, stop=None, step=None, windows=\n None, is_accessible=None, fill=np.nan):\n pos = SortedIndex(pos, copy=False)\n is_accessible = asarray_ndim(is_accessible, 1, allow_none=True)\n mpd = mean_pairwise_difference_between(ac1, ac2, fill=0)\n mpd_sum, windows, counts = windowed_statistic(pos, values=mpd,\n statistic=np.sum, size=size, start=start, stop=stop, step=step,\n windows=windows, fill=0)\n dxy, n_bases = per_base(mpd_sum, windows, is_accessible=is_accessible,\n fill=fill)\n return dxy, windows, n_bases, counts\n", "role": "user" }, { "content": "windowed_divergence", "role": "assistant" } ]
def x(pos, ac1, ac2, size=None, start=None, stop=None, step=None, windows= None, is_accessible=None, fill=np.nan): pos = SortedIndex(pos, copy=False) is_accessible = asarray_ndim(is_accessible, 1, allow_none=True) loc_df = locate_fixed_differences(ac1, ac2) n_df, windows, counts = windowed_statistic(pos, values=loc_df, statistic=np.count_nonzero, size=size, start=start, stop=stop, step =step, windows=windows, fill=0) df, n_bases = per_base(n_df, windows, is_accessible=is_accessible, fill =fill) return df, windows, n_bases, counts
windowed_df
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(pos, ac1, ac2, size=None, start=None, stop=None, step=None, windows=\n None, is_accessible=None, fill=np.nan):\n pos = SortedIndex(pos, copy=False)\n is_accessible = asarray_ndim(is_accessible, 1, allow_none=True)\n loc_df = locate_fixed_differences(ac1, ac2)\n n_df, windows, counts = windowed_statistic(pos, values=loc_df,\n statistic=np.count_nonzero, size=size, start=start, stop=stop, step\n =step, windows=windows, fill=0)\n df, n_bases = per_base(n_df, windows, is_accessible=is_accessible, fill\n =fill)\n return df, windows, n_bases, counts\n", "role": "user" }, { "content": "windowed_df", "role": "assistant" } ]
def x(pos, ac, start=None, stop=None, is_accessible=None): if not isinstance(pos, SortedIndex): pos = SortedIndex(pos, copy=False) is_accessible = asarray_ndim(is_accessible, 1, allow_none=True) if not hasattr(ac, 'count_segregating'): ac = AlleleCountsArray(ac, copy=False) if start is not None or stop is not None: loc = pos.locate_range(start, stop) pos = pos[loc] ac = ac[loc] if start is None: start = pos[0] if stop is None: stop = pos[-1] S = ac.count_segregating() n = ac.sum(axis=1).max() a1 = np.sum(1 / np.arange(1, n)) theta_hat_w_abs = S / a1 if is_accessible is None: n_bases = stop - start + 1 else: n_bases = np.count_nonzero(is_accessible[start - 1:stop]) theta_hat_w = theta_hat_w_abs / n_bases return theta_hat_w
watterson_theta
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(pos, ac, start=None, stop=None, is_accessible=None):\n if not isinstance(pos, SortedIndex):\n pos = SortedIndex(pos, copy=False)\n is_accessible = asarray_ndim(is_accessible, 1, allow_none=True)\n if not hasattr(ac, 'count_segregating'):\n ac = AlleleCountsArray(ac, copy=False)\n if start is not None or stop is not None:\n loc = pos.locate_range(start, stop)\n pos = pos[loc]\n ac = ac[loc]\n if start is None:\n start = pos[0]\n if stop is None:\n stop = pos[-1]\n S = ac.count_segregating()\n n = ac.sum(axis=1).max()\n a1 = np.sum(1 / np.arange(1, n))\n theta_hat_w_abs = S / a1\n if is_accessible is None:\n n_bases = stop - start + 1\n else:\n n_bases = np.count_nonzero(is_accessible[start - 1:stop])\n theta_hat_w = theta_hat_w_abs / n_bases\n return theta_hat_w\n", "role": "user" }, { "content": "watterson_theta", "role": "assistant" } ]
def x(pos, ac, size=None, start=None, stop=None, step=None, windows=None, is_accessible=None, fill=np.nan): if not isinstance(pos, SortedIndex): pos = SortedIndex(pos, copy=False) is_accessible = asarray_ndim(is_accessible, 1, allow_none=True) if not hasattr(ac, 'count_segregating'): ac = AlleleCountsArray(ac, copy=False) is_seg = ac.is_segregating() S, windows, counts = windowed_statistic(pos, is_seg, statistic=np. count_nonzero, size=size, start=start, stop=stop, step=step, windows=windows, fill=0) n = ac.sum(axis=1).max() a1 = np.sum(1 / np.arange(1, n)) theta_hat_w_abs = S / a1 theta_hat_w, n_bases = per_base(theta_hat_w_abs, windows=windows, is_accessible=is_accessible, fill=fill) return theta_hat_w, windows, n_bases, counts
windowed_watterson_theta
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(pos, ac, size=None, start=None, stop=None, step=None, windows=None,\n is_accessible=None, fill=np.nan):\n if not isinstance(pos, SortedIndex):\n pos = SortedIndex(pos, copy=False)\n is_accessible = asarray_ndim(is_accessible, 1, allow_none=True)\n if not hasattr(ac, 'count_segregating'):\n ac = AlleleCountsArray(ac, copy=False)\n is_seg = ac.is_segregating()\n S, windows, counts = windowed_statistic(pos, is_seg, statistic=np.\n count_nonzero, size=size, start=start, stop=stop, step=step,\n windows=windows, fill=0)\n n = ac.sum(axis=1).max()\n a1 = np.sum(1 / np.arange(1, n))\n theta_hat_w_abs = S / a1\n theta_hat_w, n_bases = per_base(theta_hat_w_abs, windows=windows,\n is_accessible=is_accessible, fill=fill)\n return theta_hat_w, windows, n_bases, counts\n", "role": "user" }, { "content": "windowed_watterson_theta", "role": "assistant" } ]
def x(ac, pos=None, start=None, stop=None, min_sites=3): if not hasattr(ac, 'count_segregating'): ac = AlleleCountsArray(ac, copy=False) if pos is not None and (start is not None or stop is not None): if not isinstance(pos, SortedIndex): pos = SortedIndex(pos, copy=False) loc = pos.locate_range(start, stop) ac = ac[loc] S = ac.count_segregating() if S < min_sites: return np.nan n = ac.sum(axis=1).max() a1 = np.sum(1 / np.arange(1, n)) theta_hat_w_abs = S / a1 mpd = mean_pairwise_difference(ac, fill=0) theta_hat_pi_abs = np.sum(mpd) d = theta_hat_pi_abs - theta_hat_w_abs a2 = np.sum(1 / np.arange(1, n) ** 2) b1 = (n + 1) / (3 * (n - 1)) b2 = 2 * (n ** 2 + n + 3) / (9 * n * (n - 1)) c1 = b1 - 1 / a1 c2 = b2 - (n + 2) / (a1 * n) + a2 / a1 ** 2 e1 = c1 / a1 e2 = c2 / (a1 ** 2 + a2) d_stdev = np.sqrt(e1 * S + e2 * S * (S - 1)) D = d / d_stdev return D
tajima_d
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(ac, pos=None, start=None, stop=None, min_sites=3):\n if not hasattr(ac, 'count_segregating'):\n ac = AlleleCountsArray(ac, copy=False)\n if pos is not None and (start is not None or stop is not None):\n if not isinstance(pos, SortedIndex):\n pos = SortedIndex(pos, copy=False)\n loc = pos.locate_range(start, stop)\n ac = ac[loc]\n S = ac.count_segregating()\n if S < min_sites:\n return np.nan\n n = ac.sum(axis=1).max()\n a1 = np.sum(1 / np.arange(1, n))\n theta_hat_w_abs = S / a1\n mpd = mean_pairwise_difference(ac, fill=0)\n theta_hat_pi_abs = np.sum(mpd)\n d = theta_hat_pi_abs - theta_hat_w_abs\n a2 = np.sum(1 / np.arange(1, n) ** 2)\n b1 = (n + 1) / (3 * (n - 1))\n b2 = 2 * (n ** 2 + n + 3) / (9 * n * (n - 1))\n c1 = b1 - 1 / a1\n c2 = b2 - (n + 2) / (a1 * n) + a2 / a1 ** 2\n e1 = c1 / a1\n e2 = c2 / (a1 ** 2 + a2)\n d_stdev = np.sqrt(e1 * S + e2 * S * (S - 1))\n D = d / d_stdev\n return D\n", "role": "user" }, { "content": "tajima_d", "role": "assistant" } ]
def x(pos, ac, size=None, start=None, stop=None, step=None, windows=None, min_sites=3): if not isinstance(pos, SortedIndex): pos = SortedIndex(pos, copy=False) if not hasattr(ac, 'count_segregating'): ac = AlleleCountsArray(ac, copy=False) n = ac.sum(axis=1).max() a1 = np.sum(1 / np.arange(1, n)) a2 = np.sum(1 / np.arange(1, n) ** 2) b1 = (n + 1) / (3 * (n - 1)) b2 = 2 * (n ** 2 + n + 3) / (9 * n * (n - 1)) c1 = b1 - 1 / a1 c2 = b2 - (n + 2) / (a1 * n) + a2 / a1 ** 2 e1 = c1 / a1 e2 = c2 / (a1 ** 2 + a2) is_seg = ac.is_segregating() mpd = mean_pairwise_difference(ac, fill=0) def x(w_is_seg, w_mpd): S = np.count_nonzero(w_is_seg) if S < min_sites: return np.nan pi = np.sum(w_mpd) d = pi - S / a1 d_stdev = np.sqrt(e1 * S + e2 * S * (S - 1)) wD = d / d_stdev return wD D, windows, counts = windowed_statistic(pos, values=(is_seg, mpd), statistic=statistic, size=size, start=start, stop=stop, step=step, windows=windows, fill=np.nan) return D, windows, counts
windowed_tajima_d
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(pos, ac, size=None, start=None, stop=None, step=None, windows=None,\n min_sites=3):\n if not isinstance(pos, SortedIndex):\n pos = SortedIndex(pos, copy=False)\n if not hasattr(ac, 'count_segregating'):\n ac = AlleleCountsArray(ac, copy=False)\n n = ac.sum(axis=1).max()\n a1 = np.sum(1 / np.arange(1, n))\n a2 = np.sum(1 / np.arange(1, n) ** 2)\n b1 = (n + 1) / (3 * (n - 1))\n b2 = 2 * (n ** 2 + n + 3) / (9 * n * (n - 1))\n c1 = b1 - 1 / a1\n c2 = b2 - (n + 2) / (a1 * n) + a2 / a1 ** 2\n e1 = c1 / a1\n e2 = c2 / (a1 ** 2 + a2)\n is_seg = ac.is_segregating()\n mpd = mean_pairwise_difference(ac, fill=0)\n\n def x(w_is_seg, w_mpd):\n S = np.count_nonzero(w_is_seg)\n if S < min_sites:\n return np.nan\n pi = np.sum(w_mpd)\n d = pi - S / a1\n d_stdev = np.sqrt(e1 * S + e2 * S * (S - 1))\n wD = d / d_stdev\n return wD\n D, windows, counts = windowed_statistic(pos, values=(is_seg, mpd),\n statistic=statistic, size=size, start=start, stop=stop, step=step,\n windows=windows, fill=np.nan)\n return D, windows, counts\n", "role": "user" }, { "content": "windowed_tajima_d", "role": "assistant" } ]
def x(ac, size, start=0, stop=None, step=None, min_sites=3): d = moving_statistic(values=ac, statistic=tajima_d, size=size, start= start, stop=stop, step=step, min_sites=min_sites) return d
moving_tajima_d
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(ac, size, start=0, stop=None, step=None, min_sites=3):\n d = moving_statistic(values=ac, statistic=tajima_d, size=size, start=\n start, stop=stop, step=step, min_sites=min_sites)\n return d\n", "role": "user" }, { "content": "moving_tajima_d", "role": "assistant" } ]
def x(gv, pos, phet_roh=0.001, phet_nonroh=(0.0025, 0.01), transition=1e-06, min_roh=0, is_accessible=None, contig_size=None): from hmmlearn import hmm if isinstance(phet_nonroh, float): phet_nonroh = phet_nonroh, gv = GenotypeVector(gv) pos = asarray_ndim(pos, 1) check_dim0_aligned(gv, pos) is_accessible = asarray_ndim(is_accessible, 1, dtype=bool) het_px = np.concatenate([(phet_roh,), phet_nonroh]) start_prob = np.repeat(1 / het_px.size, het_px.size) transition_mx = _hmm_derive_transition_matrix(transition, het_px.size) if is_accessible is None: if contig_size is None: raise ValueError( 'If is_accessibile argument is not provided, you must provide contig_size' ) p_accessible = 1.0 else: p_accessible = is_accessible.mean() contig_size = is_accessible.size emission_mx = _mhmm_derive_emission_matrix(het_px, p_accessible) roh_hmm = hmm.MultinomialHMM(n_components=het_px.size) roh_hmm.n_symbols_ = 3 roh_hmm.startprob_ = start_prob roh_hmm.transmat_ = transition_mx roh_hmm.emissionprob_ = emission_mx is_het = gv.is_het() pred, obs = _mhmm_predict_roh_state(roh_hmm, is_het, pos, is_accessible, contig_size) df_blocks = tabulate_state_blocks(pred, states=list(range(len(het_px)))) df_roh = df_blocks[df_blocks.state == 0].reset_index(drop=True) for col in ('state', 'support', 'start_lidx', 'stop_ridx', 'size_max'): del df_roh[col] df_roh.rename(columns={'start_ridx': 'start', 'stop_lidx': 'stop', 'size_min': 'length'}, inplace=True) df_roh['start'] = df_roh['start'] + 1 df_roh['stop'] = df_roh['stop'] + 1 if min_roh > 0: df_roh = df_roh[df_roh.length >= min_roh] froh = df_roh.length.sum() / contig_size return df_roh, froh
roh_mhmm
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(gv, pos, phet_roh=0.001, phet_nonroh=(0.0025, 0.01), transition=1e-06,\n min_roh=0, is_accessible=None, contig_size=None):\n from hmmlearn import hmm\n if isinstance(phet_nonroh, float):\n phet_nonroh = phet_nonroh,\n gv = GenotypeVector(gv)\n pos = asarray_ndim(pos, 1)\n check_dim0_aligned(gv, pos)\n is_accessible = asarray_ndim(is_accessible, 1, dtype=bool)\n het_px = np.concatenate([(phet_roh,), phet_nonroh])\n start_prob = np.repeat(1 / het_px.size, het_px.size)\n transition_mx = _hmm_derive_transition_matrix(transition, het_px.size)\n if is_accessible is None:\n if contig_size is None:\n raise ValueError(\n 'If is_accessibile argument is not provided, you must provide contig_size'\n )\n p_accessible = 1.0\n else:\n p_accessible = is_accessible.mean()\n contig_size = is_accessible.size\n emission_mx = _mhmm_derive_emission_matrix(het_px, p_accessible)\n roh_hmm = hmm.MultinomialHMM(n_components=het_px.size)\n roh_hmm.n_symbols_ = 3\n roh_hmm.startprob_ = start_prob\n roh_hmm.transmat_ = transition_mx\n roh_hmm.emissionprob_ = emission_mx\n is_het = gv.is_het()\n pred, obs = _mhmm_predict_roh_state(roh_hmm, is_het, pos, is_accessible,\n contig_size)\n df_blocks = tabulate_state_blocks(pred, states=list(range(len(het_px))))\n df_roh = df_blocks[df_blocks.state == 0].reset_index(drop=True)\n for col in ('state', 'support', 'start_lidx', 'stop_ridx', 'size_max'):\n del df_roh[col]\n df_roh.rename(columns={'start_ridx': 'start', 'stop_lidx': 'stop',\n 'size_min': 'length'}, inplace=True)\n df_roh['start'] = df_roh['start'] + 1\n df_roh['stop'] = df_roh['stop'] + 1\n if min_roh > 0:\n df_roh = df_roh[df_roh.length >= min_roh]\n froh = df_roh.length.sum() / contig_size\n return df_roh, froh\n", "role": "user" }, { "content": "roh_mhmm", "role": "assistant" } ]
def x(gv, pos, phet_roh=0.001, phet_nonroh=(0.0025, 0.01), transition=0.001, window_size=1000, min_roh=0, is_accessible=None, contig_size=None): from pomegranate import HiddenMarkovModel, PoissonDistribution if is_accessible is None: if contig_size is None: raise ValueError( 'If is_accessibile argument is not provided, you must provide contig_size' ) is_accessible = np.ones((contig_size,), dtype='bool') else: contig_size = is_accessible.size eqw = equally_accessible_windows(is_accessible, window_size) ishet = GenotypeVector(gv).is_het() counts, wins, records = windowed_statistic(pos, ishet, np.sum, windows=eqw) het_px = np.concatenate([(phet_roh,), phet_nonroh]) start_prob = np.repeat(1 / het_px.size, het_px.size) transition_mx = _hmm_derive_transition_matrix(transition, het_px.size) dists = [PoissonDistribution(x * window_size) for x in het_px] model = HiddenMarkovModel.from_matrix(transition_probabilities= transition_mx, distributions=dists, starts=start_prob) prediction = np.array(model.predict(counts[:, None])) df_blocks = tabulate_state_blocks(prediction, states=list(range(len( het_px)))) df_roh = df_blocks[df_blocks.state == 0].reset_index(drop=True) df_roh['start'] = df_roh.start_ridx.apply(lambda y: eqw[y, 0]) df_roh['stop'] = df_roh.stop_lidx.apply(lambda y: eqw[y, 1]) df_roh['length'] = df_roh.stop - df_roh.start if min_roh > 0: df_roh = df_roh[df_roh.length >= min_roh] froh = df_roh.length.sum() / contig_size return df_roh[['start', 'stop', 'length', 'is_marginal']], froh
roh_poissonhmm
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(gv, pos, phet_roh=0.001, phet_nonroh=(0.0025, 0.01), transition=0.001,\n window_size=1000, min_roh=0, is_accessible=None, contig_size=None):\n from pomegranate import HiddenMarkovModel, PoissonDistribution\n if is_accessible is None:\n if contig_size is None:\n raise ValueError(\n 'If is_accessibile argument is not provided, you must provide contig_size'\n )\n is_accessible = np.ones((contig_size,), dtype='bool')\n else:\n contig_size = is_accessible.size\n eqw = equally_accessible_windows(is_accessible, window_size)\n ishet = GenotypeVector(gv).is_het()\n counts, wins, records = windowed_statistic(pos, ishet, np.sum, windows=eqw)\n het_px = np.concatenate([(phet_roh,), phet_nonroh])\n start_prob = np.repeat(1 / het_px.size, het_px.size)\n transition_mx = _hmm_derive_transition_matrix(transition, het_px.size)\n dists = [PoissonDistribution(x * window_size) for x in het_px]\n model = HiddenMarkovModel.from_matrix(transition_probabilities=\n transition_mx, distributions=dists, starts=start_prob)\n prediction = np.array(model.predict(counts[:, None]))\n df_blocks = tabulate_state_blocks(prediction, states=list(range(len(\n het_px))))\n df_roh = df_blocks[df_blocks.state == 0].reset_index(drop=True)\n df_roh['start'] = df_roh.start_ridx.apply(lambda y: eqw[y, 0])\n df_roh['stop'] = df_roh.stop_lidx.apply(lambda y: eqw[y, 1])\n df_roh['length'] = df_roh.stop - df_roh.start\n if min_roh > 0:\n df_roh = df_roh[df_roh.length >= min_roh]\n froh = df_roh.length.sum() / contig_size\n return df_roh[['start', 'stop', 'length', 'is_marginal']], froh\n", "role": "user" }, { "content": "roh_poissonhmm", "role": "assistant" } ]
def x(dac, n=None): dac, n = _check_dac_n(dac, n) dac = dac.astype(int, copy=False) x = n + 1 s = np.bincount(dac, minlength=x) return s
sfs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(dac, n=None):\n dac, n = _check_dac_n(dac, n)\n dac = dac.astype(int, copy=False)\n x = n + 1\n s = np.bincount(dac, minlength=x)\n return s\n", "role": "user" }, { "content": "sfs", "role": "assistant" } ]
def x(ac, n=None): ac, n = _check_ac_n(ac, n) mac = np.amin(ac, axis=1) mac = mac.astype(int, copy=False) x = n // 2 + 1 s = np.bincount(mac, minlength=x) return s
sfs_folded
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(ac, n=None):\n ac, n = _check_ac_n(ac, n)\n mac = np.amin(ac, axis=1)\n mac = mac.astype(int, copy=False)\n x = n // 2 + 1\n s = np.bincount(mac, minlength=x)\n return s\n", "role": "user" }, { "content": "sfs_folded", "role": "assistant" } ]
def x(dac, n=None): s = sfs(dac, n=n) s = scale_sfs(s) return s
sfs_scaled
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(dac, n=None):\n s = sfs(dac, n=n)\n s = scale_sfs(s)\n return s\n", "role": "user" }, { "content": "sfs_scaled", "role": "assistant" } ]
def x(s): k = np.arange(s.size) out = s * k return out
scale_sfs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(s):\n k = np.arange(s.size)\n out = s * k\n return out\n", "role": "user" }, { "content": "scale_sfs", "role": "assistant" } ]
def x(ac, n=None): ac, n = _check_ac_n(ac, n) s = sfs_folded(ac, n=n) s = scale_sfs_folded(s, n) return s
sfs_folded_scaled
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(ac, n=None):\n ac, n = _check_ac_n(ac, n)\n s = sfs_folded(ac, n=n)\n s = scale_sfs_folded(s, n)\n return s\n", "role": "user" }, { "content": "sfs_folded_scaled", "role": "assistant" } ]
def x(s, n): k = np.arange(s.shape[0]) out = s * k * (n - k) / n return out
scale_sfs_folded
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(s, n):\n k = np.arange(s.shape[0])\n out = s * k * (n - k) / n\n return out\n", "role": "user" }, { "content": "scale_sfs_folded", "role": "assistant" } ]
def x(dac1, dac2, n1=None, n2=None): dac1, n1 = _check_dac_n(dac1, n1) dac2, n2 = _check_dac_n(dac2, n2) x = n1 + 1 y = n2 + 1 tmp = (dac1 * y + dac2).astype(int, copy=False) s = np.bincount(tmp) s.resize(x, y) return s
joint_sfs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(dac1, dac2, n1=None, n2=None):\n dac1, n1 = _check_dac_n(dac1, n1)\n dac2, n2 = _check_dac_n(dac2, n2)\n x = n1 + 1\n y = n2 + 1\n tmp = (dac1 * y + dac2).astype(int, copy=False)\n s = np.bincount(tmp)\n s.resize(x, y)\n return s\n", "role": "user" }, { "content": "joint_sfs", "role": "assistant" } ]
def x(ac1, ac2, n1=None, n2=None): ac1, n1 = _check_ac_n(ac1, n1) ac2, n2 = _check_ac_n(ac2, n2) mac1 = np.amin(ac1, axis=1) mac2 = np.amin(ac2, axis=1) x = n1 // 2 + 1 y = n2 // 2 + 1 tmp = (mac1 * y + mac2).astype(int, copy=False) s = np.bincount(tmp) s.resize(x, y) return s
joint_sfs_folded
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(ac1, ac2, n1=None, n2=None):\n ac1, n1 = _check_ac_n(ac1, n1)\n ac2, n2 = _check_ac_n(ac2, n2)\n mac1 = np.amin(ac1, axis=1)\n mac2 = np.amin(ac2, axis=1)\n x = n1 // 2 + 1\n y = n2 // 2 + 1\n tmp = (mac1 * y + mac2).astype(int, copy=False)\n s = np.bincount(tmp)\n s.resize(x, y)\n return s\n", "role": "user" }, { "content": "joint_sfs_folded", "role": "assistant" } ]
def x(dac1, dac2, n1=None, n2=None): s = joint_sfs(dac1, dac2, n1=n1, n2=n2) s = scale_joint_sfs(s) return s
joint_sfs_scaled
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(dac1, dac2, n1=None, n2=None):\n s = joint_sfs(dac1, dac2, n1=n1, n2=n2)\n s = scale_joint_sfs(s)\n return s\n", "role": "user" }, { "content": "joint_sfs_scaled", "role": "assistant" } ]
def x(s): i = np.arange(s.shape[0])[:, None] j = np.arange(s.shape[1])[None, :] out = s * i * j return out
scale_joint_sfs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(s):\n i = np.arange(s.shape[0])[:, None]\n j = np.arange(s.shape[1])[None, :]\n out = s * i * j\n return out\n", "role": "user" }, { "content": "scale_joint_sfs", "role": "assistant" } ]
def x(ac1, ac2, n1=None, n2=None): ac1, n1 = _check_ac_n(ac1, n1) ac2, n2 = _check_ac_n(ac2, n2) s = joint_sfs_folded(ac1, ac2, n1=n1, n2=n2) s = scale_joint_sfs_folded(s, n1, n2) return s
joint_sfs_folded_scaled
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(ac1, ac2, n1=None, n2=None):\n ac1, n1 = _check_ac_n(ac1, n1)\n ac2, n2 = _check_ac_n(ac2, n2)\n s = joint_sfs_folded(ac1, ac2, n1=n1, n2=n2)\n s = scale_joint_sfs_folded(s, n1, n2)\n return s\n", "role": "user" }, { "content": "joint_sfs_folded_scaled", "role": "assistant" } ]
def x(s, n1, n2): out = np.empty_like(s) for i in range(s.shape[0]): for j in range(s.shape[1]): out[i, j] = s[i, j] * i * j * (n1 - i) * (n2 - j) return out
scale_joint_sfs_folded
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(s, n1, n2):\n out = np.empty_like(s)\n for i in range(s.shape[0]):\n for j in range(s.shape[1]):\n out[i, j] = s[i, j] * i * j * (n1 - i) * (n2 - j)\n return out\n", "role": "user" }, { "content": "scale_joint_sfs_folded", "role": "assistant" } ]
def x(s, n): s = asarray_ndim(s, 1) assert s.shape[0] <= n + 1, 'invalid number of chromosomes' if s.shape[0] < n + 1: sn = np.zeros(n + 1, dtype=s.dtype) sn[:s.shape[0]] = s s = sn nf = (n + 1) // 2 n = nf * 2 o = s[:nf] + s[nf:n][::-1] return o
fold_sfs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(s, n):\n s = asarray_ndim(s, 1)\n assert s.shape[0] <= n + 1, 'invalid number of chromosomes'\n if s.shape[0] < n + 1:\n sn = np.zeros(n + 1, dtype=s.dtype)\n sn[:s.shape[0]] = s\n s = sn\n nf = (n + 1) // 2\n n = nf * 2\n o = s[:nf] + s[nf:n][::-1]\n return o\n", "role": "user" }, { "content": "fold_sfs", "role": "assistant" } ]
def x(s, n1, n2): s = asarray_ndim(s, 2) assert s.shape[0] <= n1 + 1, 'invalid number of chromosomes' assert s.shape[1] <= n2 + 1, 'invalid number of chromosomes' if s.shape[0] < n1 + 1: sm = np.zeros((n1 + 1, s.shape[1]), dtype=s.dtype) sm[:s.shape[0]] = s s = sm if s.shape[1] < n2 + 1: sn = np.zeros((s.shape[0], n2 + 1), dtype=s.dtype) sn[:, :s.shape[1]] = s s = sn mf = (n1 + 1) // 2 nf = (n2 + 1) // 2 n1 = mf * 2 n2 = nf * 2 o = s[:mf, :nf] + s[mf:n1, :nf][::-1] + s[:mf, nf:n2][:, ::-1] + s[mf: n1, nf:n2][::-1, ::-1] return o
fold_joint_sfs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(s, n1, n2):\n s = asarray_ndim(s, 2)\n assert s.shape[0] <= n1 + 1, 'invalid number of chromosomes'\n assert s.shape[1] <= n2 + 1, 'invalid number of chromosomes'\n if s.shape[0] < n1 + 1:\n sm = np.zeros((n1 + 1, s.shape[1]), dtype=s.dtype)\n sm[:s.shape[0]] = s\n s = sm\n if s.shape[1] < n2 + 1:\n sn = np.zeros((s.shape[0], n2 + 1), dtype=s.dtype)\n sn[:, :s.shape[1]] = s\n s = sn\n mf = (n1 + 1) // 2\n nf = (n2 + 1) // 2\n n1 = mf * 2\n n2 = nf * 2\n o = s[:mf, :nf] + s[mf:n1, :nf][::-1] + s[:mf, nf:n2][:, ::-1] + s[mf:\n n1, nf:n2][::-1, ::-1]\n return o\n", "role": "user" }, { "content": "fold_joint_sfs", "role": "assistant" } ]
def x(s, yscale='log', bins=None, n=None, clip_endpoints=True, label=None, plot_kwargs=None, ax=None): import matplotlib.pyplot as plt import scipy s = asarray_ndim(s, 1) if ax is None: fig, ax = plt.subplots() if bins is None: if clip_endpoints: x = np.arange(1, s.shape[0] - 1) y = s[1:-1] else: x = np.arange(s.shape[0]) y = s else: if clip_endpoints: y, b, _ = scipy.stats.binned_statistic(np.arange(1, s.shape[0] - 1), values=s[1:-1], bins=bins, statistic='sum') else: y, b, _ = scipy.stats.binned_statistic(np.arange(s.shape[0]), values=s, bins=bins, statistic='sum') x = (b[:-1] + b[1:]) / 2 if n: x = x / n ax.set_xlabel('derived allele frequency') else: ax.set_xlabel('derived allele count') if plot_kwargs is None: plot_kwargs = dict() ax.plot(x, y, label=label, **plot_kwargs) ax.set_yscale(yscale) ax.set_ylabel('site frequency') ax.autoscale(axis='x', tight=True) return ax
plot_sfs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(s, yscale='log', bins=None, n=None, clip_endpoints=True, label=None,\n plot_kwargs=None, ax=None):\n import matplotlib.pyplot as plt\n import scipy\n s = asarray_ndim(s, 1)\n if ax is None:\n fig, ax = plt.subplots()\n if bins is None:\n if clip_endpoints:\n x = np.arange(1, s.shape[0] - 1)\n y = s[1:-1]\n else:\n x = np.arange(s.shape[0])\n y = s\n else:\n if clip_endpoints:\n y, b, _ = scipy.stats.binned_statistic(np.arange(1, s.shape[0] -\n 1), values=s[1:-1], bins=bins, statistic='sum')\n else:\n y, b, _ = scipy.stats.binned_statistic(np.arange(s.shape[0]),\n values=s, bins=bins, statistic='sum')\n x = (b[:-1] + b[1:]) / 2\n if n:\n x = x / n\n ax.set_xlabel('derived allele frequency')\n else:\n ax.set_xlabel('derived allele count')\n if plot_kwargs is None:\n plot_kwargs = dict()\n ax.plot(x, y, label=label, **plot_kwargs)\n ax.set_yscale(yscale)\n ax.set_ylabel('site frequency')\n ax.autoscale(axis='x', tight=True)\n return ax\n", "role": "user" }, { "content": "plot_sfs", "role": "assistant" } ]
def x(*args, **kwargs): ax = plot_sfs(*args, **kwargs) n = kwargs.get('n', None) if n: ax.set_xlabel('minor allele frequency') else: ax.set_xlabel('minor allele count') return ax
plot_sfs_folded
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(*args, **kwargs):\n ax = plot_sfs(*args, **kwargs)\n n = kwargs.get('n', None)\n if n:\n ax.set_xlabel('minor allele frequency')\n else:\n ax.set_xlabel('minor allele count')\n return ax\n", "role": "user" }, { "content": "plot_sfs_folded", "role": "assistant" } ]
def x(*args, **kwargs): kwargs.setdefault('yscale', 'linear') ax = plot_sfs(*args, **kwargs) ax.set_ylabel('scaled site frequency') return ax
plot_sfs_scaled
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(*args, **kwargs):\n kwargs.setdefault('yscale', 'linear')\n ax = plot_sfs(*args, **kwargs)\n ax.set_ylabel('scaled site frequency')\n return ax\n", "role": "user" }, { "content": "plot_sfs_scaled", "role": "assistant" } ]
def x(*args, **kwargs): kwargs.setdefault('yscale', 'linear') ax = plot_sfs_folded(*args, **kwargs) ax.set_ylabel('scaled site frequency') n = kwargs.get('n', None) if n: ax.set_xlabel('minor allele frequency') else: ax.set_xlabel('minor allele count') return ax
plot_sfs_folded_scaled
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(*args, **kwargs):\n kwargs.setdefault('yscale', 'linear')\n ax = plot_sfs_folded(*args, **kwargs)\n ax.set_ylabel('scaled site frequency')\n n = kwargs.get('n', None)\n if n:\n ax.set_xlabel('minor allele frequency')\n else:\n ax.set_xlabel('minor allele count')\n return ax\n", "role": "user" }, { "content": "plot_sfs_folded_scaled", "role": "assistant" } ]
def x(s, ax=None, imshow_kwargs=None): import matplotlib.pyplot as plt from matplotlib.colors import LogNorm s = asarray_ndim(s, 2) if ax is None: w = plt.rcParams['figure.figsize'][0] fig, ax = plt.subplots(figsize=(w, w)) if imshow_kwargs is None: imshow_kwargs = dict() imshow_kwargs.setdefault('cmap', 'jet') imshow_kwargs.setdefault('interpolation', 'none') imshow_kwargs.setdefault('aspect', 'auto') imshow_kwargs.setdefault('norm', LogNorm()) ax.imshow(s.T, **imshow_kwargs) ax.invert_yaxis() ax.set_xlabel('derived allele count (population 1)') ax.set_ylabel('derived allele count (population 2)') return ax
plot_joint_sfs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(s, ax=None, imshow_kwargs=None):\n import matplotlib.pyplot as plt\n from matplotlib.colors import LogNorm\n s = asarray_ndim(s, 2)\n if ax is None:\n w = plt.rcParams['figure.figsize'][0]\n fig, ax = plt.subplots(figsize=(w, w))\n if imshow_kwargs is None:\n imshow_kwargs = dict()\n imshow_kwargs.setdefault('cmap', 'jet')\n imshow_kwargs.setdefault('interpolation', 'none')\n imshow_kwargs.setdefault('aspect', 'auto')\n imshow_kwargs.setdefault('norm', LogNorm())\n ax.imshow(s.T, **imshow_kwargs)\n ax.invert_yaxis()\n ax.set_xlabel('derived allele count (population 1)')\n ax.set_ylabel('derived allele count (population 2)')\n return ax\n", "role": "user" }, { "content": "plot_joint_sfs", "role": "assistant" } ]
def x(*args, **kwargs): ax = plot_joint_sfs(*args, **kwargs) ax.set_xlabel('minor allele count (population 1)') ax.set_ylabel('minor allele count (population 2)') return ax
plot_joint_sfs_folded
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(*args, **kwargs):\n ax = plot_joint_sfs(*args, **kwargs)\n ax.set_xlabel('minor allele count (population 1)')\n ax.set_ylabel('minor allele count (population 2)')\n return ax\n", "role": "user" }, { "content": "plot_joint_sfs_folded", "role": "assistant" } ]
def x(*args, **kwargs): imshow_kwargs = kwargs.get('imshow_kwargs', dict()) imshow_kwargs.setdefault('norm', None) kwargs['imshow_kwargs'] = imshow_kwargs ax = plot_joint_sfs(*args, **kwargs) return ax
plot_joint_sfs_scaled
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(*args, **kwargs):\n imshow_kwargs = kwargs.get('imshow_kwargs', dict())\n imshow_kwargs.setdefault('norm', None)\n kwargs['imshow_kwargs'] = imshow_kwargs\n ax = plot_joint_sfs(*args, **kwargs)\n return ax\n", "role": "user" }, { "content": "plot_joint_sfs_scaled", "role": "assistant" } ]
def x(*args, **kwargs): imshow_kwargs = kwargs.get('imshow_kwargs', dict()) imshow_kwargs.setdefault('norm', None) kwargs['imshow_kwargs'] = imshow_kwargs ax = plot_joint_sfs_folded(*args, **kwargs) ax.set_xlabel('minor allele count (population 1)') ax.set_ylabel('minor allele count (population 2)') return ax
plot_joint_sfs_folded_scaled
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(*args, **kwargs):\n imshow_kwargs = kwargs.get('imshow_kwargs', dict())\n imshow_kwargs.setdefault('norm', None)\n kwargs['imshow_kwargs'] = imshow_kwargs\n ax = plot_joint_sfs_folded(*args, **kwargs)\n ax.set_xlabel('minor allele count (population 1)')\n ax.set_ylabel('minor allele count (population 2)')\n return ax\n", "role": "user" }, { "content": "plot_joint_sfs_folded_scaled", "role": "assistant" } ]
def x(x): if not x.flags.writeable: if not x.flags.owndata: x = x.copy(order='A') x.setflags(write=True) return x
memoryview_safe
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(x):\n if not x.flags.writeable:\n if not x.flags.owndata:\n x = x.copy(order='A')\n x.setflags(write=True)\n return x\n", "role": "user" }, { "content": "memoryview_safe", "role": "assistant" } ]
def x(fields): store_samples = False if fields is None: return True, None if isinstance(fields, str): fields = [fields] else: fields = list(fields) if 'samples' in fields: fields.remove('samples') store_samples = True elif '*' in fields: store_samples = True return store_samples, fields
prep_fields_param
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(fields):\n store_samples = False\n if fields is None:\n return True, None\n if isinstance(fields, str):\n fields = [fields]\n else:\n fields = list(fields)\n if 'samples' in fields:\n fields.remove('samples')\n store_samples = True\n elif '*' in fields:\n store_samples = True\n return store_samples, fields\n", "role": "user" }, { "content": "prep_fields_param", "role": "assistant" } ]
def x(it, log, prefix): n_variants = 0 before_all = time.time() before_chunk = before_all for chunk, chunk_length, chrom, pos in it: after_chunk = time.time() elapsed_chunk = after_chunk - before_chunk elapsed = after_chunk - before_all n_variants += chunk_length chrom = text_type(chrom, 'utf8') message = '%s %s rows in %.2fs; chunk in %.2fs (%s rows/s)' % (prefix, n_variants, elapsed, elapsed_chunk, int(chunk_length // elapsed_chunk)) if chrom: message += '; %s:%s' % (chrom, pos) print(message, file=log) log.flush() yield chunk, chunk_length, chrom, pos before_chunk = after_chunk after_all = time.time() elapsed = after_all - before_all print('%s all done (%s rows/s)' % (prefix, int(n_variants // elapsed)), file=log) log.flush()
chunk_iter_progress
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(it, log, prefix):\n n_variants = 0\n before_all = time.time()\n before_chunk = before_all\n for chunk, chunk_length, chrom, pos in it:\n after_chunk = time.time()\n elapsed_chunk = after_chunk - before_chunk\n elapsed = after_chunk - before_all\n n_variants += chunk_length\n chrom = text_type(chrom, 'utf8')\n message = '%s %s rows in %.2fs; chunk in %.2fs (%s rows/s)' % (prefix,\n n_variants, elapsed, elapsed_chunk, int(chunk_length //\n elapsed_chunk))\n if chrom:\n message += '; %s:%s' % (chrom, pos)\n print(message, file=log)\n log.flush()\n yield chunk, chunk_length, chrom, pos\n before_chunk = after_chunk\n after_all = time.time()\n elapsed = after_all - before_all\n print('%s all done (%s rows/s)' % (prefix, int(n_variants // elapsed)),\n file=log)\n log.flush()\n", "role": "user" }, { "content": "chunk_iter_progress", "role": "assistant" } ]
def x(input, fields=None, exclude_fields=None, rename_fields=None, types= None, numbers=None, alt_number=DEFAULT_ALT_NUMBER, fills=None, region= None, tabix='tabix', samples=None, transformers=None, buffer_size= DEFAULT_BUFFER_SIZE, chunk_length=DEFAULT_CHUNK_LENGTH, log=None): store_samples, fields = _prep_fields_param(fields) fields, samples, headers, it = iter_vcf_chunks(input=input, fields= fields, exclude_fields=exclude_fields, types=types, numbers=numbers, alt_number=alt_number, buffer_size=buffer_size, chunk_length= chunk_length, fills=fills, region=region, tabix=tabix, samples= samples, transformers=transformers) if rename_fields: rename_fields, it = _do_rename(it, fields=fields, rename_fields= rename_fields, headers=headers) if log is not None: it = _chunk_iter_progress(it, log, prefix='[read_vcf]') chunks = [d[0] for d in it] if chunks: output = dict() if len(samples) > 0 and store_samples: output['samples'] = samples keys = sorted(chunks[0].keys()) for k in keys: output[k] = np.concatenate([chunk[k] for chunk in chunks], axis=0) else: output = None return output
read_vcf
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(input, fields=None, exclude_fields=None, rename_fields=None, types=\n None, numbers=None, alt_number=DEFAULT_ALT_NUMBER, fills=None, region=\n None, tabix='tabix', samples=None, transformers=None, buffer_size=\n DEFAULT_BUFFER_SIZE, chunk_length=DEFAULT_CHUNK_LENGTH, log=None):\n store_samples, fields = _prep_fields_param(fields)\n fields, samples, headers, it = iter_vcf_chunks(input=input, fields=\n fields, exclude_fields=exclude_fields, types=types, numbers=numbers,\n alt_number=alt_number, buffer_size=buffer_size, chunk_length=\n chunk_length, fills=fills, region=region, tabix=tabix, samples=\n samples, transformers=transformers)\n if rename_fields:\n rename_fields, it = _do_rename(it, fields=fields, rename_fields=\n rename_fields, headers=headers)\n if log is not None:\n it = _chunk_iter_progress(it, log, prefix='[read_vcf]')\n chunks = [d[0] for d in it]\n if chunks:\n output = dict()\n if len(samples) > 0 and store_samples:\n output['samples'] = samples\n keys = sorted(chunks[0].keys())\n for k in keys:\n output[k] = np.concatenate([chunk[k] for chunk in chunks], axis=0)\n else:\n output = None\n return output\n", "role": "user" }, { "content": "read_vcf", "role": "assistant" } ]
def x(input, output, compressed=True, overwrite=False, fields=None, exclude_fields=None, rename_fields=None, types=None, numbers=None, alt_number=DEFAULT_ALT_NUMBER, fills=None, region=None, tabix=True, samples=None, transformers=None, buffer_size=DEFAULT_BUFFER_SIZE, chunk_length=DEFAULT_CHUNK_LENGTH, log=None): if not overwrite and os.path.exists(output): raise ValueError( 'file exists at path %r; use overwrite=True to replace' % output) data = read_vcf(input=input, fields=fields, exclude_fields= exclude_fields, rename_fields=rename_fields, types=types, numbers= numbers, alt_number=alt_number, buffer_size=buffer_size, chunk_length=chunk_length, log=log, fills=fills, region=region, tabix=tabix, samples=samples, transformers=transformers) if data is None: return if compressed: savez = np.savez_compressed else: savez = np.savez savez(output, **data)
vcf_to_npz
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(input, output, compressed=True, overwrite=False, fields=None,\n exclude_fields=None, rename_fields=None, types=None, numbers=None,\n alt_number=DEFAULT_ALT_NUMBER, fills=None, region=None, tabix=True,\n samples=None, transformers=None, buffer_size=DEFAULT_BUFFER_SIZE,\n chunk_length=DEFAULT_CHUNK_LENGTH, log=None):\n if not overwrite and os.path.exists(output):\n raise ValueError(\n 'file exists at path %r; use overwrite=True to replace' % output)\n data = read_vcf(input=input, fields=fields, exclude_fields=\n exclude_fields, rename_fields=rename_fields, types=types, numbers=\n numbers, alt_number=alt_number, buffer_size=buffer_size,\n chunk_length=chunk_length, log=log, fills=fills, region=region,\n tabix=tabix, samples=samples, transformers=transformers)\n if data is None:\n return\n if compressed:\n savez = np.savez_compressed\n else:\n savez = np.savez\n savez(output, **data)\n", "role": "user" }, { "content": "vcf_to_npz", "role": "assistant" } ]
def x(input, output, group='/', compression='gzip', compression_opts=1, shuffle=False, overwrite=False, vlen=True, fields=None, exclude_fields= None, rename_fields=None, types=None, numbers=None, alt_number= DEFAULT_ALT_NUMBER, fills=None, region=None, tabix='tabix', samples= None, transformers=None, buffer_size=DEFAULT_BUFFER_SIZE, chunk_length= DEFAULT_CHUNK_LENGTH, chunk_width=DEFAULT_CHUNK_WIDTH, log=None): import h5py store_samples, fields = _prep_fields_param(fields) fields, samples, headers, it = iter_vcf_chunks(input, fields=fields, exclude_fields=exclude_fields, types=types, numbers=numbers, alt_number=alt_number, buffer_size=buffer_size, chunk_length= chunk_length, fills=fills, region=region, tabix=tabix, samples= samples, transformers=transformers) if rename_fields: rename_fields, it = _do_rename(it, fields=fields, rename_fields= rename_fields, headers=headers) if log is not None: it = _chunk_iter_progress(it, log, prefix='[vcf_to_hdf5]') try: chunk, _, _, _ = next(it) except StopIteration: return with h5py.File(output, mode='a') as h5f: root = h5f.require_group(group) if len(samples) > 0 and store_samples: name = 'samples' if name in root: if overwrite: del root[name] else: raise ValueError( 'dataset exists at path %r; use overwrite=True to replace' % name) if samples.dtype.kind == 'O': if vlen: t = h5py.special_dtype(vlen=str) else: samples = samples.astype('S') t = samples.dtype else: t = samples.dtype root.create_dataset(name, data=samples, chunks=None, dtype=t) keys = _hdf5_setup_datasets(chunk=chunk, root=root, chunk_length= chunk_length, chunk_width=chunk_width, compression=compression, compression_opts=compression_opts, shuffle=shuffle, overwrite= overwrite, headers=headers, vlen=vlen) _hdf5_store_chunk(root, keys, chunk, vlen) for chunk, _, _, _ in it: _hdf5_store_chunk(root, keys, chunk, vlen)
vcf_to_hdf5
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(input, output, group='/', compression='gzip', compression_opts=1,\n shuffle=False, overwrite=False, vlen=True, fields=None, exclude_fields=\n None, rename_fields=None, types=None, numbers=None, alt_number=\n DEFAULT_ALT_NUMBER, fills=None, region=None, tabix='tabix', samples=\n None, transformers=None, buffer_size=DEFAULT_BUFFER_SIZE, chunk_length=\n DEFAULT_CHUNK_LENGTH, chunk_width=DEFAULT_CHUNK_WIDTH, log=None):\n import h5py\n store_samples, fields = _prep_fields_param(fields)\n fields, samples, headers, it = iter_vcf_chunks(input, fields=fields,\n exclude_fields=exclude_fields, types=types, numbers=numbers,\n alt_number=alt_number, buffer_size=buffer_size, chunk_length=\n chunk_length, fills=fills, region=region, tabix=tabix, samples=\n samples, transformers=transformers)\n if rename_fields:\n rename_fields, it = _do_rename(it, fields=fields, rename_fields=\n rename_fields, headers=headers)\n if log is not None:\n it = _chunk_iter_progress(it, log, prefix='[vcf_to_hdf5]')\n try:\n chunk, _, _, _ = next(it)\n except StopIteration:\n return\n with h5py.File(output, mode='a') as h5f:\n root = h5f.require_group(group)\n if len(samples) > 0 and store_samples:\n name = 'samples'\n if name in root:\n if overwrite:\n del root[name]\n else:\n raise ValueError(\n 'dataset exists at path %r; use overwrite=True to replace'\n % name)\n if samples.dtype.kind == 'O':\n if vlen:\n t = h5py.special_dtype(vlen=str)\n else:\n samples = samples.astype('S')\n t = samples.dtype\n else:\n t = samples.dtype\n root.create_dataset(name, data=samples, chunks=None, dtype=t)\n keys = _hdf5_setup_datasets(chunk=chunk, root=root, chunk_length=\n chunk_length, chunk_width=chunk_width, compression=compression,\n compression_opts=compression_opts, shuffle=shuffle, overwrite=\n overwrite, headers=headers, vlen=vlen)\n _hdf5_store_chunk(root, keys, chunk, vlen)\n for chunk, _, _, _ in it:\n _hdf5_store_chunk(root, keys, chunk, vlen)\n", "role": "user" }, { "content": "vcf_to_hdf5", "role": "assistant" } ]
def x(input, output, group='/', compressor='default', overwrite=False, fields=None, exclude_fields=None, rename_fields=None, types=None, numbers=None, alt_number=DEFAULT_ALT_NUMBER, fills=None, region=None, tabix='tabix', samples=None, transformers=None, buffer_size= DEFAULT_BUFFER_SIZE, chunk_length=DEFAULT_CHUNK_LENGTH, chunk_width= DEFAULT_CHUNK_WIDTH, log=None): import zarr store_samples, fields = _prep_fields_param(fields) fields, samples, headers, it = iter_vcf_chunks(input, fields=fields, exclude_fields=exclude_fields, types=types, numbers=numbers, alt_number=alt_number, buffer_size=buffer_size, chunk_length= chunk_length, fills=fills, region=region, tabix=tabix, samples= samples, transformers=transformers) if rename_fields: rename_fields, it = _do_rename(it, fields=fields, rename_fields= rename_fields, headers=headers) ci_field_index = defaultdict(list) for f in fields: if rename_fields: f = rename_fields.get(f, f) ci_field_index[f.lower()].append(f) for k, v in ci_field_index.items(): if len(v) > 1: msg = textwrap.fill( 'Found two or more fields with the same name when compared case-insensitive: {!r}; this is not supported because it causes problems on platforms with a case-insensitive file system, which is usually the default on Windows and Mac OS. Please rename fields so they are distinct under a case-insensitive comparison via the rename_fields argument.' .format(v), width=80) raise ValueError(msg) if log is not None: it = _chunk_iter_progress(it, log, prefix='[vcf_to_zarr]') try: chunk, _, _, _ = next(it) except StopIteration: return root = zarr.open_group(output, mode='a', path=group) if len(samples) > 0 and store_samples: if samples.dtype.kind == 'O': if PY2: dtype = 'unicode' else: dtype = 'str' else: dtype = samples.dtype root.create_dataset('samples', data=samples, compressor=None, overwrite=overwrite, dtype=dtype) keys = _zarr_setup_datasets(chunk, root=root, chunk_length=chunk_length, chunk_width=chunk_width, compressor=compressor, overwrite=overwrite, headers=headers) _zarr_store_chunk(root, keys, chunk) for chunk, _, _, _ in it: _zarr_store_chunk(root, keys, chunk)
vcf_to_zarr
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(input, output, group='/', compressor='default', overwrite=False,\n fields=None, exclude_fields=None, rename_fields=None, types=None,\n numbers=None, alt_number=DEFAULT_ALT_NUMBER, fills=None, region=None,\n tabix='tabix', samples=None, transformers=None, buffer_size=\n DEFAULT_BUFFER_SIZE, chunk_length=DEFAULT_CHUNK_LENGTH, chunk_width=\n DEFAULT_CHUNK_WIDTH, log=None):\n import zarr\n store_samples, fields = _prep_fields_param(fields)\n fields, samples, headers, it = iter_vcf_chunks(input, fields=fields,\n exclude_fields=exclude_fields, types=types, numbers=numbers,\n alt_number=alt_number, buffer_size=buffer_size, chunk_length=\n chunk_length, fills=fills, region=region, tabix=tabix, samples=\n samples, transformers=transformers)\n if rename_fields:\n rename_fields, it = _do_rename(it, fields=fields, rename_fields=\n rename_fields, headers=headers)\n ci_field_index = defaultdict(list)\n for f in fields:\n if rename_fields:\n f = rename_fields.get(f, f)\n ci_field_index[f.lower()].append(f)\n for k, v in ci_field_index.items():\n if len(v) > 1:\n msg = textwrap.fill(\n 'Found two or more fields with the same name when compared case-insensitive: {!r}; this is not supported because it causes problems on platforms with a case-insensitive file system, which is usually the default on Windows and Mac OS. Please rename fields so they are distinct under a case-insensitive comparison via the rename_fields argument.'\n .format(v), width=80)\n raise ValueError(msg)\n if log is not None:\n it = _chunk_iter_progress(it, log, prefix='[vcf_to_zarr]')\n try:\n chunk, _, _, _ = next(it)\n except StopIteration:\n return\n root = zarr.open_group(output, mode='a', path=group)\n if len(samples) > 0 and store_samples:\n if samples.dtype.kind == 'O':\n if PY2:\n dtype = 'unicode'\n else:\n dtype = 'str'\n else:\n dtype = samples.dtype\n root.create_dataset('samples', data=samples, compressor=None,\n overwrite=overwrite, dtype=dtype)\n keys = _zarr_setup_datasets(chunk, root=root, chunk_length=chunk_length,\n chunk_width=chunk_width, compressor=compressor, overwrite=overwrite,\n headers=headers)\n _zarr_store_chunk(root, keys, chunk)\n for chunk, _, _, _ in it:\n _zarr_store_chunk(root, keys, chunk)\n", "role": "user" }, { "content": "vcf_to_zarr", "role": "assistant" } ]
def x(input, fields=None, exclude_fields=None, types=None, numbers=None, alt_number=DEFAULT_ALT_NUMBER, fills=None, region=None, tabix='tabix', samples=None, transformers=None, buffer_size=DEFAULT_BUFFER_SIZE, chunk_length=DEFAULT_CHUNK_LENGTH): kwds = dict(fields=fields, exclude_fields=exclude_fields, types=types, numbers=numbers, alt_number=alt_number, chunk_length=chunk_length, fills=fills, samples=samples, region=region) stream = _setup_input_stream(input=input, region=region, tabix=tabix, buffer_size=buffer_size) fields, samples, headers, it = _iter_vcf_stream(stream, **kwds) if transformers is not None: if not isinstance(transformers, (list, tuple)): transformers = [transformers] for trans in transformers: fields = trans.transform_fields(fields) it = _chunk_iter_transform(it, transformers) return fields, samples, headers, it
iter_vcf_chunks
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(input, fields=None, exclude_fields=None, types=None, numbers=None,\n alt_number=DEFAULT_ALT_NUMBER, fills=None, region=None, tabix='tabix',\n samples=None, transformers=None, buffer_size=DEFAULT_BUFFER_SIZE,\n chunk_length=DEFAULT_CHUNK_LENGTH):\n kwds = dict(fields=fields, exclude_fields=exclude_fields, types=types,\n numbers=numbers, alt_number=alt_number, chunk_length=chunk_length,\n fills=fills, samples=samples, region=region)\n stream = _setup_input_stream(input=input, region=region, tabix=tabix,\n buffer_size=buffer_size)\n fields, samples, headers, it = _iter_vcf_stream(stream, **kwds)\n if transformers is not None:\n if not isinstance(transformers, (list, tuple)):\n transformers = [transformers]\n for trans in transformers:\n fields = trans.transform_fields(fields)\n it = _chunk_iter_transform(it, transformers)\n return fields, samples, headers, it\n", "role": "user" }, { "content": "iter_vcf_chunks", "role": "assistant" } ]
def x(input, fields=None, exclude_fields=None, types=None, numbers=None, alt_number=DEFAULT_ALT_NUMBER, fills=None, region=None, tabix='tabix', transformers=None, buffer_size=DEFAULT_BUFFER_SIZE, chunk_length= DEFAULT_CHUNK_LENGTH, log=None): import pandas _, fields = _prep_fields_param(fields) fields, _, _, it = iter_vcf_chunks(input=input, fields=fields, exclude_fields=exclude_fields, types=types, numbers=numbers, alt_number=alt_number, buffer_size=buffer_size, chunk_length= chunk_length, fills=fills, region=region, tabix=tabix, samples=[], transformers=transformers) if log is not None: it = _chunk_iter_progress(it, log, prefix='[vcf_to_dataframe]') chunks = [d[0] for d in it] output = None if chunks: output = pandas.concat([_chunk_to_dataframe(fields, chunk) for chunk in chunks]) return output
vcf_to_dataframe
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(input, fields=None, exclude_fields=None, types=None, numbers=None,\n alt_number=DEFAULT_ALT_NUMBER, fills=None, region=None, tabix='tabix',\n transformers=None, buffer_size=DEFAULT_BUFFER_SIZE, chunk_length=\n DEFAULT_CHUNK_LENGTH, log=None):\n import pandas\n _, fields = _prep_fields_param(fields)\n fields, _, _, it = iter_vcf_chunks(input=input, fields=fields,\n exclude_fields=exclude_fields, types=types, numbers=numbers,\n alt_number=alt_number, buffer_size=buffer_size, chunk_length=\n chunk_length, fills=fills, region=region, tabix=tabix, samples=[],\n transformers=transformers)\n if log is not None:\n it = _chunk_iter_progress(it, log, prefix='[vcf_to_dataframe]')\n chunks = [d[0] for d in it]\n output = None\n if chunks:\n output = pandas.concat([_chunk_to_dataframe(fields, chunk) for\n chunk in chunks])\n return output\n", "role": "user" }, { "content": "vcf_to_dataframe", "role": "assistant" } ]
def x(input, output, fields=None, exclude_fields=None, types=None, numbers= None, alt_number=DEFAULT_ALT_NUMBER, fills=None, region=None, tabix= 'tabix', transformers=None, buffer_size=DEFAULT_BUFFER_SIZE, chunk_length=DEFAULT_CHUNK_LENGTH, log=None, **kwargs): _, fields = _prep_fields_param(fields) fields, _, _, it = iter_vcf_chunks(input=input, fields=fields, exclude_fields=exclude_fields, types=types, numbers=numbers, alt_number=alt_number, buffer_size=buffer_size, chunk_length= chunk_length, fills=fills, region=region, tabix=tabix, samples=[], transformers=transformers) if log is not None: it = _chunk_iter_progress(it, log, prefix='[vcf_to_csv]') kwargs['index'] = False for i, (chunk, _, _, _) in enumerate(it): df = _chunk_to_dataframe(fields, chunk) if i == 0: kwargs['header'] = True kwargs['mode'] = 'w' else: kwargs['header'] = False kwargs['mode'] = 'a' df.to_csv(output, **kwargs)
vcf_to_csv
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(input, output, fields=None, exclude_fields=None, types=None, numbers=\n None, alt_number=DEFAULT_ALT_NUMBER, fills=None, region=None, tabix=\n 'tabix', transformers=None, buffer_size=DEFAULT_BUFFER_SIZE,\n chunk_length=DEFAULT_CHUNK_LENGTH, log=None, **kwargs):\n _, fields = _prep_fields_param(fields)\n fields, _, _, it = iter_vcf_chunks(input=input, fields=fields,\n exclude_fields=exclude_fields, types=types, numbers=numbers,\n alt_number=alt_number, buffer_size=buffer_size, chunk_length=\n chunk_length, fills=fills, region=region, tabix=tabix, samples=[],\n transformers=transformers)\n if log is not None:\n it = _chunk_iter_progress(it, log, prefix='[vcf_to_csv]')\n kwargs['index'] = False\n for i, (chunk, _, _, _) in enumerate(it):\n df = _chunk_to_dataframe(fields, chunk)\n if i == 0:\n kwargs['header'] = True\n kwargs['mode'] = 'w'\n else:\n kwargs['header'] = False\n kwargs['mode'] = 'a'\n df.to_csv(output, **kwargs)\n", "role": "user" }, { "content": "vcf_to_csv", "role": "assistant" } ]
def x(input, fields=None, exclude_fields=None, types=None, numbers=None, alt_number=DEFAULT_ALT_NUMBER, fills=None, region=None, tabix='tabix', transformers=None, buffer_size=DEFAULT_BUFFER_SIZE, chunk_length= DEFAULT_CHUNK_LENGTH, log=None): _, fields = _prep_fields_param(fields) fields, _, _, it = iter_vcf_chunks(input=input, fields=fields, exclude_fields=exclude_fields, types=types, numbers=numbers, alt_number=alt_number, buffer_size=buffer_size, chunk_length= chunk_length, fills=fills, region=region, tabix=tabix, samples=[], transformers=transformers) if log is not None: it = _chunk_iter_progress(it, log, prefix='[vcf_to_recarray]') chunks = [d[0] for d in it] output = None if chunks: output = np.concatenate([_chunk_to_recarray(fields, chunk) for chunk in chunks]) return output
vcf_to_recarray
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(input, fields=None, exclude_fields=None, types=None, numbers=None,\n alt_number=DEFAULT_ALT_NUMBER, fills=None, region=None, tabix='tabix',\n transformers=None, buffer_size=DEFAULT_BUFFER_SIZE, chunk_length=\n DEFAULT_CHUNK_LENGTH, log=None):\n _, fields = _prep_fields_param(fields)\n fields, _, _, it = iter_vcf_chunks(input=input, fields=fields,\n exclude_fields=exclude_fields, types=types, numbers=numbers,\n alt_number=alt_number, buffer_size=buffer_size, chunk_length=\n chunk_length, fills=fills, region=region, tabix=tabix, samples=[],\n transformers=transformers)\n if log is not None:\n it = _chunk_iter_progress(it, log, prefix='[vcf_to_recarray]')\n chunks = [d[0] for d in it]\n output = None\n if chunks:\n output = np.concatenate([_chunk_to_recarray(fields, chunk) for\n chunk in chunks])\n return output\n", "role": "user" }, { "content": "vcf_to_recarray", "role": "assistant" } ]