source_code
stringlengths
45
7.24k
function_name
stringlengths
1
111
messages
list
def x(self, line): stripped = line.strip() if stripped and stripped.startswith('#') is False: rexres = self.rex.search(stripped) if rexres: return ' '.join(rexres.group(1).split()) return None
parse_cron_line
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, line):\n stripped = line.strip()\n if stripped and stripped.startswith('#') is False:\n rexres = self.rex.search(stripped)\n if rexres:\n return ' '.join(rexres.group(1).split())\n return None\n", "role": "user" }, { "content": "parse_cron_line", "role": "assistant" } ]
def x(self, new): if self.node_id != new.node_id: raise ValueError("Broker metadata {!r} doesn't match node_id={}". format(new, self.node_id)) self.node_id = new.node_id self.host = new.host self.port = new.port
updateMetadata
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, new):\n if self.node_id != new.node_id:\n raise ValueError(\"Broker metadata {!r} doesn't match node_id={}\".\n format(new, self.node_id))\n self.node_id = new.node_id\n self.host = new.host\n self.port = new.port\n", "role": "user" }, { "content": "updateMetadata", "role": "assistant" } ]
def x(self, requestId, request, expectResponse=True): if requestId in self.requests: raise DuplicateRequestError('Reuse of requestId:{}'.format(requestId)) if self._dDown: return fail(ClientError('makeRequest() called after close()')) canceller = partial(self.cancelRequest, requestId, CancelledError( message='Request correlationId={} was cancelled'.format(requestId))) tReq = _Request(requestId, request, expectResponse, canceller) self.requests[requestId] = tReq tReq.d.addErrback(self._handleRequestFailure, requestId) if self.proto: self._sendRequest(tReq) elif not self.connector: self._connect() return tReq.d
makeRequest
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, requestId, request, expectResponse=True):\n if requestId in self.requests:\n raise DuplicateRequestError('Reuse of requestId:{}'.format(requestId))\n if self._dDown:\n return fail(ClientError('makeRequest() called after close()'))\n canceller = partial(self.cancelRequest, requestId, CancelledError(\n message='Request correlationId={} was cancelled'.format(requestId)))\n tReq = _Request(requestId, request, expectResponse, canceller)\n self.requests[requestId] = tReq\n tReq.d.addErrback(self._handleRequestFailure, requestId)\n if self.proto:\n self._sendRequest(tReq)\n elif not self.connector:\n self._connect()\n return tReq.d\n", "role": "user" }, { "content": "makeRequest", "role": "assistant" } ]
def x(self): if self.proto: log.debug('%r Disconnecting from %r', self, self.proto.transport. getPeer()) self.proto.transport.loseConnection()
disconnect
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.proto:\n log.debug('%r Disconnecting from %r', self, self.proto.transport.\n getPeer())\n self.proto.transport.loseConnection()\n", "role": "user" }, { "content": "disconnect", "role": "assistant" } ]
def x(self): log.debug('%r: close() proto=%r connector=%r', self, self.proto, self. connector) assert self._dDown is None self._dDown = Deferred() if self.proto is not None: self.proto.transport.loseConnection() elif self.connector is not None: def x(reason): log.debug('%r: connection attempt has been cancelled: %r', self, reason) self._dDown.callback(None) self.connector.addErrback(connectingFailed) self.connector.cancel() else: self._dDown.callback(None) try: raise CancelledError(message= 'Broker client for node_id={} {}:{} was closed'.format(self. node_id, self.host, self.port)) except Exception: reason = Failure() for correlation_id in list(self.requests.keys()): self.cancelRequest(correlation_id, reason) return self._dDown
close
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n log.debug('%r: close() proto=%r connector=%r', self, self.proto, self.\n connector)\n assert self._dDown is None\n self._dDown = Deferred()\n if self.proto is not None:\n self.proto.transport.loseConnection()\n elif self.connector is not None:\n\n def x(reason):\n log.debug('%r: connection attempt has been cancelled: %r', self,\n reason)\n self._dDown.callback(None)\n self.connector.addErrback(connectingFailed)\n self.connector.cancel()\n else:\n self._dDown.callback(None)\n try:\n raise CancelledError(message=\n 'Broker client for node_id={} {}:{} was closed'.format(self.\n node_id, self.host, self.port))\n except Exception:\n reason = Failure()\n for correlation_id in list(self.requests.keys()):\n self.cancelRequest(correlation_id, reason)\n return self._dDown\n", "role": "user" }, { "content": "close", "role": "assistant" } ]
def x(self, reason): log.info('%r: Connection closed: %r', self, reason) self.proto = None for tReq in self.requests.values(): tReq.sent = False if self._dDown: self._dDown.callback(None) elif self.requests: self._connect()
connectionLost
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, reason):\n log.info('%r: Connection closed: %r', self, reason)\n self.proto = None\n for tReq in self.requests.values():\n tReq.sent = False\n if self._dDown:\n self._dDown.callback(None)\n elif self.requests:\n self._connect()\n", "role": "user" }, { "content": "connectionLost", "role": "assistant" } ]
def x(self, response): requestId = KafkaCodec.get_response_correlation_id(response) tReq = self.requests.pop(requestId, None) if tReq is None: log.warning('Unexpected response with correlationId=%d: %r', requestId, reprlib.repr(response)) else: tReq.d.callback(response)
handleResponse
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, response):\n requestId = KafkaCodec.get_response_correlation_id(response)\n tReq = self.requests.pop(requestId, None)\n if tReq is None:\n log.warning('Unexpected response with correlationId=%d: %r',\n requestId, reprlib.repr(response))\n else:\n tReq.d.callback(response)\n", "role": "user" }, { "content": "handleResponse", "role": "assistant" } ]
def x(self, tReq): try: tReq.sent = True self.proto.sendString(tReq.data) except Exception as e: log.exception('%r: Failed to send request %r', self, tReq) del self.requests[tReq.id] tReq.d.errback(e) else: if not tReq.expect: del self.requests[tReq.id] tReq.d.callback(None)
sendRequest
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, tReq):\n try:\n tReq.sent = True\n self.proto.sendString(tReq.data)\n except Exception as e:\n log.exception('%r: Failed to send request %r', self, tReq)\n del self.requests[tReq.id]\n tReq.d.errback(e)\n else:\n if not tReq.expect:\n del self.requests[tReq.id]\n tReq.d.callback(None)\n", "role": "user" }, { "content": "sendRequest", "role": "assistant" } ]
def x(self): for tReq in list(self.requests.values()): if not tReq.sent: self._sendRequest(tReq)
sendQueued
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n for tReq in list(self.requests.values()):\n if not tReq.sent:\n self._sendRequest(tReq)\n", "role": "user" }, { "content": "sendQueued", "role": "assistant" } ]
def x(self, requestId, reason=None, _=None): if reason is None: reason = CancelledError() tReq = self.requests.pop(requestId) tReq.d.errback(reason)
cancelRequest
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, requestId, reason=None, _=None):\n if reason is None:\n reason = CancelledError()\n tReq = self.requests.pop(requestId)\n tReq.d.errback(reason)\n", "role": "user" }, { "content": "cancelRequest", "role": "assistant" } ]
def x(self): def x(): self.connector = d = maybeDeferred(connect) d.addCallback(cbConnect) d.addErrback(ebConnect) def x(): endpoint = self._endpointFactory(self._reactor, self.host, self.port) log.debug('%r: connecting with %s', self, endpoint) return endpoint.connect(self) def x(proto): log.debug('%r: connected to %r', self, proto.transport.getPeer()) self._failures = 0 self.connector = None self.proto = proto if self._dDown: proto.transport.loseConnection() else: self._sendQueued() def x(fail): if self._dDown: log.debug('%r: breaking connect loop due to %r after close()', self, fail) return fail self._failures += 1 delay = self._retryPolicy(self._failures) log.debug('%r: failure %d to connect -> %s; retry in %.2f seconds.', self, self._failures, fail.value, delay) self.connector = d = deferLater(self._reactor, delay, lambda : None) d.addCallback(cbDelayed) def x(result): tryConnect() self._failures = 0 tryConnect()
connect
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n\n def x():\n self.connector = d = maybeDeferred(connect)\n d.addCallback(cbConnect)\n d.addErrback(ebConnect)\n\n def x():\n endpoint = self._endpointFactory(self._reactor, self.host, self.port)\n log.debug('%r: connecting with %s', self, endpoint)\n return endpoint.connect(self)\n\n def x(proto):\n log.debug('%r: connected to %r', self, proto.transport.getPeer())\n self._failures = 0\n self.connector = None\n self.proto = proto\n if self._dDown:\n proto.transport.loseConnection()\n else:\n self._sendQueued()\n\n def x(fail):\n if self._dDown:\n log.debug('%r: breaking connect loop due to %r after close()',\n self, fail)\n return fail\n self._failures += 1\n delay = self._retryPolicy(self._failures)\n log.debug('%r: failure %d to connect -> %s; retry in %.2f seconds.',\n self, self._failures, fail.value, delay)\n self.connector = d = deferLater(self._reactor, delay, lambda : None)\n d.addCallback(cbDelayed)\n\n def x(result):\n tryConnect()\n self._failures = 0\n tryConnect()\n", "role": "user" }, { "content": "connect", "role": "assistant" } ]
def x(self): parsed = ['', '', '', '', '', '', ''] if self._expression is None or len(self._expression) == 0: raise MissingFieldException('ExpressionDescriptor.expression') else: expression_parts_temp = self._expression.split() expression_parts_temp_length = len(expression_parts_temp) if expression_parts_temp_length < 5: raise FormatException( 'Error: Expression only has {0} parts. At least 5 part are required.' .format(expression_parts_temp_length)) elif expression_parts_temp_length == 5: for i, expression_part_temp in enumerate(expression_parts_temp): parsed[i + 1] = expression_part_temp elif expression_parts_temp_length == 6: year_regex = re.compile('\\d{4}$') if year_regex.search(expression_parts_temp[5]) is not None: for i, expression_part_temp in enumerate(expression_parts_temp ): parsed[i + 1] = expression_part_temp else: for i, expression_part_temp in enumerate(expression_parts_temp ): parsed[i] = expression_part_temp elif expression_parts_temp_length == 7: parsed = expression_parts_temp else: raise FormatException( 'Error: Expression has too many parts ({0}). Expression must not have more than 7 parts.' .format(expression_parts_temp_length)) self.normalize_expression(parsed) return parsed
parse
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n parsed = ['', '', '', '', '', '', '']\n if self._expression is None or len(self._expression) == 0:\n raise MissingFieldException('ExpressionDescriptor.expression')\n else:\n expression_parts_temp = self._expression.split()\n expression_parts_temp_length = len(expression_parts_temp)\n if expression_parts_temp_length < 5:\n raise FormatException(\n 'Error: Expression only has {0} parts. At least 5 part are required.'\n .format(expression_parts_temp_length))\n elif expression_parts_temp_length == 5:\n for i, expression_part_temp in enumerate(expression_parts_temp):\n parsed[i + 1] = expression_part_temp\n elif expression_parts_temp_length == 6:\n year_regex = re.compile('\\\\d{4}$')\n if year_regex.search(expression_parts_temp[5]) is not None:\n for i, expression_part_temp in enumerate(expression_parts_temp\n ):\n parsed[i + 1] = expression_part_temp\n else:\n for i, expression_part_temp in enumerate(expression_parts_temp\n ):\n parsed[i] = expression_part_temp\n elif expression_parts_temp_length == 7:\n parsed = expression_parts_temp\n else:\n raise FormatException(\n 'Error: Expression has too many parts ({0}). Expression must not have more than 7 parts.'\n .format(expression_parts_temp_length))\n self.normalize_expression(parsed)\n return parsed\n", "role": "user" }, { "content": "parse", "role": "assistant" } ]
def x(self, expression_parts): expression_parts[3] = expression_parts[3].replace('?', '*') expression_parts[5] = expression_parts[5].replace('?', '*') if expression_parts[0].startswith('0/'): expression_parts[0] = expression_parts[0].replace('0/', '*/') if expression_parts[1].startswith('0/'): expression_parts[1] = expression_parts[1].replace('0/', '*/') if expression_parts[2].startswith('0/'): expression_parts[2] = expression_parts[2].replace('0/', '*/') if expression_parts[3].startswith('1/'): expression_parts[3] = expression_parts[3].replace('1/', '*/') if expression_parts[4].startswith('1/'): expression_parts[4] = expression_parts[4].replace('1/', '*/') if expression_parts[5].startswith('1/'): expression_parts[5] = expression_parts[5].replace('1/', '*/') if expression_parts[6].startswith('1/'): expression_parts[6] = expression_parts[6].replace('1/', '*/') if self._options.day_of_week_start_index_zero is False: expression_parts[5] = self.decrease_days_of_week(expression_parts[5]) if expression_parts[3] == '?': expression_parts[3] = '*' for day_number in self._cron_days: expression_parts[5] = expression_parts[5].upper().replace(self. _cron_days[day_number], str(day_number)) for month_number in self._cron_months: expression_parts[4] = expression_parts[4].upper().replace(self. _cron_months[month_number], str(month_number)) if expression_parts[0] == '0': expression_parts[0] = '' length = len(expression_parts) for i in range(length): if expression_parts[i] == '*/1': expression_parts[i] = '*' """ Convert Month,DOW,Year step values with a starting value (i.e. not '*') to between expressions. This allows us to reuse the between expression handling for step values. For Example: - month part '3/2' will be converted to '3-12/2' (every 2 months between March and December) - DOW part '3/2' will be converted to '3-6/2' (every 2 days between Tuesday and Saturday) """ if '/' in expression_parts[i] and any(exp in expression_parts[i] for exp in ['*', '-', ',']) is False: choices = {(4): '12', (5): '6', (6): '9999'} step_range_through = choices.get(i) if step_range_through is not None: parts = expression_parts[i].split('/') expression_parts[i] = '{0}-{1}/{2}'.format(parts[0], step_range_through, parts[1])
normalize_expression
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, expression_parts):\n expression_parts[3] = expression_parts[3].replace('?', '*')\n expression_parts[5] = expression_parts[5].replace('?', '*')\n if expression_parts[0].startswith('0/'):\n expression_parts[0] = expression_parts[0].replace('0/', '*/')\n if expression_parts[1].startswith('0/'):\n expression_parts[1] = expression_parts[1].replace('0/', '*/')\n if expression_parts[2].startswith('0/'):\n expression_parts[2] = expression_parts[2].replace('0/', '*/')\n if expression_parts[3].startswith('1/'):\n expression_parts[3] = expression_parts[3].replace('1/', '*/')\n if expression_parts[4].startswith('1/'):\n expression_parts[4] = expression_parts[4].replace('1/', '*/')\n if expression_parts[5].startswith('1/'):\n expression_parts[5] = expression_parts[5].replace('1/', '*/')\n if expression_parts[6].startswith('1/'):\n expression_parts[6] = expression_parts[6].replace('1/', '*/')\n if self._options.day_of_week_start_index_zero is False:\n expression_parts[5] = self.decrease_days_of_week(expression_parts[5])\n if expression_parts[3] == '?':\n expression_parts[3] = '*'\n for day_number in self._cron_days:\n expression_parts[5] = expression_parts[5].upper().replace(self.\n _cron_days[day_number], str(day_number))\n for month_number in self._cron_months:\n expression_parts[4] = expression_parts[4].upper().replace(self.\n _cron_months[month_number], str(month_number))\n if expression_parts[0] == '0':\n expression_parts[0] = ''\n length = len(expression_parts)\n for i in range(length):\n if expression_parts[i] == '*/1':\n expression_parts[i] = '*'\n \"\"\"\n Convert Month,DOW,Year step values with a starting value (i.e. not '*') to between expressions.\n This allows us to reuse the between expression handling for step values.\n\n For Example:\n - month part '3/2' will be converted to '3-12/2' (every 2 months between March and December)\n - DOW part '3/2' will be converted to '3-6/2' (every 2 days between Tuesday and Saturday)\n \"\"\"\n if '/' in expression_parts[i] and any(exp in expression_parts[i] for\n exp in ['*', '-', ',']) is False:\n choices = {(4): '12', (5): '6', (6): '9999'}\n step_range_through = choices.get(i)\n if step_range_through is not None:\n parts = expression_parts[i].split('/')\n expression_parts[i] = '{0}-{1}/{2}'.format(parts[0],\n step_range_through, parts[1])\n", "role": "user" }, { "content": "normalize_expression", "role": "assistant" } ]
def x(envlist): groups = {} for env in envlist: envpy, category = env.split('-')[0:2] if category == 'lint': category = 'unit' try: groups[envpy, category].append(env) except KeyError: groups[envpy, category] = [env] return sorted((envpy, category, envs) for (envpy, category), envs in groups.items())
group_envs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(envlist):\n groups = {}\n for env in envlist:\n envpy, category = env.split('-')[0:2]\n if category == 'lint':\n category = 'unit'\n try:\n groups[envpy, category].append(env)\n except KeyError:\n groups[envpy, category] = [env]\n return sorted((envpy, category, envs) for (envpy, category), envs in\n groups.items())\n", "role": "user" }, { "content": "group_envs", "role": "assistant" } ]
def x(payload, key=None): assert payload is None or isinstance(payload, bytes ), 'payload={!r} should be bytes or None'.format(payload) assert key is None or isinstance(key, bytes ), 'key={!r} should be bytes or None'.format(key) return Message(0, 0, key, payload)
create_message
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(payload, key=None):\n assert payload is None or isinstance(payload, bytes\n ), 'payload={!r} should be bytes or None'.format(payload)\n assert key is None or isinstance(key, bytes\n ), 'key={!r} should be bytes or None'.format(key)\n return Message(0, 0, key, payload)\n", "role": "user" }, { "content": "create_message", "role": "assistant" } ]
def x(message_set): encoded_message_set = KafkaCodec._encode_message_set(message_set) gzipped = gzip_encode(encoded_message_set) return Message(0, CODEC_GZIP, None, gzipped)
create_gzip_message
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(message_set):\n encoded_message_set = KafkaCodec._encode_message_set(message_set)\n gzipped = gzip_encode(encoded_message_set)\n return Message(0, CODEC_GZIP, None, gzipped)\n", "role": "user" }, { "content": "create_gzip_message", "role": "assistant" } ]
def x(message_set): encoded_message_set = KafkaCodec._encode_message_set(message_set) snapped = snappy_encode(encoded_message_set) return Message(0, CODEC_SNAPPY, None, snapped)
create_snappy_message
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(message_set):\n encoded_message_set = KafkaCodec._encode_message_set(message_set)\n snapped = snappy_encode(encoded_message_set)\n return Message(0, CODEC_SNAPPY, None, snapped)\n", "role": "user" }, { "content": "create_snappy_message", "role": "assistant" } ]
def x(requests, codec=CODEC_NONE): msglist = [] for req in requests: msglist.extend([create_message(m, key=req.key) for m in req.messages]) if codec == CODEC_NONE: return msglist elif codec == CODEC_GZIP: return [create_gzip_message(msglist)] elif codec == CODEC_SNAPPY: return [create_snappy_message(msglist)] else: raise UnsupportedCodecError('Codec 0x%02x unsupported' % codec)
create_message_set
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(requests, codec=CODEC_NONE):\n msglist = []\n for req in requests:\n msglist.extend([create_message(m, key=req.key) for m in req.messages])\n if codec == CODEC_NONE:\n return msglist\n elif codec == CODEC_GZIP:\n return [create_gzip_message(msglist)]\n elif codec == CODEC_SNAPPY:\n return [create_snappy_message(msglist)]\n else:\n raise UnsupportedCodecError('Codec 0x%02x unsupported' % codec)\n", "role": "user" }, { "content": "create_message_set", "role": "assistant" } ]
def x(cls, client_id, correlation_id, request_key, api_version=0): return struct.pack('>hhih', request_key, api_version, correlation_id, len(client_id)) + client_id
encode_message_header
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, client_id, correlation_id, request_key, api_version=0):\n return struct.pack('>hhih', request_key, api_version, correlation_id,\n len(client_id)) + client_id\n", "role": "user" }, { "content": "encode_message_header", "role": "assistant" } ]
def x(cls, messages, offset=None): message_set = [] incr = 1 if offset is None: incr = 0 offset = 0 for message in messages: encoded_message = KafkaCodec._encode_message(message) message_set.append(struct.pack('>qi', offset, len(encoded_message))) message_set.append(encoded_message) offset += incr return b''.join(message_set)
encode_message_set
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, messages, offset=None):\n message_set = []\n incr = 1\n if offset is None:\n incr = 0\n offset = 0\n for message in messages:\n encoded_message = KafkaCodec._encode_message(message)\n message_set.append(struct.pack('>qi', offset, len(encoded_message)))\n message_set.append(encoded_message)\n offset += incr\n return b''.join(message_set)\n", "role": "user" }, { "content": "encode_message_set", "role": "assistant" } ]
def x(cls, message): if message.magic == 0: msg = struct.pack('>BB', message.magic, message.attributes) msg += write_int_string(message.key) msg += write_int_string(message.value) crc = zlib.crc32(msg) & 4294967295 msg = struct.pack('>I', crc) + msg else: raise ProtocolError('Unexpected magic number: %d' % message.magic) return msg
encode_message
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, message):\n if message.magic == 0:\n msg = struct.pack('>BB', message.magic, message.attributes)\n msg += write_int_string(message.key)\n msg += write_int_string(message.value)\n crc = zlib.crc32(msg) & 4294967295\n msg = struct.pack('>I', crc) + msg\n else:\n raise ProtocolError('Unexpected magic number: %d' % message.magic)\n return msg\n", "role": "user" }, { "content": "encode_message", "role": "assistant" } ]
def x(cls, client_id, correlation_id, payloads=None, acks=1, timeout= DEFAULT_REPLICAS_ACK_TIMEOUT_MSECS): if not isinstance(client_id, bytes): raise TypeError('client_id={!r} should be bytes'.format(client_id)) payloads = [] if payloads is None else payloads grouped_payloads = group_by_topic_and_partition(payloads) message = cls._encode_message_header(client_id, correlation_id, KafkaCodec.PRODUCE_KEY) message += struct.pack('>hii', acks, timeout, len(grouped_payloads)) for topic, topic_payloads in grouped_payloads.items(): message += write_short_ascii(topic) message += struct.pack('>i', len(topic_payloads)) for partition, payload in topic_payloads.items(): msg_set = KafkaCodec._encode_message_set(payload.messages) message += struct.pack('>ii', partition, len(msg_set)) message += msg_set return message
encode_produce_request
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, client_id, correlation_id, payloads=None, acks=1, timeout=\n DEFAULT_REPLICAS_ACK_TIMEOUT_MSECS):\n if not isinstance(client_id, bytes):\n raise TypeError('client_id={!r} should be bytes'.format(client_id))\n payloads = [] if payloads is None else payloads\n grouped_payloads = group_by_topic_and_partition(payloads)\n message = cls._encode_message_header(client_id, correlation_id,\n KafkaCodec.PRODUCE_KEY)\n message += struct.pack('>hii', acks, timeout, len(grouped_payloads))\n for topic, topic_payloads in grouped_payloads.items():\n message += write_short_ascii(topic)\n message += struct.pack('>i', len(topic_payloads))\n for partition, payload in topic_payloads.items():\n msg_set = KafkaCodec._encode_message_set(payload.messages)\n message += struct.pack('>ii', partition, len(msg_set))\n message += msg_set\n return message\n", "role": "user" }, { "content": "encode_produce_request", "role": "assistant" } ]
def x(cls, data): (correlation_id, num_topics), cur = relative_unpack('>ii', data, 0) for _i in range(num_topics): topic, cur = read_short_ascii(data, cur) (num_partitions,), cur = relative_unpack('>i', data, cur) for _i in range(num_partitions): (partition, error, offset), cur = relative_unpack('>ihq', data, cur ) yield ProduceResponse(topic, partition, error, offset)
decode_produce_response
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, data):\n (correlation_id, num_topics), cur = relative_unpack('>ii', data, 0)\n for _i in range(num_topics):\n topic, cur = read_short_ascii(data, cur)\n (num_partitions,), cur = relative_unpack('>i', data, cur)\n for _i in range(num_partitions):\n (partition, error, offset), cur = relative_unpack('>ihq', data, cur\n )\n yield ProduceResponse(topic, partition, error, offset)\n", "role": "user" }, { "content": "decode_produce_response", "role": "assistant" } ]
def x(cls, client_id, correlation_id, payloads=None, max_wait_time=100, min_bytes=4096): payloads = [] if payloads is None else payloads grouped_payloads = group_by_topic_and_partition(payloads) message = cls._encode_message_header(client_id, correlation_id, KafkaCodec.FETCH_KEY) assert isinstance(max_wait_time, int) message += struct.pack('>iiii', -1, max_wait_time, min_bytes, len( grouped_payloads)) for topic, topic_payloads in grouped_payloads.items(): message += write_short_ascii(topic) message += struct.pack('>i', len(topic_payloads)) for partition, payload in topic_payloads.items(): message += struct.pack('>iqi', partition, payload.offset, payload.max_bytes) return message
encode_fetch_request
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, client_id, correlation_id, payloads=None, max_wait_time=100,\n min_bytes=4096):\n payloads = [] if payloads is None else payloads\n grouped_payloads = group_by_topic_and_partition(payloads)\n message = cls._encode_message_header(client_id, correlation_id,\n KafkaCodec.FETCH_KEY)\n assert isinstance(max_wait_time, int)\n message += struct.pack('>iiii', -1, max_wait_time, min_bytes, len(\n grouped_payloads))\n for topic, topic_payloads in grouped_payloads.items():\n message += write_short_ascii(topic)\n message += struct.pack('>i', len(topic_payloads))\n for partition, payload in topic_payloads.items():\n message += struct.pack('>iqi', partition, payload.offset,\n payload.max_bytes)\n return message\n", "role": "user" }, { "content": "encode_fetch_request", "role": "assistant" } ]
def x(cls, data): (correlation_id, num_topics), cur = relative_unpack('>ii', data, 0) for _i in range(num_topics): topic, cur = read_short_ascii(data, cur) (num_partitions,), cur = relative_unpack('>i', data, cur) for _i in range(num_partitions): (partition, error, highwater_mark_offset), cur = relative_unpack( '>ihq', data, cur) message_set, cur = read_int_string(data, cur) yield FetchResponse(topic, partition, error, highwater_mark_offset, KafkaCodec._decode_message_set_iter( message_set))
decode_fetch_response
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, data):\n (correlation_id, num_topics), cur = relative_unpack('>ii', data, 0)\n for _i in range(num_topics):\n topic, cur = read_short_ascii(data, cur)\n (num_partitions,), cur = relative_unpack('>i', data, cur)\n for _i in range(num_partitions):\n (partition, error, highwater_mark_offset), cur = relative_unpack(\n '>ihq', data, cur)\n message_set, cur = read_int_string(data, cur)\n yield FetchResponse(topic, partition, error,\n highwater_mark_offset, KafkaCodec._decode_message_set_iter(\n message_set))\n", "role": "user" }, { "content": "decode_fetch_response", "role": "assistant" } ]
def x(cls, data): (correlation_id, num_topics), cur = relative_unpack('>ii', data, 0) for _i in range(num_topics): topic, cur = read_short_ascii(data, cur) (num_partitions,), cur = relative_unpack('>i', data, cur) for _i in range(num_partitions): (partition, error, num_offsets), cur = relative_unpack('>ihi', data, cur) offsets = [] for _i in range(num_offsets): (offset,), cur = relative_unpack('>q', data, cur) offsets.append(offset) yield OffsetResponse(topic, partition, error, tuple(offsets))
decode_offset_response
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, data):\n (correlation_id, num_topics), cur = relative_unpack('>ii', data, 0)\n for _i in range(num_topics):\n topic, cur = read_short_ascii(data, cur)\n (num_partitions,), cur = relative_unpack('>i', data, cur)\n for _i in range(num_partitions):\n (partition, error, num_offsets), cur = relative_unpack('>ihi',\n data, cur)\n offsets = []\n for _i in range(num_offsets):\n (offset,), cur = relative_unpack('>q', data, cur)\n offsets.append(offset)\n yield OffsetResponse(topic, partition, error, tuple(offsets))\n", "role": "user" }, { "content": "decode_offset_response", "role": "assistant" } ]
def x(cls, client_id, correlation_id, topics=None): topics = [] if topics is None else topics message = [cls._encode_message_header(client_id, correlation_id, KafkaCodec.METADATA_KEY), struct.pack('>i', len(topics))] for topic in topics: message.append(write_short_ascii(topic)) return b''.join(message)
encode_metadata_request
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, client_id, correlation_id, topics=None):\n topics = [] if topics is None else topics\n message = [cls._encode_message_header(client_id, correlation_id,\n KafkaCodec.METADATA_KEY), struct.pack('>i', len(topics))]\n for topic in topics:\n message.append(write_short_ascii(topic))\n return b''.join(message)\n", "role": "user" }, { "content": "encode_metadata_request", "role": "assistant" } ]
def x(cls, data): (correlation_id, numbrokers), cur = relative_unpack('>ii', data, 0) if numbrokers > MAX_BROKERS: raise InvalidMessageError('Brokers:{} exceeds max:{}'.format( numbrokers, MAX_BROKERS)) brokers = {} for _i in range(numbrokers): (nodeId,), cur = relative_unpack('>i', data, cur) host, cur = read_short_ascii(data, cur) (port,), cur = relative_unpack('>i', data, cur) brokers[nodeId] = BrokerMetadata(nodeId, nativeString(host), port) (num_topics,), cur = relative_unpack('>i', data, cur) topic_metadata = {} for _i in range(num_topics): (topic_error,), cur = relative_unpack('>h', data, cur) topic_name, cur = read_short_ascii(data, cur) (num_partitions,), cur = relative_unpack('>i', data, cur) partition_metadata = {} for _j in range(num_partitions): (partition_error_code, partition, leader, numReplicas ), cur = relative_unpack('>hiii', data, cur) replicas, cur = relative_unpack('>%di' % numReplicas, data, cur) (num_isr,), cur = relative_unpack('>i', data, cur) isr, cur = relative_unpack('>%di' % num_isr, data, cur) partition_metadata[partition] = PartitionMetadata(topic_name, partition, partition_error_code, leader, replicas, isr) topic_metadata[topic_name] = TopicMetadata(topic_name, topic_error, partition_metadata) return brokers, topic_metadata
decode_metadata_response
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, data):\n (correlation_id, numbrokers), cur = relative_unpack('>ii', data, 0)\n if numbrokers > MAX_BROKERS:\n raise InvalidMessageError('Brokers:{} exceeds max:{}'.format(\n numbrokers, MAX_BROKERS))\n brokers = {}\n for _i in range(numbrokers):\n (nodeId,), cur = relative_unpack('>i', data, cur)\n host, cur = read_short_ascii(data, cur)\n (port,), cur = relative_unpack('>i', data, cur)\n brokers[nodeId] = BrokerMetadata(nodeId, nativeString(host), port)\n (num_topics,), cur = relative_unpack('>i', data, cur)\n topic_metadata = {}\n for _i in range(num_topics):\n (topic_error,), cur = relative_unpack('>h', data, cur)\n topic_name, cur = read_short_ascii(data, cur)\n (num_partitions,), cur = relative_unpack('>i', data, cur)\n partition_metadata = {}\n for _j in range(num_partitions):\n (partition_error_code, partition, leader, numReplicas\n ), cur = relative_unpack('>hiii', data, cur)\n replicas, cur = relative_unpack('>%di' % numReplicas, data, cur)\n (num_isr,), cur = relative_unpack('>i', data, cur)\n isr, cur = relative_unpack('>%di' % num_isr, data, cur)\n partition_metadata[partition] = PartitionMetadata(topic_name,\n partition, partition_error_code, leader, replicas, isr)\n topic_metadata[topic_name] = TopicMetadata(topic_name, topic_error,\n partition_metadata)\n return brokers, topic_metadata\n", "role": "user" }, { "content": "decode_metadata_response", "role": "assistant" } ]
def x(cls, client_id, correlation_id, consumer_group): message = cls._encode_message_header(client_id, correlation_id, KafkaCodec.CONSUMER_METADATA_KEY) message += write_short_ascii(consumer_group) return message
encode_consumermetadata_request
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, client_id, correlation_id, consumer_group):\n message = cls._encode_message_header(client_id, correlation_id,\n KafkaCodec.CONSUMER_METADATA_KEY)\n message += write_short_ascii(consumer_group)\n return message\n", "role": "user" }, { "content": "encode_consumermetadata_request", "role": "assistant" } ]
def x(cls, data): (correlation_id, error_code, node_id), cur = relative_unpack('>ihi', data, 0) host, cur = read_short_ascii(data, cur) (port,), cur = relative_unpack('>i', data, cur) return ConsumerMetadataResponse(error_code, node_id, nativeString(host), port)
decode_consumermetadata_response
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, data):\n (correlation_id, error_code, node_id), cur = relative_unpack('>ihi',\n data, 0)\n host, cur = read_short_ascii(data, cur)\n (port,), cur = relative_unpack('>i', data, cur)\n return ConsumerMetadataResponse(error_code, node_id, nativeString(host),\n port)\n", "role": "user" }, { "content": "decode_consumermetadata_response", "role": "assistant" } ]
def x(cls, client_id, correlation_id, group, group_generation_id, consumer_id, payloads): grouped_payloads = group_by_topic_and_partition(payloads) message = cls._encode_message_header(client_id, correlation_id, KafkaCodec.OFFSET_COMMIT_KEY, api_version=1) message += write_short_ascii(group) message += struct.pack('>i', group_generation_id) message += write_short_ascii(consumer_id) message += struct.pack('>i', len(grouped_payloads)) for topic, topic_payloads in grouped_payloads.items(): message += write_short_ascii(topic) message += struct.pack('>i', len(topic_payloads)) for partition, payload in topic_payloads.items(): message += struct.pack('>iqq', partition, payload.offset, payload.timestamp) message += write_short_bytes(payload.metadata) return message
encode_offset_commit_request
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, client_id, correlation_id, group, group_generation_id,\n consumer_id, payloads):\n grouped_payloads = group_by_topic_and_partition(payloads)\n message = cls._encode_message_header(client_id, correlation_id,\n KafkaCodec.OFFSET_COMMIT_KEY, api_version=1)\n message += write_short_ascii(group)\n message += struct.pack('>i', group_generation_id)\n message += write_short_ascii(consumer_id)\n message += struct.pack('>i', len(grouped_payloads))\n for topic, topic_payloads in grouped_payloads.items():\n message += write_short_ascii(topic)\n message += struct.pack('>i', len(topic_payloads))\n for partition, payload in topic_payloads.items():\n message += struct.pack('>iqq', partition, payload.offset,\n payload.timestamp)\n message += write_short_bytes(payload.metadata)\n return message\n", "role": "user" }, { "content": "encode_offset_commit_request", "role": "assistant" } ]
def x(cls, data): (correlation_id,), cur = relative_unpack('>i', data, 0) (num_topics,), cur = relative_unpack('>i', data, cur) for _i in range(num_topics): topic, cur = read_short_ascii(data, cur) (num_partitions,), cur = relative_unpack('>i', data, cur) for _i in range(num_partitions): (partition, error), cur = relative_unpack('>ih', data, cur) yield OffsetCommitResponse(topic, partition, error)
decode_offset_commit_response
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, data):\n (correlation_id,), cur = relative_unpack('>i', data, 0)\n (num_topics,), cur = relative_unpack('>i', data, cur)\n for _i in range(num_topics):\n topic, cur = read_short_ascii(data, cur)\n (num_partitions,), cur = relative_unpack('>i', data, cur)\n for _i in range(num_partitions):\n (partition, error), cur = relative_unpack('>ih', data, cur)\n yield OffsetCommitResponse(topic, partition, error)\n", "role": "user" }, { "content": "decode_offset_commit_response", "role": "assistant" } ]
def x(cls, client_id, correlation_id, group, payloads): grouped_payloads = group_by_topic_and_partition(payloads) message = cls._encode_message_header(client_id, correlation_id, KafkaCodec.OFFSET_FETCH_KEY, api_version=1) message += write_short_ascii(group) message += struct.pack('>i', len(grouped_payloads)) for topic, topic_payloads in grouped_payloads.items(): message += write_short_ascii(topic) message += struct.pack('>i', len(topic_payloads)) for partition in topic_payloads: message += struct.pack('>i', partition) return message
encode_offset_fetch_request
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, client_id, correlation_id, group, payloads):\n grouped_payloads = group_by_topic_and_partition(payloads)\n message = cls._encode_message_header(client_id, correlation_id,\n KafkaCodec.OFFSET_FETCH_KEY, api_version=1)\n message += write_short_ascii(group)\n message += struct.pack('>i', len(grouped_payloads))\n for topic, topic_payloads in grouped_payloads.items():\n message += write_short_ascii(topic)\n message += struct.pack('>i', len(topic_payloads))\n for partition in topic_payloads:\n message += struct.pack('>i', partition)\n return message\n", "role": "user" }, { "content": "encode_offset_fetch_request", "role": "assistant" } ]
def x(cls, data): (correlation_id,), cur = relative_unpack('>i', data, 0) (num_topics,), cur = relative_unpack('>i', data, cur) for _i in range(num_topics): topic, cur = read_short_ascii(data, cur) (num_partitions,), cur = relative_unpack('>i', data, cur) for _i in range(num_partitions): (partition, offset), cur = relative_unpack('>iq', data, cur) metadata, cur = read_short_bytes(data, cur) (error,), cur = relative_unpack('>h', data, cur) yield OffsetFetchResponse(topic, partition, offset, metadata, error )
decode_offset_fetch_response
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, data):\n (correlation_id,), cur = relative_unpack('>i', data, 0)\n (num_topics,), cur = relative_unpack('>i', data, cur)\n for _i in range(num_topics):\n topic, cur = read_short_ascii(data, cur)\n (num_partitions,), cur = relative_unpack('>i', data, cur)\n for _i in range(num_partitions):\n (partition, offset), cur = relative_unpack('>iq', data, cur)\n metadata, cur = read_short_bytes(data, cur)\n (error,), cur = relative_unpack('>h', data, cur)\n yield OffsetFetchResponse(topic, partition, offset, metadata, error\n )\n", "role": "user" }, { "content": "decode_offset_fetch_response", "role": "assistant" } ]
def x(self, response): correlation_id = response[0:4] try: d = self._pending.pop(correlation_id) except KeyError: self._log.warn( 'Response has unknown correlation ID {correlation_id!r}. Dropping connection to {peer}.' , correlation_id=correlation_id, peer=self.transport.getPeer()) self.transport.loseConnection() else: d.callback(response)
stringReceived
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, response):\n correlation_id = response[0:4]\n try:\n d = self._pending.pop(correlation_id)\n except KeyError:\n self._log.warn(\n 'Response has unknown correlation ID {correlation_id!r}. Dropping connection to {peer}.'\n , correlation_id=correlation_id, peer=self.transport.getPeer())\n self.transport.loseConnection()\n else:\n d.callback(response)\n", "role": "user" }, { "content": "stringReceived", "role": "assistant" } ]
def x(self, reason=connectionDone): self._failed = reason pending, self._pending = self._pending, None for d in pending.values(): d.errback(reason)
connectionLost
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, reason=connectionDone):\n self._failed = reason\n pending, self._pending = self._pending, None\n for d in pending.values():\n d.errback(reason)\n", "role": "user" }, { "content": "connectionLost", "role": "assistant" } ]
def x(self, request): if self._failed is not None: return fail(self._failed) correlation_id = request[4:8] assert correlation_id not in self._pending d = Deferred() self.sendString(request) self._pending[correlation_id] = d return d
request
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, request):\n if self._failed is not None:\n return fail(self._failed)\n correlation_id = request[4:8]\n assert correlation_id not in self._pending\n d = Deferred()\n self.sendString(request)\n self._pending[correlation_id] = d\n return d\n", "role": "user" }, { "content": "request", "role": "assistant" } ]
def x(self, topic, key=None, msgs=()): try: topic = _coerce_topic(topic) if key is not None and not isinstance(key, bytes): raise TypeError('key={!r} must be bytes or None'.format(key)) if not msgs: raise ValueError('msgs must be a non-empty sequence') msg_cnt = len(msgs) byte_cnt = 0 for index, m in enumerate(msgs): if m is None: continue if not isinstance(m, bytes): raise TypeError( 'Message {} to topic {} ({!r:.100}) has type {}, but must have type {}' .format(index, topic, m, type(m).__name__, type(bytes). __name__)) byte_cnt += len(m) except Exception: return fail() d = Deferred(self._cancel_send_messages) self._batch_reqs.append(SendRequest(topic, key, msgs, d)) self._waitingMsgCount += msg_cnt self._waitingByteCount += byte_cnt self._outstanding.append(d) d.addBoth(self._remove_from_outstanding, d) self._check_send_batch() return d
send_messages
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, topic, key=None, msgs=()):\n try:\n topic = _coerce_topic(topic)\n if key is not None and not isinstance(key, bytes):\n raise TypeError('key={!r} must be bytes or None'.format(key))\n if not msgs:\n raise ValueError('msgs must be a non-empty sequence')\n msg_cnt = len(msgs)\n byte_cnt = 0\n for index, m in enumerate(msgs):\n if m is None:\n continue\n if not isinstance(m, bytes):\n raise TypeError(\n 'Message {} to topic {} ({!r:.100}) has type {}, but must have type {}'\n .format(index, topic, m, type(m).__name__, type(bytes).\n __name__))\n byte_cnt += len(m)\n except Exception:\n return fail()\n d = Deferred(self._cancel_send_messages)\n self._batch_reqs.append(SendRequest(topic, key, msgs, d))\n self._waitingMsgCount += msg_cnt\n self._waitingByteCount += byte_cnt\n self._outstanding.append(d)\n d.addBoth(self._remove_from_outstanding, d)\n self._check_send_batch()\n return d\n", "role": "user" }, { "content": "send_messages", "role": "assistant" } ]
def x(self): self.stopping = True if self._batch_send_d: self._batch_send_d.cancel() if self.batch_every_t is not None: if self._sendLooper is not None: self._sendLooper.stop() self._cancel_outstanding() return self._sendLooperD or succeed(None)
stop
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n self.stopping = True\n if self._batch_send_d:\n self._batch_send_d.cancel()\n if self.batch_every_t is not None:\n if self._sendLooper is not None:\n self._sendLooper.stop()\n self._cancel_outstanding()\n return self._sendLooperD or succeed(None)\n", "role": "user" }, { "content": "stop", "role": "assistant" } ]
def x(self, fail): log.warning('_send_timer_failed:%r: %s', fail, fail.getBriefTraceback()) self._sendLooperD = self._sendLooper.start(self.batch_every_t, now=False)
send_timer_failed
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, fail):\n log.warning('_send_timer_failed:%r: %s', fail, fail.getBriefTraceback())\n self._sendLooperD = self._sendLooper.start(self.batch_every_t, now=False)\n", "role": "user" }, { "content": "send_timer_failed", "role": "assistant" } ]
def x(self, lCall): if self._sendLooper is not lCall: log.warning('commitTimerStopped with wrong timer:%s not:%s', lCall, self._sendLooper) else: self._sendLooper = None self._sendLooperD = None
send_timer_stopped
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, lCall):\n if self._sendLooper is not lCall:\n log.warning('commitTimerStopped with wrong timer:%s not:%s', lCall,\n self._sendLooper)\n else:\n self._sendLooper = None\n self._sendLooperD = None\n", "role": "user" }, { "content": "send_timer_stopped", "role": "assistant" } ]
def x(self, topic, key=None): while self.client.metadata_error_for_topic(topic): if self._req_attempts >= self._max_attempts: _check_error(self.client.metadata_error_for_topic(topic)) yield self.client.load_metadata_for_topics(topic) if not self.client.metadata_error_for_topic(topic): break self._req_attempts += 1 d = Deferred() self.client.reactor.callLater(self._retry_interval, d.callback, True) self._retry_interval *= self.RETRY_INTERVAL_FACTOR yield d partitions = self.client.topic_partitions[topic] if topic not in self.partitioners: self.partitioners[topic] = self.partitioner_class(topic, partitions) partition = self.partitioners[topic].partition(key, partitions) returnValue(partition)
next_partition
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, topic, key=None):\n while self.client.metadata_error_for_topic(topic):\n if self._req_attempts >= self._max_attempts:\n _check_error(self.client.metadata_error_for_topic(topic))\n yield self.client.load_metadata_for_topics(topic)\n if not self.client.metadata_error_for_topic(topic):\n break\n self._req_attempts += 1\n d = Deferred()\n self.client.reactor.callLater(self._retry_interval, d.callback, True)\n self._retry_interval *= self.RETRY_INTERVAL_FACTOR\n yield d\n partitions = self.client.topic_partitions[topic]\n if topic not in self.partitioners:\n self.partitioners[topic] = self.partitioner_class(topic, partitions)\n partition = self.partitioners[topic].partition(key, partitions)\n returnValue(partition)\n", "role": "user" }, { "content": "next_partition", "role": "assistant" } ]
def x(self, parts_results, requests): reqsByTopicPart = defaultdict(list) payloadsByTopicPart = defaultdict(list) deferredsByTopicPart = defaultdict(list) for (success, part_or_failure), req in zip(parts_results, requests): if req.deferred.called: continue if not success: req.deferred.errback(part_or_failure) continue topicPart = TopicAndPartition(req.topic, part_or_failure) reqsByTopicPart[topicPart].append(req) deferredsByTopicPart[topicPart].append(req.deferred) payloads = [] for (topic, partition), reqs in reqsByTopicPart.items(): msgSet = create_message_set(reqs, self.codec) req = ProduceRequest(topic, partition, msgSet) topicPart = TopicAndPartition(topic, partition) payloads.append(req) payloadsByTopicPart[topicPart] = req if not payloads: return d = self.client.send_produce_request(payloads, acks=self.req_acks, timeout=self.ack_timeout, fail_on_error=False) self._req_attempts += 1 d.addBoth(self._handle_send_response, payloadsByTopicPart, deferredsByTopicPart) return d
send_requests
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, parts_results, requests):\n reqsByTopicPart = defaultdict(list)\n payloadsByTopicPart = defaultdict(list)\n deferredsByTopicPart = defaultdict(list)\n for (success, part_or_failure), req in zip(parts_results, requests):\n if req.deferred.called:\n continue\n if not success:\n req.deferred.errback(part_or_failure)\n continue\n topicPart = TopicAndPartition(req.topic, part_or_failure)\n reqsByTopicPart[topicPart].append(req)\n deferredsByTopicPart[topicPart].append(req.deferred)\n payloads = []\n for (topic, partition), reqs in reqsByTopicPart.items():\n msgSet = create_message_set(reqs, self.codec)\n req = ProduceRequest(topic, partition, msgSet)\n topicPart = TopicAndPartition(topic, partition)\n payloads.append(req)\n payloadsByTopicPart[topicPart] = req\n if not payloads:\n return\n d = self.client.send_produce_request(payloads, acks=self.req_acks,\n timeout=self.ack_timeout, fail_on_error=False)\n self._req_attempts += 1\n d.addBoth(self._handle_send_response, payloadsByTopicPart,\n deferredsByTopicPart)\n return d\n", "role": "user" }, { "content": "send_requests", "role": "assistant" } ]
def x(self, resp): self._batch_send_d = None self._req_attempts = 0 self._retry_interval = self._init_retry_interval if isinstance(resp, Failure) and not resp.check(tid_CancelledError, CancelledError): log.error('Failure detected in _complete_batch_send: %r\n%r', resp, resp.getTraceback()) return
complete_batch_send
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, resp):\n self._batch_send_d = None\n self._req_attempts = 0\n self._retry_interval = self._init_retry_interval\n if isinstance(resp, Failure) and not resp.check(tid_CancelledError,\n CancelledError):\n log.error('Failure detected in _complete_batch_send: %r\\n%r', resp,\n resp.getTraceback())\n return\n", "role": "user" }, { "content": "complete_batch_send", "role": "assistant" } ]
def x(self, result=None): if (self.batch_every_n and self.batch_every_n <= self._waitingMsgCount or self.batch_every_b and self.batch_every_b <= self._waitingByteCount): self._send_batch() return result
check_send_batch
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, result=None):\n if (self.batch_every_n and self.batch_every_n <= self._waitingMsgCount or\n self.batch_every_b and self.batch_every_b <= self._waitingByteCount):\n self._send_batch()\n return result\n", "role": "user" }, { "content": "check_send_batch", "role": "assistant" } ]
def x(self): if not self._batch_reqs or self._batch_send_d: return requests, self._batch_reqs = self._batch_reqs, [] self._waitingByteCount = 0 self._waitingMsgCount = 0 d_list = [] for req in requests: d_list.append(self._next_partition(req.topic, req.key)) d = self._batch_send_d = Deferred() d.addCallback(lambda r: DeferredList(d_list, consumeErrors=True)) d.addCallback(self._send_requests, requests) d.addBoth(self._complete_batch_send) d.addBoth(self._check_send_batch) d.callback(None)
send_batch
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if not self._batch_reqs or self._batch_send_d:\n return\n requests, self._batch_reqs = self._batch_reqs, []\n self._waitingByteCount = 0\n self._waitingMsgCount = 0\n d_list = []\n for req in requests:\n d_list.append(self._next_partition(req.topic, req.key))\n d = self._batch_send_d = Deferred()\n d.addCallback(lambda r: DeferredList(d_list, consumeErrors=True))\n d.addCallback(self._send_requests, requests)\n d.addBoth(self._complete_batch_send)\n d.addBoth(self._check_send_batch)\n d.callback(None)\n", "role": "user" }, { "content": "send_batch", "role": "assistant" } ]
def x(self, d): for req in self._batch_reqs: if req.deferred == d: msgs = req.messages self._waitingMsgCount -= len(msgs) for m in (_m for _m in msgs if _m is not None): self._waitingByteCount -= len(m) self._batch_reqs.remove(req) d.errback(CancelledError(request_sent=False)) return d.errback(CancelledError(request_sent=self._batch_send_d is not None)) return
cancel_send_messages
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, d):\n for req in self._batch_reqs:\n if req.deferred == d:\n msgs = req.messages\n self._waitingMsgCount -= len(msgs)\n for m in (_m for _m in msgs if _m is not None):\n self._waitingByteCount -= len(m)\n self._batch_reqs.remove(req)\n d.errback(CancelledError(request_sent=False))\n return\n d.errback(CancelledError(request_sent=self._batch_send_d is not None))\n return\n", "role": "user" }, { "content": "cancel_send_messages", "role": "assistant" } ]
def x(self, result, payloadsByTopicPart, deferredsByTopicPart): def x(d_list, result=None): for d in d_list: if not isinstance(d, Deferred): _deliver_result(d, result) elif not d.called: d.callback(result) def x(payloads): d = self.client.send_produce_request(payloads, acks=self.req_acks, timeout=self.ack_timeout, fail_on_error=False) self._req_attempts += 1 d.addBoth(self._handle_send_response, payloadsByTopicPart, deferredsByTopicPart) return d def x(failure, dc): dc.cancel() _deliver_result(deferredsByTopicPart.values(), failure) return failure def x(failed_payloads_with_errs): if self._req_attempts >= self._max_attempts: for p, f in failed_payloads_with_errs: t_and_p = TopicAndPartition(p.topic, p.partition) _deliver_result(deferredsByTopicPart[t_and_p], f) return d = Deferred() dc = self.client.reactor.callLater(self._retry_interval, d.callback, [p for p, f in failed_payloads]) self._retry_interval *= self.RETRY_INTERVAL_FACTOR d.addErrback(_cancel_retry, dc) reset_topics = set() for payload, e in failed_payloads: if isinstance(e, NotLeaderForPartitionError) or isinstance(e, UnknownTopicOrPartitionError): reset_topics.add(payload.topic) if reset_topics: self.client.reset_topic_metadata(*reset_topics) d.addCallback(_do_retry) return d failed_payloads = [] if not result: if self.req_acks == PRODUCER_ACK_NOT_REQUIRED: result = None else: result = Failure(NoResponseError()) _deliver_result(deferredsByTopicPart.values(), result) return elif isinstance(result, Failure): if not result.check(FailedPayloadsError): if result.check(KafkaError): failure, result = result, [] failed_payloads = [(p, failure) for p in payloadsByTopicPart.values()] else: if not result.check(tid_CancelledError): log.error('Unexpected failure: %r in _handle_send_response' , result) _deliver_result(deferredsByTopicPart.values(), result) return else: failure = result result = failure.value.args[0] failed_payloads = failure.value.args[1] for res in result: t_and_p = TopicAndPartition(res.topic, res.partition) t_and_p_err = _check_error(res, raiseException=False) if not t_and_p_err: d_list = deferredsByTopicPart[t_and_p] _deliver_result(d_list, res) else: p = payloadsByTopicPart[t_and_p] failed_payloads.append((p, t_and_p_err)) if failed_payloads: return _check_retry_payloads(failed_payloads) return
handle_send_response
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, result, payloadsByTopicPart, deferredsByTopicPart):\n\n def x(d_list, result=None):\n for d in d_list:\n if not isinstance(d, Deferred):\n _deliver_result(d, result)\n elif not d.called:\n d.callback(result)\n\n def x(payloads):\n d = self.client.send_produce_request(payloads, acks=self.req_acks,\n timeout=self.ack_timeout, fail_on_error=False)\n self._req_attempts += 1\n d.addBoth(self._handle_send_response, payloadsByTopicPart,\n deferredsByTopicPart)\n return d\n\n def x(failure, dc):\n dc.cancel()\n _deliver_result(deferredsByTopicPart.values(), failure)\n return failure\n\n def x(failed_payloads_with_errs):\n if self._req_attempts >= self._max_attempts:\n for p, f in failed_payloads_with_errs:\n t_and_p = TopicAndPartition(p.topic, p.partition)\n _deliver_result(deferredsByTopicPart[t_and_p], f)\n return\n d = Deferred()\n dc = self.client.reactor.callLater(self._retry_interval, d.callback,\n [p for p, f in failed_payloads])\n self._retry_interval *= self.RETRY_INTERVAL_FACTOR\n d.addErrback(_cancel_retry, dc)\n reset_topics = set()\n for payload, e in failed_payloads:\n if isinstance(e, NotLeaderForPartitionError) or isinstance(e,\n UnknownTopicOrPartitionError):\n reset_topics.add(payload.topic)\n if reset_topics:\n self.client.reset_topic_metadata(*reset_topics)\n d.addCallback(_do_retry)\n return d\n failed_payloads = []\n if not result:\n if self.req_acks == PRODUCER_ACK_NOT_REQUIRED:\n result = None\n else:\n result = Failure(NoResponseError())\n _deliver_result(deferredsByTopicPart.values(), result)\n return\n elif isinstance(result, Failure):\n if not result.check(FailedPayloadsError):\n if result.check(KafkaError):\n failure, result = result, []\n failed_payloads = [(p, failure) for p in\n payloadsByTopicPart.values()]\n else:\n if not result.check(tid_CancelledError):\n log.error('Unexpected failure: %r in _handle_send_response'\n , result)\n _deliver_result(deferredsByTopicPart.values(), result)\n return\n else:\n failure = result\n result = failure.value.args[0]\n failed_payloads = failure.value.args[1]\n for res in result:\n t_and_p = TopicAndPartition(res.topic, res.partition)\n t_and_p_err = _check_error(res, raiseException=False)\n if not t_and_p_err:\n d_list = deferredsByTopicPart[t_and_p]\n _deliver_result(d_list, res)\n else:\n p = payloadsByTopicPart[t_and_p]\n failed_payloads.append((p, t_and_p_err))\n if failed_payloads:\n return _check_retry_payloads(failed_payloads)\n return\n", "role": "user" }, { "content": "handle_send_response", "role": "assistant" } ]
def x(self): for d in list(self._outstanding): d.addErrback(lambda _: None) d.cancel()
cancel_outstanding
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n for d in list(self._outstanding):\n d.addErrback(lambda _: None)\n d.cancel()\n", "role": "user" }, { "content": "cancel_outstanding", "role": "assistant" } ]
def x(hosts): if isinstance(hosts, bytes): hosts = hosts.split(b',') elif isinstance(hosts, _unicode): hosts = hosts.split(u',') result = set() for host_port in hosts: res = nativeString(host_port).split(':') host = res[0].strip() port = int(res[1].strip()) if len(res) > 1 else DefaultKafkaPort result.add((host, port)) return sorted(result)
normalize_hosts
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(hosts):\n if isinstance(hosts, bytes):\n hosts = hosts.split(b',')\n elif isinstance(hosts, _unicode):\n hosts = hosts.split(u',')\n result = set()\n for host_port in hosts:\n res = nativeString(host_port).split(':')\n host = res[0].strip()\n port = int(res[1].strip()) if len(res) > 1 else DefaultKafkaPort\n result.add((host, port))\n return sorted(result)\n", "role": "user" }, { "content": "normalize_hosts", "role": "assistant" } ]
def x(self, *groups): groups = tuple(_coerce_consumer_group(g) for g in groups) for group in groups: if group in self.consumer_group_to_brokers: del self.consumer_group_to_brokers[group]
reset_consumer_group_metadata
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, *groups):\n groups = tuple(_coerce_consumer_group(g) for g in groups)\n for group in groups:\n if group in self.consumer_group_to_brokers:\n del self.consumer_group_to_brokers[group]\n", "role": "user" }, { "content": "reset_consumer_group_metadata", "role": "assistant" } ]
def x(self): self.topics_to_brokers.clear() self.topic_partitions.clear() self.topic_errors.clear() self.consumer_group_to_brokers.clear()
reset_all_metadata
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n self.topics_to_brokers.clear()\n self.topic_partitions.clear()\n self.topic_errors.clear()\n self.consumer_group_to_brokers.clear()\n", "role": "user" }, { "content": "reset_all_metadata", "role": "assistant" } ]
def x(self, topic): topic = _coerce_topic(topic) if topic not in self.topic_partitions: return False if not self.topic_partitions[topic]: return False return all(self.partition_fully_replicated(TopicAndPartition(topic, p)) for p in self.topic_partitions[topic])
topic_fully_replicated
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, topic):\n topic = _coerce_topic(topic)\n if topic not in self.topic_partitions:\n return False\n if not self.topic_partitions[topic]:\n return False\n return all(self.partition_fully_replicated(TopicAndPartition(topic, p)) for\n p in self.topic_partitions[topic])\n", "role": "user" }, { "content": "topic_fully_replicated", "role": "assistant" } ]
def x(self): log.debug('%r: close', self) self._closing = True brokerclients, self.clients = self.clients, None self._close_brokerclients(brokerclients.values()) self.reset_all_metadata() return self.close_dlist or defer.succeed(None)
close
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n log.debug('%r: close', self)\n self._closing = True\n brokerclients, self.clients = self.clients, None\n self._close_brokerclients(brokerclients.values())\n self.reset_all_metadata()\n return self.close_dlist or defer.succeed(None)\n", "role": "user" }, { "content": "close", "role": "assistant" } ]
def x(self, *topics): topics = tuple(_coerce_topic(t) for t in topics) log.debug('%r: load_metadata_for_topics(%s)', self, ', '.join(repr(t) for t in topics)) fetch_all_metadata = not topics requestId = self._next_id() request = KafkaCodec.encode_metadata_request(self._clientIdBytes, requestId, topics) def x(response): brokers, topics = KafkaCodec.decode_metadata_response(response) log.debug('%r: got metadata brokers=%r topics=%r', self, brokers, topics) if fetch_all_metadata: self._brokers = brokers self._topics = topics ok_to_remove = fetch_all_metadata and len(brokers) self._update_brokers(brokers.values(), remove=ok_to_remove) for topic, topic_metadata in topics.items(): _, topic_error, partitions = topic_metadata self.reset_topic_metadata(topic) self.topic_errors[topic] = topic_error if not partitions: log.warning('No partitions for %s, Err:%d', topic, topic_error) continue self.topic_partitions[topic] = [] for partition, meta in partitions.items(): self.topic_partitions[topic].append(partition) topic_part = TopicAndPartition(topic, partition) self.partition_meta[topic_part] = meta if meta.leader == -1: log.warning('No leader for topic %s partition %s', topic, partition) self.topics_to_brokers[topic_part] = None else: self.topics_to_brokers[topic_part] = brokers[meta.leader] self.topic_partitions[topic] = sorted(self.topic_partitions[topic]) return True def x(err): if err.check(t_CancelledError, CancelledError): return None log.error('Failed to retrieve metadata:%s', err) raise KafkaUnavailableError( 'Unable to load metadata from configured hosts: {!r}'.format(err)) d = self._send_broker_unaware_request(requestId, request) d.addCallbacks(_handleMetadataResponse, _handleMetadataErr) return d
load_metadata_for_topics
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, *topics):\n topics = tuple(_coerce_topic(t) for t in topics)\n log.debug('%r: load_metadata_for_topics(%s)', self, ', '.join(repr(t) for\n t in topics))\n fetch_all_metadata = not topics\n requestId = self._next_id()\n request = KafkaCodec.encode_metadata_request(self._clientIdBytes,\n requestId, topics)\n\n def x(response):\n brokers, topics = KafkaCodec.decode_metadata_response(response)\n log.debug('%r: got metadata brokers=%r topics=%r', self, brokers,\n topics)\n if fetch_all_metadata:\n self._brokers = brokers\n self._topics = topics\n ok_to_remove = fetch_all_metadata and len(brokers)\n self._update_brokers(brokers.values(), remove=ok_to_remove)\n for topic, topic_metadata in topics.items():\n _, topic_error, partitions = topic_metadata\n self.reset_topic_metadata(topic)\n self.topic_errors[topic] = topic_error\n if not partitions:\n log.warning('No partitions for %s, Err:%d', topic, topic_error)\n continue\n self.topic_partitions[topic] = []\n for partition, meta in partitions.items():\n self.topic_partitions[topic].append(partition)\n topic_part = TopicAndPartition(topic, partition)\n self.partition_meta[topic_part] = meta\n if meta.leader == -1:\n log.warning('No leader for topic %s partition %s',\n topic, partition)\n self.topics_to_brokers[topic_part] = None\n else:\n self.topics_to_brokers[topic_part] = brokers[meta.leader]\n self.topic_partitions[topic] = sorted(self.topic_partitions[topic])\n return True\n\n def x(err):\n if err.check(t_CancelledError, CancelledError):\n return None\n log.error('Failed to retrieve metadata:%s', err)\n raise KafkaUnavailableError(\n 'Unable to load metadata from configured hosts: {!r}'.format(err))\n d = self._send_broker_unaware_request(requestId, request)\n d.addCallbacks(_handleMetadataResponse, _handleMetadataErr)\n return d\n", "role": "user" }, { "content": "load_metadata_for_topics", "role": "assistant" } ]
def x(self, group): group = _coerce_consumer_group(group) log.debug('%r: load_consumer_metadata_for_group(%r)', self, group) if group in self.coordinator_fetches: d = defer.Deferred() self.coordinator_fetches[group][1].append(d) return d requestId = self._next_id() request = KafkaCodec.encode_consumermetadata_request(self. _clientIdBytes, requestId, group) def x(response_bytes): response = KafkaCodec.decode_consumermetadata_response(response_bytes) log.debug('%r: load_consumer_metadata_for_group(%r) -> %r', self, group, response) if response.error: raise BrokerResponseError.errnos.get(response.error, UnknownError)( response) bm = BrokerMetadata(response.node_id, response.host, response.port) self.consumer_group_to_brokers[group] = bm self._update_brokers([bm]) return True def x(err): log.error('Failed to retrieve consumer metadata for group %r', group, exc_info=(err.type, err.value, err.getTracebackObject())) self.reset_consumer_group_metadata(group) raise ConsumerCoordinatorNotAvailableError( 'Coordinator for group {!r} not available'.format(group)) def x(result): [_, ds] = self.coordinator_fetches.pop(group, None) for d in ds: d.callback(result) request_d = self._send_broker_unaware_request(requestId, request) d = defer.Deferred() self.coordinator_fetches[group] = request_d, [d] request_d.addCallback(_handleConsumerMetadataResponse) request_d.addErrback(_handleConsumerMetadataErr) request_d.addBoth(_propagate) return d
load_consumer_metadata_for_group
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, group):\n group = _coerce_consumer_group(group)\n log.debug('%r: load_consumer_metadata_for_group(%r)', self, group)\n if group in self.coordinator_fetches:\n d = defer.Deferred()\n self.coordinator_fetches[group][1].append(d)\n return d\n requestId = self._next_id()\n request = KafkaCodec.encode_consumermetadata_request(self.\n _clientIdBytes, requestId, group)\n\n def x(response_bytes):\n response = KafkaCodec.decode_consumermetadata_response(response_bytes)\n log.debug('%r: load_consumer_metadata_for_group(%r) -> %r', self,\n group, response)\n if response.error:\n raise BrokerResponseError.errnos.get(response.error, UnknownError)(\n response)\n bm = BrokerMetadata(response.node_id, response.host, response.port)\n self.consumer_group_to_brokers[group] = bm\n self._update_brokers([bm])\n return True\n\n def x(err):\n log.error('Failed to retrieve consumer metadata for group %r',\n group, exc_info=(err.type, err.value, err.getTracebackObject()))\n self.reset_consumer_group_metadata(group)\n raise ConsumerCoordinatorNotAvailableError(\n 'Coordinator for group {!r} not available'.format(group))\n\n def x(result):\n [_, ds] = self.coordinator_fetches.pop(group, None)\n for d in ds:\n d.callback(result)\n request_d = self._send_broker_unaware_request(requestId, request)\n d = defer.Deferred()\n self.coordinator_fetches[group] = request_d, [d]\n request_d.addCallback(_handleConsumerMetadataResponse)\n request_d.addErrback(_handleConsumerMetadataErr)\n request_d.addBoth(_propagate)\n return d\n", "role": "user" }, { "content": "load_consumer_metadata_for_group", "role": "assistant" } ]
def x(self, payloads=None, acks=1, timeout=DEFAULT_REPLICAS_ACK_MSECS, fail_on_error=True, callback=None): encoder = partial(KafkaCodec.encode_produce_request, acks=acks, timeout =timeout) if acks == 0: decoder = None else: decoder = KafkaCodec.decode_produce_response resps = yield self._send_broker_aware_request(payloads, encoder, decoder) returnValue(self._handle_responses(resps, fail_on_error, callback))
send_produce_request
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, payloads=None, acks=1, timeout=DEFAULT_REPLICAS_ACK_MSECS,\n fail_on_error=True, callback=None):\n encoder = partial(KafkaCodec.encode_produce_request, acks=acks, timeout\n =timeout)\n if acks == 0:\n decoder = None\n else:\n decoder = KafkaCodec.decode_produce_response\n resps = yield self._send_broker_aware_request(payloads, encoder, decoder)\n returnValue(self._handle_responses(resps, fail_on_error, callback))\n", "role": "user" }, { "content": "send_produce_request", "role": "assistant" } ]
def x(self, payloads=None, fail_on_error=True, callback=None, max_wait_time =DEFAULT_FETCH_SERVER_WAIT_MSECS, min_bytes=DEFAULT_FETCH_MIN_BYTES): if max_wait_time / 1000 > self.timeout - 0.1: raise ValueError( '%r: max_wait_time: %d must be less than client.timeout by at least 100 milliseconds.' , self, max_wait_time) encoder = partial(KafkaCodec.encode_fetch_request, max_wait_time= max_wait_time, min_bytes=min_bytes) resps = yield self._send_broker_aware_request(payloads, encoder, KafkaCodec.decode_fetch_response) returnValue(self._handle_responses(resps, fail_on_error, callback))
send_fetch_request
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, payloads=None, fail_on_error=True, callback=None, max_wait_time\n =DEFAULT_FETCH_SERVER_WAIT_MSECS, min_bytes=DEFAULT_FETCH_MIN_BYTES):\n if max_wait_time / 1000 > self.timeout - 0.1:\n raise ValueError(\n '%r: max_wait_time: %d must be less than client.timeout by at least 100 milliseconds.'\n , self, max_wait_time)\n encoder = partial(KafkaCodec.encode_fetch_request, max_wait_time=\n max_wait_time, min_bytes=min_bytes)\n resps = yield self._send_broker_aware_request(payloads, encoder,\n KafkaCodec.decode_fetch_response)\n returnValue(self._handle_responses(resps, fail_on_error, callback))\n", "role": "user" }, { "content": "send_fetch_request", "role": "assistant" } ]
def x(self, group, payloads=None, fail_on_error=True, callback=None): encoder = partial(KafkaCodec.encode_offset_fetch_request, group=group) decoder = KafkaCodec.decode_offset_fetch_response resps = yield self._send_broker_aware_request(payloads, encoder, decoder, consumer_group=group) returnValue(self._handle_responses(resps, fail_on_error, callback, group))
send_offset_fetch_request
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, group, payloads=None, fail_on_error=True, callback=None):\n encoder = partial(KafkaCodec.encode_offset_fetch_request, group=group)\n decoder = KafkaCodec.decode_offset_fetch_response\n resps = yield self._send_broker_aware_request(payloads, encoder,\n decoder, consumer_group=group)\n returnValue(self._handle_responses(resps, fail_on_error, callback, group))\n", "role": "user" }, { "content": "send_offset_fetch_request", "role": "assistant" } ]
def x(self, group, payloads=None, fail_on_error=True, callback=None, group_generation_id=-1, consumer_id=''): group = _coerce_consumer_group(group) encoder = partial(KafkaCodec.encode_offset_commit_request, group=group, group_generation_id=group_generation_id, consumer_id=consumer_id) decoder = KafkaCodec.decode_offset_commit_response resps = yield self._send_broker_aware_request(payloads, encoder, decoder, consumer_group=group) returnValue(self._handle_responses(resps, fail_on_error, callback, group))
send_offset_commit_request
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, group, payloads=None, fail_on_error=True, callback=None,\n group_generation_id=-1, consumer_id=''):\n group = _coerce_consumer_group(group)\n encoder = partial(KafkaCodec.encode_offset_commit_request, group=group,\n group_generation_id=group_generation_id, consumer_id=consumer_id)\n decoder = KafkaCodec.decode_offset_commit_response\n resps = yield self._send_broker_aware_request(payloads, encoder,\n decoder, consumer_group=group)\n returnValue(self._handle_responses(resps, fail_on_error, callback, group))\n", "role": "user" }, { "content": "send_offset_commit_request", "role": "assistant" } ]
def x(self, node_id): if self._closing: raise ClientError( 'Cannot get broker client for node_id={}: {} has been closed'. format(node_id, self)) if node_id not in self.clients: broker_metadata = self._brokers[node_id] log.debug('%r: creating client for %s', self, broker_metadata) self.clients[node_id] = _KafkaBrokerClient(self.reactor, self. _endpoint_factory, broker_metadata, self.clientId, self. _retry_policy) return self.clients[node_id]
get_brokerclient
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, node_id):\n if self._closing:\n raise ClientError(\n 'Cannot get broker client for node_id={}: {} has been closed'.\n format(node_id, self))\n if node_id not in self.clients:\n broker_metadata = self._brokers[node_id]\n log.debug('%r: creating client for %s', self, broker_metadata)\n self.clients[node_id] = _KafkaBrokerClient(self.reactor, self.\n _endpoint_factory, broker_metadata, self.clientId, self.\n _retry_policy)\n return self.clients[node_id]\n", "role": "user" }, { "content": "get_brokerclient", "role": "assistant" } ]
def x(self, clients): def x(failure, brokerclient): log.debug('BrokerClient: %s close result: %s: %s', brokerclient, failure.type.__name__, failure.getErrorMessage()) def x(result, close_dlist): if close_dlist == self.close_dlist: self.close_dlist = None if not self.close_dlist: dList = [] else: log.debug('%r: _close_brokerclients has nested deferredlist: %r', self, self.close_dlist) dList = [self.close_dlist] for brokerClient in clients: log.debug('Calling close on: %r', brokerClient) d = brokerClient.close().addErrback(_log_close_failure, brokerClient) dList.append(d) self.close_dlist = DeferredList(dList) self.close_dlist.addBoth(_clean_close_dlist, self.close_dlist)
close_brokerclients
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, clients):\n\n def x(failure, brokerclient):\n log.debug('BrokerClient: %s close result: %s: %s', brokerclient,\n failure.type.__name__, failure.getErrorMessage())\n\n def x(result, close_dlist):\n if close_dlist == self.close_dlist:\n self.close_dlist = None\n if not self.close_dlist:\n dList = []\n else:\n log.debug('%r: _close_brokerclients has nested deferredlist: %r',\n self, self.close_dlist)\n dList = [self.close_dlist]\n for brokerClient in clients:\n log.debug('Calling close on: %r', brokerClient)\n d = brokerClient.close().addErrback(_log_close_failure, brokerClient)\n dList.append(d)\n self.close_dlist = DeferredList(dList)\n self.close_dlist.addBoth(_clean_close_dlist, self.close_dlist)\n", "role": "user" }, { "content": "close_brokerclients", "role": "assistant" } ]
def x(self, brokers, remove=False): log.debug('%r: _update_brokers(%r, remove=%r)', self, brokers, remove) brokers_by_id = {bm.node_id: bm for bm in brokers} self._brokers.update(brokers_by_id) for node_id, broker_meta in brokers_by_id.items(): if node_id not in self.clients: continue self.clients[node_id].updateMetadata(broker_meta) if remove: to_close = [self.clients.pop(node_id) for node_id in set(self. clients) - set(brokers_by_id)] if to_close: self._close_brokerclients(to_close)
update_brokers
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, brokers, remove=False):\n log.debug('%r: _update_brokers(%r, remove=%r)', self, brokers, remove)\n brokers_by_id = {bm.node_id: bm for bm in brokers}\n self._brokers.update(brokers_by_id)\n for node_id, broker_meta in brokers_by_id.items():\n if node_id not in self.clients:\n continue\n self.clients[node_id].updateMetadata(broker_meta)\n if remove:\n to_close = [self.clients.pop(node_id) for node_id in set(self.\n clients) - set(brokers_by_id)]\n if to_close:\n self._close_brokerclients(to_close)\n", "role": "user" }, { "content": "update_brokers", "role": "assistant" } ]
def x(self, topic, partition): key = TopicAndPartition(topic, partition) if self.topics_to_brokers.get(key) is None: yield self.load_metadata_for_topics(topic) if key not in self.topics_to_brokers: raise PartitionUnavailableError('%s not available' % str(key)) returnValue(self.topics_to_brokers[key])
get_leader_for_partition
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, topic, partition):\n key = TopicAndPartition(topic, partition)\n if self.topics_to_brokers.get(key) is None:\n yield self.load_metadata_for_topics(topic)\n if key not in self.topics_to_brokers:\n raise PartitionUnavailableError('%s not available' % str(key))\n returnValue(self.topics_to_brokers[key])\n", "role": "user" }, { "content": "get_leader_for_partition", "role": "assistant" } ]
def x(self, consumer_group): if self.consumer_group_to_brokers.get(consumer_group) is None: yield self.load_consumer_metadata_for_group(consumer_group) returnValue(self.consumer_group_to_brokers.get(consumer_group))
get_coordinator_for_group
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, consumer_group):\n if self.consumer_group_to_brokers.get(consumer_group) is None:\n yield self.load_consumer_metadata_for_group(consumer_group)\n returnValue(self.consumer_group_to_brokers.get(consumer_group))\n", "role": "user" }, { "content": "get_coordinator_for_group", "role": "assistant" } ]
def x(self, broker, requestId, request, **kwArgs): def x(broker, requestId): try: broker.cancelRequest(requestId, reason=RequestTimedOutError( 'Request: {} cancelled due to timeout'.format(requestId))) except KeyError: log.exception( 'ERROR: Failed to find key for timed-out request. Broker: %r Req: %d' , broker, requestId) raise if self._disconnect_on_timeout: broker.disconnect() def x(timeout, start): now = self.reactor.seconds() if now >= start + timeout: log.warning('Reactor was starved for %r seconds', now - start) def x(result, dc): if dc.active(): dc.cancel() return result log.debug('_mrtb: sending request: %d to broker: %r', requestId, broker) d = broker.makeRequest(requestId, request, **kwArgs) dc = self.reactor.callLater(self.timeout, _timeout_request, broker, requestId) rc = self.reactor.callLater(self.timeout * 0.9, _alert_blocked_reactor, self.timeout, self.reactor.seconds()) d.addBoth(_cancel_timeout, dc) d.addBoth(_cancel_timeout, rc) return d
make_request_to_broker
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, broker, requestId, request, **kwArgs):\n\n def x(broker, requestId):\n try:\n broker.cancelRequest(requestId, reason=RequestTimedOutError(\n 'Request: {} cancelled due to timeout'.format(requestId)))\n except KeyError:\n log.exception(\n 'ERROR: Failed to find key for timed-out request. Broker: %r Req: %d'\n , broker, requestId)\n raise\n if self._disconnect_on_timeout:\n broker.disconnect()\n\n def x(timeout, start):\n now = self.reactor.seconds()\n if now >= start + timeout:\n log.warning('Reactor was starved for %r seconds', now - start)\n\n def x(result, dc):\n if dc.active():\n dc.cancel()\n return result\n log.debug('_mrtb: sending request: %d to broker: %r', requestId, broker)\n d = broker.makeRequest(requestId, request, **kwArgs)\n dc = self.reactor.callLater(self.timeout, _timeout_request, broker,\n requestId)\n rc = self.reactor.callLater(self.timeout * 0.9, _alert_blocked_reactor,\n self.timeout, self.reactor.seconds())\n d.addBoth(_cancel_timeout, dc)\n d.addBoth(_cancel_timeout, rc)\n return d\n", "role": "user" }, { "content": "make_request_to_broker", "role": "assistant" } ]
def x(self, requestId, request): node_ids = list(self._brokers.keys()) random.shuffle(node_ids) def x(node_id): try: return self.clients[node_id].connected() except KeyError: return False node_ids.sort(reverse=True, key=connected) for node_id in node_ids: broker = self._get_brokerclient(node_id) try: log.debug('_sbur: sending request %d to broker %r', requestId, broker) d = self._make_request_to_broker(broker, requestId, request) resp = yield d returnValue(resp) except KafkaError as e: log.warning( 'Will try next server after request with correlationId=%d failed against server %s:%i. Error: %s' , requestId, broker.host, broker.port, e) returnValue((yield self._send_bootstrap_request(request)))
send_broker_unaware_request
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, requestId, request):\n node_ids = list(self._brokers.keys())\n random.shuffle(node_ids)\n\n def x(node_id):\n try:\n return self.clients[node_id].connected()\n except KeyError:\n return False\n node_ids.sort(reverse=True, key=connected)\n for node_id in node_ids:\n broker = self._get_brokerclient(node_id)\n try:\n log.debug('_sbur: sending request %d to broker %r', requestId,\n broker)\n d = self._make_request_to_broker(broker, requestId, request)\n resp = yield d\n returnValue(resp)\n except KafkaError as e:\n log.warning(\n 'Will try next server after request with correlationId=%d failed against server %s:%i. Error: %s'\n , requestId, broker.host, broker.port, e)\n returnValue((yield self._send_bootstrap_request(request)))\n", "role": "user" }, { "content": "send_broker_unaware_request", "role": "assistant" } ]
def x(self, request): hostports = list(self._bootstrap_hosts) random.shuffle(hostports) for host, port in hostports: ep = self._endpoint_factory(self.reactor, host, port) try: protocol = yield ep.connect(_bootstrapFactory) except Exception as e: log.debug('%s: bootstrap connect to %s:%s -> %s', self, host, port, e) continue try: response = yield protocol.request(request).addTimeout(self. timeout, self.reactor) except Exception: log.debug('%s: bootstrap request to %s:%s failed', self, host, port, exc_info=True) else: returnValue(response) finally: protocol.transport.loseConnection() raise KafkaUnavailableError('Failed to bootstrap from hosts {}'.format( hostports))
send_bootstrap_request
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, request):\n hostports = list(self._bootstrap_hosts)\n random.shuffle(hostports)\n for host, port in hostports:\n ep = self._endpoint_factory(self.reactor, host, port)\n try:\n protocol = yield ep.connect(_bootstrapFactory)\n except Exception as e:\n log.debug('%s: bootstrap connect to %s:%s -> %s', self, host,\n port, e)\n continue\n try:\n response = yield protocol.request(request).addTimeout(self.\n timeout, self.reactor)\n except Exception:\n log.debug('%s: bootstrap request to %s:%s failed', self, host,\n port, exc_info=True)\n else:\n returnValue(response)\n finally:\n protocol.transport.loseConnection()\n raise KafkaUnavailableError('Failed to bootstrap from hosts {}'.format(\n hostports))\n", "role": "user" }, { "content": "send_bootstrap_request", "role": "assistant" } ]
def x(self, payloads, encoder_fn, decode_fn, consumer_group=None): if not payloads: raise ValueError('Payloads parameter is empty') original_keys = [] payloads_by_broker = collections.defaultdict(list) for payload in payloads: if consumer_group is None: leader = yield self._get_leader_for_partition(payload.topic, payload.partition) if leader is None: raise LeaderUnavailableError( 'Leader not available for topic %s partition %s' % ( payload.topic, payload.partition)) else: leader = yield self._get_coordinator_for_group(consumer_group) if leader is None: raise ConsumerCoordinatorNotAvailableError( 'Coordinator not available for group: %s' % consumer_group) payloads_by_broker[leader].append(payload) original_keys.append((payload.topic, payload.partition)) acc = {} expectResponse = decode_fn is not None failed_payloads = [] inFlight = [] payloadsList = [] for broker_meta, payloads in payloads_by_broker.items(): broker = self._get_brokerclient(broker_meta.node_id) requestId = self._next_id() request = encoder_fn(client_id=self._clientIdBytes, correlation_id= requestId, payloads=payloads) d = self._make_request_to_broker(broker, requestId, request, expectResponse=expectResponse) inFlight.append(d) payloadsList.append(payloads) results = yield DeferredList(inFlight, consumeErrors=True) for (success, response), payloads in zip(results, payloadsList): if not success: log.debug('%r: request:%r to broker failed: %r', self, payloads, response) failed_payloads.extend([(p, response) for p in payloads]) continue if not expectResponse: continue for response in decode_fn(response): acc[response.topic, response.partition] = response responses = [acc[k] for k in original_keys if k in acc] if acc else [] if failed_payloads: self.reset_all_metadata() raise FailedPayloadsError(responses, failed_payloads) returnValue(responses)
send_broker_aware_request
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, payloads, encoder_fn, decode_fn, consumer_group=None):\n if not payloads:\n raise ValueError('Payloads parameter is empty')\n original_keys = []\n payloads_by_broker = collections.defaultdict(list)\n for payload in payloads:\n if consumer_group is None:\n leader = yield self._get_leader_for_partition(payload.topic,\n payload.partition)\n if leader is None:\n raise LeaderUnavailableError(\n 'Leader not available for topic %s partition %s' % (\n payload.topic, payload.partition))\n else:\n leader = yield self._get_coordinator_for_group(consumer_group)\n if leader is None:\n raise ConsumerCoordinatorNotAvailableError(\n 'Coordinator not available for group: %s' % consumer_group)\n payloads_by_broker[leader].append(payload)\n original_keys.append((payload.topic, payload.partition))\n acc = {}\n expectResponse = decode_fn is not None\n failed_payloads = []\n inFlight = []\n payloadsList = []\n for broker_meta, payloads in payloads_by_broker.items():\n broker = self._get_brokerclient(broker_meta.node_id)\n requestId = self._next_id()\n request = encoder_fn(client_id=self._clientIdBytes, correlation_id=\n requestId, payloads=payloads)\n d = self._make_request_to_broker(broker, requestId, request,\n expectResponse=expectResponse)\n inFlight.append(d)\n payloadsList.append(payloads)\n results = yield DeferredList(inFlight, consumeErrors=True)\n for (success, response), payloads in zip(results, payloadsList):\n if not success:\n log.debug('%r: request:%r to broker failed: %r', self, payloads,\n response)\n failed_payloads.extend([(p, response) for p in payloads])\n continue\n if not expectResponse:\n continue\n for response in decode_fn(response):\n acc[response.topic, response.partition] = response\n responses = [acc[k] for k in original_keys if k in acc] if acc else []\n if failed_payloads:\n self.reset_all_metadata()\n raise FailedPayloadsError(responses, failed_payloads)\n returnValue(responses)\n", "role": "user" }, { "content": "send_broker_aware_request", "role": "assistant" } ]
def x(expression, options=None): descripter = ExpressionDescriptor(expression, options) return descripter.get_description(DescriptionTypeEnum.FULL)
get_description
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(expression, options=None):\n descripter = ExpressionDescriptor(expression, options)\n return descripter.get_description(DescriptionTypeEnum.FULL)\n", "role": "user" }, { "content": "get_description", "role": "assistant" } ]
def x(self, description_type=DescriptionTypeEnum.FULL): try: if self._parsed is False: parser = ExpressionParser(self._expression, self._options) self._expression_parts = parser.parse() self._parsed = True choices = {DescriptionTypeEnum.FULL: self.get_full_description, DescriptionTypeEnum.TIMEOFDAY: self.get_time_of_day_description, DescriptionTypeEnum.HOURS: self.get_hours_description, DescriptionTypeEnum.MINUTES: self.get_minutes_description, DescriptionTypeEnum.SECONDS: self.get_seconds_description, DescriptionTypeEnum.DAYOFMONTH: self. get_day_of_month_description, DescriptionTypeEnum.MONTH: self. get_month_description, DescriptionTypeEnum.DAYOFWEEK: self. get_day_of_week_description, DescriptionTypeEnum.YEAR: self. get_year_description} description = choices.get(description_type, self. get_seconds_description)() except Exception as ex: if self._options.throw_exception_on_parse_error: raise else: description = str(ex) return description
get_description
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, description_type=DescriptionTypeEnum.FULL):\n try:\n if self._parsed is False:\n parser = ExpressionParser(self._expression, self._options)\n self._expression_parts = parser.parse()\n self._parsed = True\n choices = {DescriptionTypeEnum.FULL: self.get_full_description,\n DescriptionTypeEnum.TIMEOFDAY: self.get_time_of_day_description,\n DescriptionTypeEnum.HOURS: self.get_hours_description,\n DescriptionTypeEnum.MINUTES: self.get_minutes_description,\n DescriptionTypeEnum.SECONDS: self.get_seconds_description,\n DescriptionTypeEnum.DAYOFMONTH: self.\n get_day_of_month_description, DescriptionTypeEnum.MONTH: self.\n get_month_description, DescriptionTypeEnum.DAYOFWEEK: self.\n get_day_of_week_description, DescriptionTypeEnum.YEAR: self.\n get_year_description}\n description = choices.get(description_type, self.\n get_seconds_description)()\n except Exception as ex:\n if self._options.throw_exception_on_parse_error:\n raise\n else:\n description = str(ex)\n return description\n", "role": "user" }, { "content": "get_description", "role": "assistant" } ]
def x(self): try: time_segment = self.get_time_of_day_description() day_of_month_desc = self.get_day_of_month_description() month_desc = self.get_month_description() day_of_week_desc = self.get_day_of_week_description() year_desc = self.get_year_description() description = '{0}{1}{2}{3}{4}'.format(time_segment, day_of_month_desc, day_of_week_desc, month_desc, year_desc) description = self.transform_verbosity(description, self._options. verbose) description = self.transform_case(description, self._options. casing_type) except Exception: description = _( 'An error occured when generating the expression description. Check the cron expression syntax.' ) if self._options.throw_exception_on_parse_error: raise FormatException(description) return description
get_full_description
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n try:\n time_segment = self.get_time_of_day_description()\n day_of_month_desc = self.get_day_of_month_description()\n month_desc = self.get_month_description()\n day_of_week_desc = self.get_day_of_week_description()\n year_desc = self.get_year_description()\n description = '{0}{1}{2}{3}{4}'.format(time_segment,\n day_of_month_desc, day_of_week_desc, month_desc, year_desc)\n description = self.transform_verbosity(description, self._options.\n verbose)\n description = self.transform_case(description, self._options.\n casing_type)\n except Exception:\n description = _(\n 'An error occured when generating the expression description. Check the cron expression syntax.'\n )\n if self._options.throw_exception_on_parse_error:\n raise FormatException(description)\n return description\n", "role": "user" }, { "content": "get_full_description", "role": "assistant" } ]
def x(self): seconds_expression = self._expression_parts[0] minute_expression = self._expression_parts[1] hour_expression = self._expression_parts[2] description = StringBuilder() if any(exp in minute_expression for exp in self._special_characters ) is False and any(exp in hour_expression for exp in self. _special_characters) is False and any(exp in seconds_expression for exp in self._special_characters) is False: description.append(_('At ')) description.append(self.format_time(hour_expression, minute_expression, seconds_expression)) elif '-' in minute_expression and ',' not in minute_expression and any( exp in hour_expression for exp in self._special_characters) is False: minute_parts = minute_expression.split('-') description.append(_('Every minute between {0} and {1}').format( self.format_time(hour_expression, minute_parts[0]), self. format_time(hour_expression, minute_parts[1]))) elif ',' in hour_expression and '-' not in hour_expression and any(exp in minute_expression for exp in self._special_characters) is False: hour_parts = hour_expression.split(',') description.append(_('At')) for i, hour_part in enumerate(hour_parts): description.append(' ') description.append(self.format_time(hour_part, minute_expression)) if i < len(hour_parts) - 2: description.append(',') if i == len(hour_parts) - 2: description.append(_(' and')) else: seconds_description = self.get_seconds_description() minutes_description = self.get_minutes_description() hours_description = self.get_hours_description() description.append(seconds_description) if description: description.append(', ') description.append(minutes_description) if description: description.append(', ') description.append(hours_description) return str(description)
get_time_of_day_description
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n seconds_expression = self._expression_parts[0]\n minute_expression = self._expression_parts[1]\n hour_expression = self._expression_parts[2]\n description = StringBuilder()\n if any(exp in minute_expression for exp in self._special_characters\n ) is False and any(exp in hour_expression for exp in self.\n _special_characters) is False and any(exp in seconds_expression for\n exp in self._special_characters) is False:\n description.append(_('At '))\n description.append(self.format_time(hour_expression,\n minute_expression, seconds_expression))\n elif '-' in minute_expression and ',' not in minute_expression and any(\n exp in hour_expression for exp in self._special_characters) is False:\n minute_parts = minute_expression.split('-')\n description.append(_('Every minute between {0} and {1}').format(\n self.format_time(hour_expression, minute_parts[0]), self.\n format_time(hour_expression, minute_parts[1])))\n elif ',' in hour_expression and '-' not in hour_expression and any(exp in\n minute_expression for exp in self._special_characters) is False:\n hour_parts = hour_expression.split(',')\n description.append(_('At'))\n for i, hour_part in enumerate(hour_parts):\n description.append(' ')\n description.append(self.format_time(hour_part, minute_expression))\n if i < len(hour_parts) - 2:\n description.append(',')\n if i == len(hour_parts) - 2:\n description.append(_(' and'))\n else:\n seconds_description = self.get_seconds_description()\n minutes_description = self.get_minutes_description()\n hours_description = self.get_hours_description()\n description.append(seconds_description)\n if description:\n description.append(', ')\n description.append(minutes_description)\n if description:\n description.append(', ')\n description.append(hours_description)\n return str(description)\n", "role": "user" }, { "content": "get_time_of_day_description", "role": "assistant" } ]
def x(self): return self.get_segment_description(self._expression_parts[0], _( 'every second'), lambda s: s, lambda s: _('every {0} seconds'). format(s), lambda s: _('seconds {0} through {1} past the minute'), lambda s: _('at {0} seconds past the minute'))
get_seconds_description
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return self.get_segment_description(self._expression_parts[0], _(\n 'every second'), lambda s: s, lambda s: _('every {0} seconds').\n format(s), lambda s: _('seconds {0} through {1} past the minute'), \n lambda s: _('at {0} seconds past the minute'))\n", "role": "user" }, { "content": "get_seconds_description", "role": "assistant" } ]
def x(self): return self.get_segment_description(self._expression_parts[1], _( 'every minute'), lambda s: s, lambda s: _('every {0} minutes'). format(s), lambda s: _('minutes {0} through {1} past the hour'), lambda s: '' if s == '0' else _('at {0} minutes past the hour'))
get_minutes_description
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return self.get_segment_description(self._expression_parts[1], _(\n 'every minute'), lambda s: s, lambda s: _('every {0} minutes').\n format(s), lambda s: _('minutes {0} through {1} past the hour'), lambda\n s: '' if s == '0' else _('at {0} minutes past the hour'))\n", "role": "user" }, { "content": "get_minutes_description", "role": "assistant" } ]
def x(self): expression = self._expression_parts[2] return self.get_segment_description(expression, _('every hour'), lambda s: self.format_time(s, '0'), lambda s: _('every {0} hours').format( s), lambda s: _('between {0} and {1}'), lambda s: _('at {0}'))
get_hours_description
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n expression = self._expression_parts[2]\n return self.get_segment_description(expression, _('every hour'), lambda\n s: self.format_time(s, '0'), lambda s: _('every {0} hours').format(\n s), lambda s: _('between {0} and {1}'), lambda s: _('at {0}'))\n", "role": "user" }, { "content": "get_hours_description", "role": "assistant" } ]
def x(self): if self._expression_parts[5] == '*' and self._expression_parts[3] != '*': return '' def x(s): exp = s if '#' in s: exp, useless = s.split('#', 2) elif 'L' in s: exp = exp.replace('L', '') return self.number_to_day(int(exp)) def x(s): if '#' in s: day_of_week_of_month = s[s.find('#') + 1:] try: day_of_week_of_month_number = int(day_of_week_of_month) choices = {(1): _('first'), (2): _('second'), (3): _( 'third'), (4): _('forth'), (5): _('fifth')} day_of_week_of_month_description = choices.get( day_of_week_of_month_number, '') except ValueError: day_of_week_of_month_description = '' formated = '{}{}{}'.format(_(', on the '), day_of_week_of_month_description, _(' {0} of the month')) elif 'L' in s: formated = _(', on the last {0} of the month') else: formated = _(', only on {0}') return formated return self.get_segment_description(self._expression_parts[5], _( ', every day'), lambda s: get_day_name(s), lambda s: _( ', every {0} days of the week').format(s), lambda s: _( ', {0} through {1}'), lambda s: get_format(s))
get_day_of_week_description
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self._expression_parts[5] == '*' and self._expression_parts[3] != '*':\n return ''\n\n def x(s):\n exp = s\n if '#' in s:\n exp, useless = s.split('#', 2)\n elif 'L' in s:\n exp = exp.replace('L', '')\n return self.number_to_day(int(exp))\n\n def x(s):\n if '#' in s:\n day_of_week_of_month = s[s.find('#') + 1:]\n try:\n day_of_week_of_month_number = int(day_of_week_of_month)\n choices = {(1): _('first'), (2): _('second'), (3): _(\n 'third'), (4): _('forth'), (5): _('fifth')}\n day_of_week_of_month_description = choices.get(\n day_of_week_of_month_number, '')\n except ValueError:\n day_of_week_of_month_description = ''\n formated = '{}{}{}'.format(_(', on the '),\n day_of_week_of_month_description, _(' {0} of the month'))\n elif 'L' in s:\n formated = _(', on the last {0} of the month')\n else:\n formated = _(', only on {0}')\n return formated\n return self.get_segment_description(self._expression_parts[5], _(\n ', every day'), lambda s: get_day_name(s), lambda s: _(\n ', every {0} days of the week').format(s), lambda s: _(\n ', {0} through {1}'), lambda s: get_format(s))\n", "role": "user" }, { "content": "get_day_of_week_description", "role": "assistant" } ]
def x(self): return self.get_segment_description(self._expression_parts[4], '', lambda s: datetime.date(datetime.date.today().year, int(s), 1).strftime( '%B'), lambda s: _(', every {0} months').format(s), lambda s: _( ', {0} through {1}'), lambda s: _(', only in {0}'))
get_month_description
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return self.get_segment_description(self._expression_parts[4], '', lambda\n s: datetime.date(datetime.date.today().year, int(s), 1).strftime(\n '%B'), lambda s: _(', every {0} months').format(s), lambda s: _(\n ', {0} through {1}'), lambda s: _(', only in {0}'))\n", "role": "user" }, { "content": "get_month_description", "role": "assistant" } ]
def x(self): expression = self._expression_parts[3] expression = expression.replace('?', '*') if expression == 'L': description = _(', on the last day of the month') elif expression == 'LW' or expression == 'WL': description = _(', on the last weekday of the month') else: regex = re.compile('(\\d{1,2}W)|(W\\d{1,2})') if regex.match(expression): m = regex.match(expression) day_number = int(m.group().replace('W', '')) day_string = _('first weekday') if day_number == 1 else _( 'weekday nearest day {0}').format(day_number) description = _(', on the {0} of the month').format(day_string) else: description = self.get_segment_description(expression, _( ', every day'), lambda s: s, lambda s: _(', every day') if s == '1' else _(', every {0} days'), lambda s: _( ', between day {0} and {1} of the month'), lambda s: _( ', on day {0} of the month')) return description
get_day_of_month_description
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n expression = self._expression_parts[3]\n expression = expression.replace('?', '*')\n if expression == 'L':\n description = _(', on the last day of the month')\n elif expression == 'LW' or expression == 'WL':\n description = _(', on the last weekday of the month')\n else:\n regex = re.compile('(\\\\d{1,2}W)|(W\\\\d{1,2})')\n if regex.match(expression):\n m = regex.match(expression)\n day_number = int(m.group().replace('W', ''))\n day_string = _('first weekday') if day_number == 1 else _(\n 'weekday nearest day {0}').format(day_number)\n description = _(', on the {0} of the month').format(day_string)\n else:\n description = self.get_segment_description(expression, _(\n ', every day'), lambda s: s, lambda s: _(', every day') if \n s == '1' else _(', every {0} days'), lambda s: _(\n ', between day {0} and {1} of the month'), lambda s: _(\n ', on day {0} of the month'))\n return description\n", "role": "user" }, { "content": "get_day_of_month_description", "role": "assistant" } ]
def x(self): def x(s): regex = re.compile('^\\d+$') if regex.match(s): year_int = int(s) if year_int < 1900: return year_int return datetime.date(year_int, 1, 1).strftime('%Y') else: return s return self.get_segment_description(self._expression_parts[6], '', lambda s: format_year(s), lambda s: _(', every {0} years').format(s), lambda s: _(', {0} through {1}'), lambda s: _(', only in {0}'))
get_year_description
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n\n def x(s):\n regex = re.compile('^\\\\d+$')\n if regex.match(s):\n year_int = int(s)\n if year_int < 1900:\n return year_int\n return datetime.date(year_int, 1, 1).strftime('%Y')\n else:\n return s\n return self.get_segment_description(self._expression_parts[6], '', lambda\n s: format_year(s), lambda s: _(', every {0} years').format(s), lambda\n s: _(', {0} through {1}'), lambda s: _(', only in {0}'))\n", "role": "user" }, { "content": "get_year_description", "role": "assistant" } ]
def x(self, expression, all_description, get_single_item_description, get_interval_description_format, get_between_description_format, get_description_format): description = None if expression is None or expression == '': description = '' elif expression == '*': description = all_description elif any(ext in expression for ext in ['/', '-', ',']) is False: description = get_description_format(expression).format( get_single_item_description(expression)) elif '/' in expression: segments = expression.split('/') description = get_interval_description_format(segments[1]).format( get_single_item_description(segments[1])) if '-' in segments[0]: between_segment_description = (self. generate_between_segment_description(segments[0], get_between_description_format, get_single_item_description)) if not between_segment_description.startswith(', '): description += ', ' description += between_segment_description elif any(ext in segments[0] for ext in ['*', ',']) is False: range_item_description = get_description_format(segments[0] ).format(get_single_item_description(segments[0])) range_item_description = range_item_description.replace(', ', '') description += _(', starting {0}').format(range_item_description) elif ',' in expression: segments = expression.split(',') description_content = '' for i, segment in enumerate(segments): if i > 0 and len(segments) > 2: description_content += ',' if i < len(segments) - 1: description_content += ' ' if i > 0 and len(segments) > 1 and (i == len(segments) - 1 or len(segments) == 2): description_content += _(' and ') if '-' in segment: between_description = (self. generate_between_segment_description(segment, lambda s: _(', {0} through {1}'), get_single_item_description)) between_description = between_description.replace(', ', '') description_content += between_description else: description_content += get_single_item_description(segment) description = get_description_format(expression).format( description_content) elif '-' in expression: description = self.generate_between_segment_description(expression, get_between_description_format, get_single_item_description) return description
get_segment_description
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, expression, all_description, get_single_item_description,\n get_interval_description_format, get_between_description_format,\n get_description_format):\n description = None\n if expression is None or expression == '':\n description = ''\n elif expression == '*':\n description = all_description\n elif any(ext in expression for ext in ['/', '-', ',']) is False:\n description = get_description_format(expression).format(\n get_single_item_description(expression))\n elif '/' in expression:\n segments = expression.split('/')\n description = get_interval_description_format(segments[1]).format(\n get_single_item_description(segments[1]))\n if '-' in segments[0]:\n between_segment_description = (self.\n generate_between_segment_description(segments[0],\n get_between_description_format, get_single_item_description))\n if not between_segment_description.startswith(', '):\n description += ', '\n description += between_segment_description\n elif any(ext in segments[0] for ext in ['*', ',']) is False:\n range_item_description = get_description_format(segments[0]\n ).format(get_single_item_description(segments[0]))\n range_item_description = range_item_description.replace(', ', '')\n description += _(', starting {0}').format(range_item_description)\n elif ',' in expression:\n segments = expression.split(',')\n description_content = ''\n for i, segment in enumerate(segments):\n if i > 0 and len(segments) > 2:\n description_content += ','\n if i < len(segments) - 1:\n description_content += ' '\n if i > 0 and len(segments) > 1 and (i == len(segments) - 1 or \n len(segments) == 2):\n description_content += _(' and ')\n if '-' in segment:\n between_description = (self.\n generate_between_segment_description(segment, lambda s:\n _(', {0} through {1}'), get_single_item_description))\n between_description = between_description.replace(', ', '')\n description_content += between_description\n else:\n description_content += get_single_item_description(segment)\n description = get_description_format(expression).format(\n description_content)\n elif '-' in expression:\n description = self.generate_between_segment_description(expression,\n get_between_description_format, get_single_item_description)\n return description\n", "role": "user" }, { "content": "get_segment_description", "role": "assistant" } ]
def x(self, between_expression, get_between_description_format, get_single_item_description): description = '' between_segments = between_expression.split('-') between_segment_1_description = get_single_item_description( between_segments[0]) between_segment_2_description = get_single_item_description( between_segments[1]) between_segment_2_description = between_segment_2_description.replace(':00' , ':59') between_description_format = get_between_description_format( between_expression) description += between_description_format.format( between_segment_1_description, between_segment_2_description) return description
generate_between_segment_description
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, between_expression, get_between_description_format,\n get_single_item_description):\n description = ''\n between_segments = between_expression.split('-')\n between_segment_1_description = get_single_item_description(\n between_segments[0])\n between_segment_2_description = get_single_item_description(\n between_segments[1])\n between_segment_2_description = between_segment_2_description.replace(':00'\n , ':59')\n between_description_format = get_between_description_format(\n between_expression)\n description += between_description_format.format(\n between_segment_1_description, between_segment_2_description)\n return description\n", "role": "user" }, { "content": "generate_between_segment_description", "role": "assistant" } ]
def x(self, hour_expression, minute_expression, second_expression=''): hour = int(hour_expression) period = '' if self._options.use_24hour_time_format is False: period = ' PM' if hour >= 12 else ' AM' if hour > 12: hour -= 12 minute = str(int(minute_expression)) second = '' if second_expression is not None and second_expression: second = '{}{}'.format(':', str(int(second_expression)).zfill(2)) return '{0}:{1}{2}{3}'.format(str(hour).zfill(2), minute.zfill(2), second, period)
format_time
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, hour_expression, minute_expression, second_expression=''):\n hour = int(hour_expression)\n period = ''\n if self._options.use_24hour_time_format is False:\n period = ' PM' if hour >= 12 else ' AM'\n if hour > 12:\n hour -= 12\n minute = str(int(minute_expression))\n second = ''\n if second_expression is not None and second_expression:\n second = '{}{}'.format(':', str(int(second_expression)).zfill(2))\n return '{0}:{1}{2}{3}'.format(str(hour).zfill(2), minute.zfill(2),\n second, period)\n", "role": "user" }, { "content": "format_time", "role": "assistant" } ]
def x(self, description, use_verbose_format): if use_verbose_format is False: description = description.replace(_(', every minute'), '') description = description.replace(_(', every hour'), '') description = description.replace(_(', every day'), '') return description
transform_verbosity
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, description, use_verbose_format):\n if use_verbose_format is False:\n description = description.replace(_(', every minute'), '')\n description = description.replace(_(', every hour'), '')\n description = description.replace(_(', every day'), '')\n return description\n", "role": "user" }, { "content": "transform_verbosity", "role": "assistant" } ]
def x(self, description, case_type): if case_type == CasingTypeEnum.Sentence: description = '{}{}'.format(description[0].upper(), description[1:]) elif case_type == CasingTypeEnum.Title: description = description.title() else: description = description.lower() return description
transform_case
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, description, case_type):\n if case_type == CasingTypeEnum.Sentence:\n description = '{}{}'.format(description[0].upper(), description[1:])\n elif case_type == CasingTypeEnum.Title:\n description = description.title()\n else:\n description = description.lower()\n return description\n", "role": "user" }, { "content": "transform_case", "role": "assistant" } ]
def x(self, day_number): return [calendar.day_name[6], calendar.day_name[0], calendar.day_name[1 ], calendar.day_name[2], calendar.day_name[3], calendar.day_name[4], calendar.day_name[5]][day_number]
number_to_day
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, day_number):\n return [calendar.day_name[6], calendar.day_name[0], calendar.day_name[1\n ], calendar.day_name[2], calendar.day_name[3], calendar.day_name[4],\n calendar.day_name[5]][day_number]\n", "role": "user" }, { "content": "number_to_day", "role": "assistant" } ]
def x(self, start_offset): if self._start_d is not None: raise RestartError('Start called on already-started consumer') self._state = '[started]' start_d = self._start_d = Deferred() self._fetch_offset = start_offset self._do_fetch() if self.consumer_group and self.auto_commit_every_s: self._commit_looper = LoopingCall(self._auto_commit) self._commit_looper.clock = self.client.reactor self._commit_looper_d = self._commit_looper.start(self. auto_commit_every_s, now=False) self._commit_looper_d.addCallbacks(self._commit_timer_stopped, self ._commit_timer_failed) return start_d
start
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, start_offset):\n if self._start_d is not None:\n raise RestartError('Start called on already-started consumer')\n self._state = '[started]'\n start_d = self._start_d = Deferred()\n self._fetch_offset = start_offset\n self._do_fetch()\n if self.consumer_group and self.auto_commit_every_s:\n self._commit_looper = LoopingCall(self._auto_commit)\n self._commit_looper.clock = self.client.reactor\n self._commit_looper_d = self._commit_looper.start(self.\n auto_commit_every_s, now=False)\n self._commit_looper_d.addCallbacks(self._commit_timer_stopped, self\n ._commit_timer_failed)\n return start_d\n", "role": "user" }, { "content": "start", "role": "assistant" } ]
def x(self): def x(result): self._shutdown_d, d = None, self._shutdown_d self.stop() self._shuttingdown = False d.callback((self._last_processed_offset, self._last_committed_offset)) def x(failure): if failure.check(OperationInProgress): failure.value.deferred.addCallback(_commit_and_stop) return self._shutdown_d, d = None, self._shutdown_d self.stop() self._shuttingdown = False d.errback(failure) def x(result): if not self.consumer_group: return _handle_shutdown_commit_success(None) self.commit().addCallbacks(_handle_shutdown_commit_success, _handle_shutdown_commit_failure) if self._start_d is None: return fail(Failure(RestopError( 'Shutdown called on non-running consumer'))) if self._shutdown_d: return fail(Failure(RestopError('Shutdown called more than once.'))) self._shuttingdown = True self._state = '[shutting down]' self._shutdown_d = d = Deferred() if self._processor_d: self._processor_d.addCallback(_commit_and_stop) else: _commit_and_stop(None) return d
shutdown
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n\n def x(result):\n self._shutdown_d, d = None, self._shutdown_d\n self.stop()\n self._shuttingdown = False\n d.callback((self._last_processed_offset, self._last_committed_offset))\n\n def x(failure):\n if failure.check(OperationInProgress):\n failure.value.deferred.addCallback(_commit_and_stop)\n return\n self._shutdown_d, d = None, self._shutdown_d\n self.stop()\n self._shuttingdown = False\n d.errback(failure)\n\n def x(result):\n if not self.consumer_group:\n return _handle_shutdown_commit_success(None)\n self.commit().addCallbacks(_handle_shutdown_commit_success,\n _handle_shutdown_commit_failure)\n if self._start_d is None:\n return fail(Failure(RestopError(\n 'Shutdown called on non-running consumer')))\n if self._shutdown_d:\n return fail(Failure(RestopError('Shutdown called more than once.')))\n self._shuttingdown = True\n self._state = '[shutting down]'\n self._shutdown_d = d = Deferred()\n if self._processor_d:\n self._processor_d.addCallback(_commit_and_stop)\n else:\n _commit_and_stop(None)\n return d\n", "role": "user" }, { "content": "shutdown", "role": "assistant" } ]
def x(self): if self._start_d is None: raise RestopError('Stop called on non-running consumer') self._stopping = True self._state = '[stopping]' if self._request_d: self._request_d.cancel() if self._msg_block_d: _msg_block_d, self._msg_block_d = self._msg_block_d, None _msg_block_d.addErrback(lambda fail: fail.trap(CancelledError)) _msg_block_d.cancel() if self._processor_d: self._processor_d.cancel() if self._retry_call: self._retry_call.cancel() if self._commit_ds: while self._commit_ds: d = self._commit_ds.pop() d.cancel() if self._commit_req: self._commit_req.cancel() if self._commit_call: self._commit_call.cancel() if self._commit_looper is not None: self._commit_looper.stop() self._stopping = False self._state = '[stopped]' self._start_d, d = None, self._start_d if not d.called: d.callback((self._last_processed_offset, self._last_committed_offset)) return self._last_processed_offset
stop
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self._start_d is None:\n raise RestopError('Stop called on non-running consumer')\n self._stopping = True\n self._state = '[stopping]'\n if self._request_d:\n self._request_d.cancel()\n if self._msg_block_d:\n _msg_block_d, self._msg_block_d = self._msg_block_d, None\n _msg_block_d.addErrback(lambda fail: fail.trap(CancelledError))\n _msg_block_d.cancel()\n if self._processor_d:\n self._processor_d.cancel()\n if self._retry_call:\n self._retry_call.cancel()\n if self._commit_ds:\n while self._commit_ds:\n d = self._commit_ds.pop()\n d.cancel()\n if self._commit_req:\n self._commit_req.cancel()\n if self._commit_call:\n self._commit_call.cancel()\n if self._commit_looper is not None:\n self._commit_looper.stop()\n self._stopping = False\n self._state = '[stopped]'\n self._start_d, d = None, self._start_d\n if not d.called:\n d.callback((self._last_processed_offset, self._last_committed_offset))\n return self._last_processed_offset\n", "role": "user" }, { "content": "stop", "role": "assistant" } ]
def x(self): if not self.consumer_group: return fail(Failure(InvalidConsumerGroupError('Bad Group_id:{0!r}'. format(self.consumer_group)))) if (self._last_processed_offset is None or self._last_processed_offset == self._last_committed_offset): return succeed(self._last_committed_offset) if self._commit_ds: d = Deferred() self._commit_ds.append(d) return fail(OperationInProgress(d)) d = Deferred() self._commit_ds.append(d) self._send_commit_request() if self._commit_looper is not None: self._commit_looper.reset() return d
commit
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if not self.consumer_group:\n return fail(Failure(InvalidConsumerGroupError('Bad Group_id:{0!r}'.\n format(self.consumer_group))))\n if (self._last_processed_offset is None or self._last_processed_offset ==\n self._last_committed_offset):\n return succeed(self._last_committed_offset)\n if self._commit_ds:\n d = Deferred()\n self._commit_ds.append(d)\n return fail(OperationInProgress(d))\n d = Deferred()\n self._commit_ds.append(d)\n self._send_commit_request()\n if self._commit_looper is not None:\n self._commit_looper.reset()\n return d\n", "role": "user" }, { "content": "commit", "role": "assistant" } ]
def x(self, by_count=False): if (self._stopping or self._shuttingdown or not self._start_d or self. _last_processed_offset is None or not self.consumer_group or by_count and not self.auto_commit_every_n): return if (not by_count or self._last_committed_offset is None or self. _last_processed_offset - self._last_committed_offset >= self. auto_commit_every_n): if not self._commit_ds: commit_d = self.commit() commit_d.addErrback(self._handle_auto_commit_error) else: d = Deferred() d.addCallback(self._retry_auto_commit, by_count) self._commit_ds.append(d)
auto_commit
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, by_count=False):\n if (self._stopping or self._shuttingdown or not self._start_d or self.\n _last_processed_offset is None or not self.consumer_group or \n by_count and not self.auto_commit_every_n):\n return\n if (not by_count or self._last_committed_offset is None or self.\n _last_processed_offset - self._last_committed_offset >= self.\n auto_commit_every_n):\n if not self._commit_ds:\n commit_d = self.commit()\n commit_d.addErrback(self._handle_auto_commit_error)\n else:\n d = Deferred()\n d.addCallback(self._retry_auto_commit, by_count)\n self._commit_ds.append(d)\n", "role": "user" }, { "content": "auto_commit", "role": "assistant" } ]
def x(self, after=None): if self._stopping or self._shuttingdown or self._start_d is None: return if self._retry_call is None: if after is None: after = self.retry_delay self.retry_delay = min(self.retry_delay * REQUEST_RETRY_FACTOR, self.retry_max_delay) self._fetch_attempt_count += 1 self._retry_call = self.client.reactor.callLater(after, self._do_fetch)
retry_fetch
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, after=None):\n if self._stopping or self._shuttingdown or self._start_d is None:\n return\n if self._retry_call is None:\n if after is None:\n after = self.retry_delay\n self.retry_delay = min(self.retry_delay * REQUEST_RETRY_FACTOR,\n self.retry_max_delay)\n self._fetch_attempt_count += 1\n self._retry_call = self.client.reactor.callLater(after, self._do_fetch)\n", "role": "user" }, { "content": "retry_fetch", "role": "assistant" } ]
def x(self, response): self._request_d = None self.retry_delay = self.retry_init_delay self._fetch_attempt_count = 1 response = response[0] if hasattr(response, 'offsets'): self._fetch_offset = response.offsets[0] elif response.offset == OFFSET_NOT_COMMITTED: self._fetch_offset = OFFSET_EARLIEST else: self._fetch_offset = response.offset + 1 self._last_committed_offset = response.offset self._do_fetch()
handle_offset_response
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, response):\n self._request_d = None\n self.retry_delay = self.retry_init_delay\n self._fetch_attempt_count = 1\n response = response[0]\n if hasattr(response, 'offsets'):\n self._fetch_offset = response.offsets[0]\n elif response.offset == OFFSET_NOT_COMMITTED:\n self._fetch_offset = OFFSET_EARLIEST\n else:\n self._fetch_offset = response.offset + 1\n self._last_committed_offset = response.offset\n self._do_fetch()\n", "role": "user" }, { "content": "handle_offset_response", "role": "assistant" } ]
def x(self, failure): self._request_d = None if self._stopping and failure.check(CancelledError): return if (self.request_retry_max_attempts != 0 and self._fetch_attempt_count >= self.request_retry_max_attempts): log.debug('%r: Exhausted attempts: %d fetching offset from kafka: %r', self, self.request_retry_max_attempts, failure) self._start_d.errback(failure) return if (self.retry_delay < self.retry_max_delay or 0 == self. _fetch_attempt_count % 2): log.debug('%r: Failure fetching offset from kafka: %r', self, failure) else: log.warning('%r: Still failing fetching offset from kafka: %r', self, failure) self._retry_fetch()
handle_offset_error
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, failure):\n self._request_d = None\n if self._stopping and failure.check(CancelledError):\n return\n if (self.request_retry_max_attempts != 0 and self._fetch_attempt_count >=\n self.request_retry_max_attempts):\n log.debug('%r: Exhausted attempts: %d fetching offset from kafka: %r',\n self, self.request_retry_max_attempts, failure)\n self._start_d.errback(failure)\n return\n if (self.retry_delay < self.retry_max_delay or 0 == self.\n _fetch_attempt_count % 2):\n log.debug('%r: Failure fetching offset from kafka: %r', self, failure)\n else:\n log.warning('%r: Still failing fetching offset from kafka: %r',\n self, failure)\n self._retry_fetch()\n", "role": "user" }, { "content": "handle_offset_error", "role": "assistant" } ]
def x(self, retry_delay=None, attempt=None): if self._commit_call and not self._commit_call.active(): self._commit_call = None if self._commit_req is not None: raise OperationInProgress(self._commit_req) if retry_delay is None: retry_delay = self.retry_init_delay if attempt is None: attempt = 1 commit_offset = self._last_processed_offset commit_request = OffsetCommitRequest(self.topic, self.partition, commit_offset, TIMESTAMP_INVALID, self.commit_metadata) log.debug('Committing off=%d grp=%s tpc=%s part=%s req=%r', self. _last_processed_offset, self.consumer_group, self.topic, self. partition, commit_request) self._commit_req = d = self.client.send_offset_commit_request(self. consumer_group, [commit_request]) d.addBoth(self._clear_commit_req) d.addCallbacks(self._update_committed_offset, self._handle_commit_error, callbackArgs=(commit_offset,), errbackArgs=(retry_delay, attempt))
send_commit_request
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, retry_delay=None, attempt=None):\n if self._commit_call and not self._commit_call.active():\n self._commit_call = None\n if self._commit_req is not None:\n raise OperationInProgress(self._commit_req)\n if retry_delay is None:\n retry_delay = self.retry_init_delay\n if attempt is None:\n attempt = 1\n commit_offset = self._last_processed_offset\n commit_request = OffsetCommitRequest(self.topic, self.partition,\n commit_offset, TIMESTAMP_INVALID, self.commit_metadata)\n log.debug('Committing off=%d grp=%s tpc=%s part=%s req=%r', self.\n _last_processed_offset, self.consumer_group, self.topic, self.\n partition, commit_request)\n self._commit_req = d = self.client.send_offset_commit_request(self.\n consumer_group, [commit_request])\n d.addBoth(self._clear_commit_req)\n d.addCallbacks(self._update_committed_offset, self._handle_commit_error,\n callbackArgs=(commit_offset,), errbackArgs=(retry_delay, attempt))\n", "role": "user" }, { "content": "send_commit_request", "role": "assistant" } ]
def x(self, failure, retry_delay, attempt): if self._stopping and failure.check(CancelledError): return self._deliver_commit_result(self._last_committed_offset) if not failure.check(KafkaError): log.error('Unhandleable failure during commit attempt: %r\n\t%r', failure, failure.getBriefTraceback()) return self._deliver_commit_result(failure) if (self.request_retry_max_attempts != 0 and attempt >= self. request_retry_max_attempts): log.debug('%r: Exhausted attempts: %d to commit offset: %r', self, self.request_retry_max_attempts, failure) return self._deliver_commit_result(failure) if retry_delay < self.retry_max_delay or 0 == attempt % 2: log.debug('%r: Failure committing offset to kafka: %r', self, failure) else: log.warning('%r: Still failing committing offset to kafka: %r', self, failure) retry_delay = min(retry_delay * REQUEST_RETRY_FACTOR, self.retry_max_delay) self._commit_call = self.client.reactor.callLater(retry_delay, self. _send_commit_request, retry_delay, attempt + 1)
handle_commit_error
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, failure, retry_delay, attempt):\n if self._stopping and failure.check(CancelledError):\n return self._deliver_commit_result(self._last_committed_offset)\n if not failure.check(KafkaError):\n log.error('Unhandleable failure during commit attempt: %r\\n\\t%r',\n failure, failure.getBriefTraceback())\n return self._deliver_commit_result(failure)\n if (self.request_retry_max_attempts != 0 and attempt >= self.\n request_retry_max_attempts):\n log.debug('%r: Exhausted attempts: %d to commit offset: %r', self,\n self.request_retry_max_attempts, failure)\n return self._deliver_commit_result(failure)\n if retry_delay < self.retry_max_delay or 0 == attempt % 2:\n log.debug('%r: Failure committing offset to kafka: %r', self, failure)\n else:\n log.warning('%r: Still failing committing offset to kafka: %r',\n self, failure)\n retry_delay = min(retry_delay * REQUEST_RETRY_FACTOR, self.retry_max_delay)\n self._commit_call = self.client.reactor.callLater(retry_delay, self.\n _send_commit_request, retry_delay, attempt + 1)\n", "role": "user" }, { "content": "handle_commit_error", "role": "assistant" } ]
def x(self, failure): if not (self._stopping and failure.check(CancelledError)): if self._start_d: self._start_d.errback(failure)
handle_processor_error
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, failure):\n if not (self._stopping and failure.check(CancelledError)):\n if self._start_d:\n self._start_d.errback(failure)\n", "role": "user" }, { "content": "handle_processor_error", "role": "assistant" } ]
def x(self, failure): self._request_d = None if failure.check(OffsetOutOfRangeError): if self.auto_offset_reset is None: self._start_d.errback(failure) return self._fetch_offset = self.auto_offset_reset if self._stopping and failure.check(CancelledError): return if (self.request_retry_max_attempts != 0 and self._fetch_attempt_count >= self.request_retry_max_attempts): log.debug('%r: Exhausted attempts: %d fetching messages from kafka: %r' , self, self.request_retry_max_attempts, failure) self._start_d.errback(failure) return if (self.retry_delay < self.retry_max_delay or 0 == self. _fetch_attempt_count % 2): log.debug('%r: Failure fetching messages from kafka: %r', self, failure ) else: log.warning('%r: Still failing fetching messages from kafka: %r', self, failure) self._retry_fetch()
handle_fetch_error
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, failure):\n self._request_d = None\n if failure.check(OffsetOutOfRangeError):\n if self.auto_offset_reset is None:\n self._start_d.errback(failure)\n return\n self._fetch_offset = self.auto_offset_reset\n if self._stopping and failure.check(CancelledError):\n return\n if (self.request_retry_max_attempts != 0 and self._fetch_attempt_count >=\n self.request_retry_max_attempts):\n log.debug('%r: Exhausted attempts: %d fetching messages from kafka: %r'\n , self, self.request_retry_max_attempts, failure)\n self._start_d.errback(failure)\n return\n if (self.retry_delay < self.retry_max_delay or 0 == self.\n _fetch_attempt_count % 2):\n log.debug('%r: Failure fetching messages from kafka: %r', self, failure\n )\n else:\n log.warning('%r: Still failing fetching messages from kafka: %r',\n self, failure)\n self._retry_fetch()\n", "role": "user" }, { "content": "handle_fetch_error", "role": "assistant" } ]
def x(self, responses): self.retry_delay = self.retry_init_delay self._fetch_attempt_count = 1 if self._msg_block_d: self._msg_block_d.addCallback(lambda _: self._handle_fetch_response (responses)) return self._request_d = None messages = [] try: for resp in responses: if resp.partition != self.partition: log.warning( '%r: Got response with partition: %r not our own: %r', self, resp.partition, self.partition) continue for message in resp.messages: if message.offset < self._fetch_offset: log.debug( 'Skipping message at offset: %d, because its offset is less that our fetch offset: %d.' , message.offset, self._fetch_offset) continue messages.append(SourcedMessage(message=message.message, offset=message.offset, topic=self.topic, partition=self .partition)) self._fetch_offset = message.offset + 1 except ConsumerFetchSizeTooSmall: factor = 2 if self.buffer_size <= 2 ** 20: factor = 16 if self.max_buffer_size is None: self.buffer_size *= factor elif self.max_buffer_size is not None and self.buffer_size < self.max_buffer_size: self.buffer_size = min(self.buffer_size * factor, self. max_buffer_size) else: log.error('Max fetch size %d too small', self.max_buffer_size) failure = Failure(ConsumerFetchSizeTooSmall( 'Max buffer size:%d too small for message', self. max_buffer_size)) self._start_d.errback(failure) return log.debug( 'Next message larger than fetch size, increasing to %d (~2x) and retrying' , self.buffer_size) finally: if messages: self._msg_block_d = Deferred() self._process_messages(messages) self._retry_fetch(0)
handle_fetch_response
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, responses):\n self.retry_delay = self.retry_init_delay\n self._fetch_attempt_count = 1\n if self._msg_block_d:\n self._msg_block_d.addCallback(lambda _: self._handle_fetch_response\n (responses))\n return\n self._request_d = None\n messages = []\n try:\n for resp in responses:\n if resp.partition != self.partition:\n log.warning(\n '%r: Got response with partition: %r not our own: %r',\n self, resp.partition, self.partition)\n continue\n for message in resp.messages:\n if message.offset < self._fetch_offset:\n log.debug(\n 'Skipping message at offset: %d, because its offset is less that our fetch offset: %d.'\n , message.offset, self._fetch_offset)\n continue\n messages.append(SourcedMessage(message=message.message,\n offset=message.offset, topic=self.topic, partition=self\n .partition))\n self._fetch_offset = message.offset + 1\n except ConsumerFetchSizeTooSmall:\n factor = 2\n if self.buffer_size <= 2 ** 20:\n factor = 16\n if self.max_buffer_size is None:\n self.buffer_size *= factor\n elif self.max_buffer_size is not None and self.buffer_size < self.max_buffer_size:\n self.buffer_size = min(self.buffer_size * factor, self.\n max_buffer_size)\n else:\n log.error('Max fetch size %d too small', self.max_buffer_size)\n failure = Failure(ConsumerFetchSizeTooSmall(\n 'Max buffer size:%d too small for message', self.\n max_buffer_size))\n self._start_d.errback(failure)\n return\n log.debug(\n 'Next message larger than fetch size, increasing to %d (~2x) and retrying'\n , self.buffer_size)\n finally:\n if messages:\n self._msg_block_d = Deferred()\n self._process_messages(messages)\n self._retry_fetch(0)\n", "role": "user" }, { "content": "handle_fetch_response", "role": "assistant" } ]
def x(self, messages): if self._shuttingdown: return if not messages: if self._msg_block_d: _msg_block_d, self._msg_block_d = self._msg_block_d, None _msg_block_d.callback(True) return proc_block_size = sys.maxsize if self.auto_commit_every_n: proc_block_size = self.auto_commit_every_n msgs_to_proc = messages[:proc_block_size] msgs_remainder = messages[proc_block_size:] last_offset = msgs_to_proc[-1].offset self._processor_d = d = maybeDeferred(self.processor, self, msgs_to_proc) log.debug('self.processor return: %r, last_offset: %r', d, last_offset) d.addBoth(self._clear_processor_deferred) d.addCallback(self._update_processed_offset, last_offset) if self._stopping or self._start_d is None: d.cancel() else: d.addCallback(lambda _: self._process_messages(msgs_remainder)) d.addErrback(self._handle_processor_error)
process_messages
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, messages):\n if self._shuttingdown:\n return\n if not messages:\n if self._msg_block_d:\n _msg_block_d, self._msg_block_d = self._msg_block_d, None\n _msg_block_d.callback(True)\n return\n proc_block_size = sys.maxsize\n if self.auto_commit_every_n:\n proc_block_size = self.auto_commit_every_n\n msgs_to_proc = messages[:proc_block_size]\n msgs_remainder = messages[proc_block_size:]\n last_offset = msgs_to_proc[-1].offset\n self._processor_d = d = maybeDeferred(self.processor, self, msgs_to_proc)\n log.debug('self.processor return: %r, last_offset: %r', d, last_offset)\n d.addBoth(self._clear_processor_deferred)\n d.addCallback(self._update_processed_offset, last_offset)\n if self._stopping or self._start_d is None:\n d.cancel()\n else:\n d.addCallback(lambda _: self._process_messages(msgs_remainder))\n d.addErrback(self._handle_processor_error)\n", "role": "user" }, { "content": "process_messages", "role": "assistant" } ]