source_code
stringlengths 45
7.24k
| function_name
stringlengths 1
111
| messages
list |
|---|---|---|
def x(self, line):
stripped = line.strip()
if stripped and stripped.startswith('#') is False:
rexres = self.rex.search(stripped)
if rexres:
return ' '.join(rexres.group(1).split())
return None
|
parse_cron_line
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, line):\n stripped = line.strip()\n if stripped and stripped.startswith('#') is False:\n rexres = self.rex.search(stripped)\n if rexres:\n return ' '.join(rexres.group(1).split())\n return None\n",
"role": "user"
},
{
"content": "parse_cron_line",
"role": "assistant"
}
] |
def x(self, new):
if self.node_id != new.node_id:
raise ValueError("Broker metadata {!r} doesn't match node_id={}".
format(new, self.node_id))
self.node_id = new.node_id
self.host = new.host
self.port = new.port
|
updateMetadata
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, new):\n if self.node_id != new.node_id:\n raise ValueError(\"Broker metadata {!r} doesn't match node_id={}\".\n format(new, self.node_id))\n self.node_id = new.node_id\n self.host = new.host\n self.port = new.port\n",
"role": "user"
},
{
"content": "updateMetadata",
"role": "assistant"
}
] |
def x(self, requestId, request, expectResponse=True):
if requestId in self.requests:
raise DuplicateRequestError('Reuse of requestId:{}'.format(requestId))
if self._dDown:
return fail(ClientError('makeRequest() called after close()'))
canceller = partial(self.cancelRequest, requestId, CancelledError(
message='Request correlationId={} was cancelled'.format(requestId)))
tReq = _Request(requestId, request, expectResponse, canceller)
self.requests[requestId] = tReq
tReq.d.addErrback(self._handleRequestFailure, requestId)
if self.proto:
self._sendRequest(tReq)
elif not self.connector:
self._connect()
return tReq.d
|
makeRequest
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, requestId, request, expectResponse=True):\n if requestId in self.requests:\n raise DuplicateRequestError('Reuse of requestId:{}'.format(requestId))\n if self._dDown:\n return fail(ClientError('makeRequest() called after close()'))\n canceller = partial(self.cancelRequest, requestId, CancelledError(\n message='Request correlationId={} was cancelled'.format(requestId)))\n tReq = _Request(requestId, request, expectResponse, canceller)\n self.requests[requestId] = tReq\n tReq.d.addErrback(self._handleRequestFailure, requestId)\n if self.proto:\n self._sendRequest(tReq)\n elif not self.connector:\n self._connect()\n return tReq.d\n",
"role": "user"
},
{
"content": "makeRequest",
"role": "assistant"
}
] |
def x(self):
if self.proto:
log.debug('%r Disconnecting from %r', self, self.proto.transport.
getPeer())
self.proto.transport.loseConnection()
|
disconnect
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.proto:\n log.debug('%r Disconnecting from %r', self, self.proto.transport.\n getPeer())\n self.proto.transport.loseConnection()\n",
"role": "user"
},
{
"content": "disconnect",
"role": "assistant"
}
] |
def x(self):
log.debug('%r: close() proto=%r connector=%r', self, self.proto, self.
connector)
assert self._dDown is None
self._dDown = Deferred()
if self.proto is not None:
self.proto.transport.loseConnection()
elif self.connector is not None:
def x(reason):
log.debug('%r: connection attempt has been cancelled: %r', self,
reason)
self._dDown.callback(None)
self.connector.addErrback(connectingFailed)
self.connector.cancel()
else:
self._dDown.callback(None)
try:
raise CancelledError(message=
'Broker client for node_id={} {}:{} was closed'.format(self.
node_id, self.host, self.port))
except Exception:
reason = Failure()
for correlation_id in list(self.requests.keys()):
self.cancelRequest(correlation_id, reason)
return self._dDown
|
close
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n log.debug('%r: close() proto=%r connector=%r', self, self.proto, self.\n connector)\n assert self._dDown is None\n self._dDown = Deferred()\n if self.proto is not None:\n self.proto.transport.loseConnection()\n elif self.connector is not None:\n\n def x(reason):\n log.debug('%r: connection attempt has been cancelled: %r', self,\n reason)\n self._dDown.callback(None)\n self.connector.addErrback(connectingFailed)\n self.connector.cancel()\n else:\n self._dDown.callback(None)\n try:\n raise CancelledError(message=\n 'Broker client for node_id={} {}:{} was closed'.format(self.\n node_id, self.host, self.port))\n except Exception:\n reason = Failure()\n for correlation_id in list(self.requests.keys()):\n self.cancelRequest(correlation_id, reason)\n return self._dDown\n",
"role": "user"
},
{
"content": "close",
"role": "assistant"
}
] |
def x(self, reason):
log.info('%r: Connection closed: %r', self, reason)
self.proto = None
for tReq in self.requests.values():
tReq.sent = False
if self._dDown:
self._dDown.callback(None)
elif self.requests:
self._connect()
|
connectionLost
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, reason):\n log.info('%r: Connection closed: %r', self, reason)\n self.proto = None\n for tReq in self.requests.values():\n tReq.sent = False\n if self._dDown:\n self._dDown.callback(None)\n elif self.requests:\n self._connect()\n",
"role": "user"
},
{
"content": "connectionLost",
"role": "assistant"
}
] |
def x(self, response):
requestId = KafkaCodec.get_response_correlation_id(response)
tReq = self.requests.pop(requestId, None)
if tReq is None:
log.warning('Unexpected response with correlationId=%d: %r',
requestId, reprlib.repr(response))
else:
tReq.d.callback(response)
|
handleResponse
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, response):\n requestId = KafkaCodec.get_response_correlation_id(response)\n tReq = self.requests.pop(requestId, None)\n if tReq is None:\n log.warning('Unexpected response with correlationId=%d: %r',\n requestId, reprlib.repr(response))\n else:\n tReq.d.callback(response)\n",
"role": "user"
},
{
"content": "handleResponse",
"role": "assistant"
}
] |
def x(self, tReq):
try:
tReq.sent = True
self.proto.sendString(tReq.data)
except Exception as e:
log.exception('%r: Failed to send request %r', self, tReq)
del self.requests[tReq.id]
tReq.d.errback(e)
else:
if not tReq.expect:
del self.requests[tReq.id]
tReq.d.callback(None)
|
sendRequest
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, tReq):\n try:\n tReq.sent = True\n self.proto.sendString(tReq.data)\n except Exception as e:\n log.exception('%r: Failed to send request %r', self, tReq)\n del self.requests[tReq.id]\n tReq.d.errback(e)\n else:\n if not tReq.expect:\n del self.requests[tReq.id]\n tReq.d.callback(None)\n",
"role": "user"
},
{
"content": "sendRequest",
"role": "assistant"
}
] |
def x(self):
for tReq in list(self.requests.values()):
if not tReq.sent:
self._sendRequest(tReq)
|
sendQueued
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n for tReq in list(self.requests.values()):\n if not tReq.sent:\n self._sendRequest(tReq)\n",
"role": "user"
},
{
"content": "sendQueued",
"role": "assistant"
}
] |
def x(self, requestId, reason=None, _=None):
if reason is None:
reason = CancelledError()
tReq = self.requests.pop(requestId)
tReq.d.errback(reason)
|
cancelRequest
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, requestId, reason=None, _=None):\n if reason is None:\n reason = CancelledError()\n tReq = self.requests.pop(requestId)\n tReq.d.errback(reason)\n",
"role": "user"
},
{
"content": "cancelRequest",
"role": "assistant"
}
] |
def x(self):
def x():
self.connector = d = maybeDeferred(connect)
d.addCallback(cbConnect)
d.addErrback(ebConnect)
def x():
endpoint = self._endpointFactory(self._reactor, self.host, self.port)
log.debug('%r: connecting with %s', self, endpoint)
return endpoint.connect(self)
def x(proto):
log.debug('%r: connected to %r', self, proto.transport.getPeer())
self._failures = 0
self.connector = None
self.proto = proto
if self._dDown:
proto.transport.loseConnection()
else:
self._sendQueued()
def x(fail):
if self._dDown:
log.debug('%r: breaking connect loop due to %r after close()',
self, fail)
return fail
self._failures += 1
delay = self._retryPolicy(self._failures)
log.debug('%r: failure %d to connect -> %s; retry in %.2f seconds.',
self, self._failures, fail.value, delay)
self.connector = d = deferLater(self._reactor, delay, lambda : None)
d.addCallback(cbDelayed)
def x(result):
tryConnect()
self._failures = 0
tryConnect()
|
connect
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n\n def x():\n self.connector = d = maybeDeferred(connect)\n d.addCallback(cbConnect)\n d.addErrback(ebConnect)\n\n def x():\n endpoint = self._endpointFactory(self._reactor, self.host, self.port)\n log.debug('%r: connecting with %s', self, endpoint)\n return endpoint.connect(self)\n\n def x(proto):\n log.debug('%r: connected to %r', self, proto.transport.getPeer())\n self._failures = 0\n self.connector = None\n self.proto = proto\n if self._dDown:\n proto.transport.loseConnection()\n else:\n self._sendQueued()\n\n def x(fail):\n if self._dDown:\n log.debug('%r: breaking connect loop due to %r after close()',\n self, fail)\n return fail\n self._failures += 1\n delay = self._retryPolicy(self._failures)\n log.debug('%r: failure %d to connect -> %s; retry in %.2f seconds.',\n self, self._failures, fail.value, delay)\n self.connector = d = deferLater(self._reactor, delay, lambda : None)\n d.addCallback(cbDelayed)\n\n def x(result):\n tryConnect()\n self._failures = 0\n tryConnect()\n",
"role": "user"
},
{
"content": "connect",
"role": "assistant"
}
] |
def x(self):
parsed = ['', '', '', '', '', '', '']
if self._expression is None or len(self._expression) == 0:
raise MissingFieldException('ExpressionDescriptor.expression')
else:
expression_parts_temp = self._expression.split()
expression_parts_temp_length = len(expression_parts_temp)
if expression_parts_temp_length < 5:
raise FormatException(
'Error: Expression only has {0} parts. At least 5 part are required.'
.format(expression_parts_temp_length))
elif expression_parts_temp_length == 5:
for i, expression_part_temp in enumerate(expression_parts_temp):
parsed[i + 1] = expression_part_temp
elif expression_parts_temp_length == 6:
year_regex = re.compile('\\d{4}$')
if year_regex.search(expression_parts_temp[5]) is not None:
for i, expression_part_temp in enumerate(expression_parts_temp
):
parsed[i + 1] = expression_part_temp
else:
for i, expression_part_temp in enumerate(expression_parts_temp
):
parsed[i] = expression_part_temp
elif expression_parts_temp_length == 7:
parsed = expression_parts_temp
else:
raise FormatException(
'Error: Expression has too many parts ({0}). Expression must not have more than 7 parts.'
.format(expression_parts_temp_length))
self.normalize_expression(parsed)
return parsed
|
parse
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n parsed = ['', '', '', '', '', '', '']\n if self._expression is None or len(self._expression) == 0:\n raise MissingFieldException('ExpressionDescriptor.expression')\n else:\n expression_parts_temp = self._expression.split()\n expression_parts_temp_length = len(expression_parts_temp)\n if expression_parts_temp_length < 5:\n raise FormatException(\n 'Error: Expression only has {0} parts. At least 5 part are required.'\n .format(expression_parts_temp_length))\n elif expression_parts_temp_length == 5:\n for i, expression_part_temp in enumerate(expression_parts_temp):\n parsed[i + 1] = expression_part_temp\n elif expression_parts_temp_length == 6:\n year_regex = re.compile('\\\\d{4}$')\n if year_regex.search(expression_parts_temp[5]) is not None:\n for i, expression_part_temp in enumerate(expression_parts_temp\n ):\n parsed[i + 1] = expression_part_temp\n else:\n for i, expression_part_temp in enumerate(expression_parts_temp\n ):\n parsed[i] = expression_part_temp\n elif expression_parts_temp_length == 7:\n parsed = expression_parts_temp\n else:\n raise FormatException(\n 'Error: Expression has too many parts ({0}). Expression must not have more than 7 parts.'\n .format(expression_parts_temp_length))\n self.normalize_expression(parsed)\n return parsed\n",
"role": "user"
},
{
"content": "parse",
"role": "assistant"
}
] |
def x(self, expression_parts):
expression_parts[3] = expression_parts[3].replace('?', '*')
expression_parts[5] = expression_parts[5].replace('?', '*')
if expression_parts[0].startswith('0/'):
expression_parts[0] = expression_parts[0].replace('0/', '*/')
if expression_parts[1].startswith('0/'):
expression_parts[1] = expression_parts[1].replace('0/', '*/')
if expression_parts[2].startswith('0/'):
expression_parts[2] = expression_parts[2].replace('0/', '*/')
if expression_parts[3].startswith('1/'):
expression_parts[3] = expression_parts[3].replace('1/', '*/')
if expression_parts[4].startswith('1/'):
expression_parts[4] = expression_parts[4].replace('1/', '*/')
if expression_parts[5].startswith('1/'):
expression_parts[5] = expression_parts[5].replace('1/', '*/')
if expression_parts[6].startswith('1/'):
expression_parts[6] = expression_parts[6].replace('1/', '*/')
if self._options.day_of_week_start_index_zero is False:
expression_parts[5] = self.decrease_days_of_week(expression_parts[5])
if expression_parts[3] == '?':
expression_parts[3] = '*'
for day_number in self._cron_days:
expression_parts[5] = expression_parts[5].upper().replace(self.
_cron_days[day_number], str(day_number))
for month_number in self._cron_months:
expression_parts[4] = expression_parts[4].upper().replace(self.
_cron_months[month_number], str(month_number))
if expression_parts[0] == '0':
expression_parts[0] = ''
length = len(expression_parts)
for i in range(length):
if expression_parts[i] == '*/1':
expression_parts[i] = '*'
"""
Convert Month,DOW,Year step values with a starting value (i.e. not '*') to between expressions.
This allows us to reuse the between expression handling for step values.
For Example:
- month part '3/2' will be converted to '3-12/2' (every 2 months between March and December)
- DOW part '3/2' will be converted to '3-6/2' (every 2 days between Tuesday and Saturday)
"""
if '/' in expression_parts[i] and any(exp in expression_parts[i] for
exp in ['*', '-', ',']) is False:
choices = {(4): '12', (5): '6', (6): '9999'}
step_range_through = choices.get(i)
if step_range_through is not None:
parts = expression_parts[i].split('/')
expression_parts[i] = '{0}-{1}/{2}'.format(parts[0],
step_range_through, parts[1])
|
normalize_expression
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, expression_parts):\n expression_parts[3] = expression_parts[3].replace('?', '*')\n expression_parts[5] = expression_parts[5].replace('?', '*')\n if expression_parts[0].startswith('0/'):\n expression_parts[0] = expression_parts[0].replace('0/', '*/')\n if expression_parts[1].startswith('0/'):\n expression_parts[1] = expression_parts[1].replace('0/', '*/')\n if expression_parts[2].startswith('0/'):\n expression_parts[2] = expression_parts[2].replace('0/', '*/')\n if expression_parts[3].startswith('1/'):\n expression_parts[3] = expression_parts[3].replace('1/', '*/')\n if expression_parts[4].startswith('1/'):\n expression_parts[4] = expression_parts[4].replace('1/', '*/')\n if expression_parts[5].startswith('1/'):\n expression_parts[5] = expression_parts[5].replace('1/', '*/')\n if expression_parts[6].startswith('1/'):\n expression_parts[6] = expression_parts[6].replace('1/', '*/')\n if self._options.day_of_week_start_index_zero is False:\n expression_parts[5] = self.decrease_days_of_week(expression_parts[5])\n if expression_parts[3] == '?':\n expression_parts[3] = '*'\n for day_number in self._cron_days:\n expression_parts[5] = expression_parts[5].upper().replace(self.\n _cron_days[day_number], str(day_number))\n for month_number in self._cron_months:\n expression_parts[4] = expression_parts[4].upper().replace(self.\n _cron_months[month_number], str(month_number))\n if expression_parts[0] == '0':\n expression_parts[0] = ''\n length = len(expression_parts)\n for i in range(length):\n if expression_parts[i] == '*/1':\n expression_parts[i] = '*'\n \"\"\"\n Convert Month,DOW,Year step values with a starting value (i.e. not '*') to between expressions.\n This allows us to reuse the between expression handling for step values.\n\n For Example:\n - month part '3/2' will be converted to '3-12/2' (every 2 months between March and December)\n - DOW part '3/2' will be converted to '3-6/2' (every 2 days between Tuesday and Saturday)\n \"\"\"\n if '/' in expression_parts[i] and any(exp in expression_parts[i] for\n exp in ['*', '-', ',']) is False:\n choices = {(4): '12', (5): '6', (6): '9999'}\n step_range_through = choices.get(i)\n if step_range_through is not None:\n parts = expression_parts[i].split('/')\n expression_parts[i] = '{0}-{1}/{2}'.format(parts[0],\n step_range_through, parts[1])\n",
"role": "user"
},
{
"content": "normalize_expression",
"role": "assistant"
}
] |
def x(envlist):
groups = {}
for env in envlist:
envpy, category = env.split('-')[0:2]
if category == 'lint':
category = 'unit'
try:
groups[envpy, category].append(env)
except KeyError:
groups[envpy, category] = [env]
return sorted((envpy, category, envs) for (envpy, category), envs in
groups.items())
|
group_envs
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(envlist):\n groups = {}\n for env in envlist:\n envpy, category = env.split('-')[0:2]\n if category == 'lint':\n category = 'unit'\n try:\n groups[envpy, category].append(env)\n except KeyError:\n groups[envpy, category] = [env]\n return sorted((envpy, category, envs) for (envpy, category), envs in\n groups.items())\n",
"role": "user"
},
{
"content": "group_envs",
"role": "assistant"
}
] |
def x(payload, key=None):
assert payload is None or isinstance(payload, bytes
), 'payload={!r} should be bytes or None'.format(payload)
assert key is None or isinstance(key, bytes
), 'key={!r} should be bytes or None'.format(key)
return Message(0, 0, key, payload)
|
create_message
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(payload, key=None):\n assert payload is None or isinstance(payload, bytes\n ), 'payload={!r} should be bytes or None'.format(payload)\n assert key is None or isinstance(key, bytes\n ), 'key={!r} should be bytes or None'.format(key)\n return Message(0, 0, key, payload)\n",
"role": "user"
},
{
"content": "create_message",
"role": "assistant"
}
] |
def x(message_set):
encoded_message_set = KafkaCodec._encode_message_set(message_set)
gzipped = gzip_encode(encoded_message_set)
return Message(0, CODEC_GZIP, None, gzipped)
|
create_gzip_message
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(message_set):\n encoded_message_set = KafkaCodec._encode_message_set(message_set)\n gzipped = gzip_encode(encoded_message_set)\n return Message(0, CODEC_GZIP, None, gzipped)\n",
"role": "user"
},
{
"content": "create_gzip_message",
"role": "assistant"
}
] |
def x(message_set):
encoded_message_set = KafkaCodec._encode_message_set(message_set)
snapped = snappy_encode(encoded_message_set)
return Message(0, CODEC_SNAPPY, None, snapped)
|
create_snappy_message
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(message_set):\n encoded_message_set = KafkaCodec._encode_message_set(message_set)\n snapped = snappy_encode(encoded_message_set)\n return Message(0, CODEC_SNAPPY, None, snapped)\n",
"role": "user"
},
{
"content": "create_snappy_message",
"role": "assistant"
}
] |
def x(requests, codec=CODEC_NONE):
msglist = []
for req in requests:
msglist.extend([create_message(m, key=req.key) for m in req.messages])
if codec == CODEC_NONE:
return msglist
elif codec == CODEC_GZIP:
return [create_gzip_message(msglist)]
elif codec == CODEC_SNAPPY:
return [create_snappy_message(msglist)]
else:
raise UnsupportedCodecError('Codec 0x%02x unsupported' % codec)
|
create_message_set
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(requests, codec=CODEC_NONE):\n msglist = []\n for req in requests:\n msglist.extend([create_message(m, key=req.key) for m in req.messages])\n if codec == CODEC_NONE:\n return msglist\n elif codec == CODEC_GZIP:\n return [create_gzip_message(msglist)]\n elif codec == CODEC_SNAPPY:\n return [create_snappy_message(msglist)]\n else:\n raise UnsupportedCodecError('Codec 0x%02x unsupported' % codec)\n",
"role": "user"
},
{
"content": "create_message_set",
"role": "assistant"
}
] |
def x(cls, client_id, correlation_id, request_key, api_version=0):
return struct.pack('>hhih', request_key, api_version, correlation_id,
len(client_id)) + client_id
|
encode_message_header
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, client_id, correlation_id, request_key, api_version=0):\n return struct.pack('>hhih', request_key, api_version, correlation_id,\n len(client_id)) + client_id\n",
"role": "user"
},
{
"content": "encode_message_header",
"role": "assistant"
}
] |
def x(cls, messages, offset=None):
message_set = []
incr = 1
if offset is None:
incr = 0
offset = 0
for message in messages:
encoded_message = KafkaCodec._encode_message(message)
message_set.append(struct.pack('>qi', offset, len(encoded_message)))
message_set.append(encoded_message)
offset += incr
return b''.join(message_set)
|
encode_message_set
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, messages, offset=None):\n message_set = []\n incr = 1\n if offset is None:\n incr = 0\n offset = 0\n for message in messages:\n encoded_message = KafkaCodec._encode_message(message)\n message_set.append(struct.pack('>qi', offset, len(encoded_message)))\n message_set.append(encoded_message)\n offset += incr\n return b''.join(message_set)\n",
"role": "user"
},
{
"content": "encode_message_set",
"role": "assistant"
}
] |
def x(cls, message):
if message.magic == 0:
msg = struct.pack('>BB', message.magic, message.attributes)
msg += write_int_string(message.key)
msg += write_int_string(message.value)
crc = zlib.crc32(msg) & 4294967295
msg = struct.pack('>I', crc) + msg
else:
raise ProtocolError('Unexpected magic number: %d' % message.magic)
return msg
|
encode_message
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, message):\n if message.magic == 0:\n msg = struct.pack('>BB', message.magic, message.attributes)\n msg += write_int_string(message.key)\n msg += write_int_string(message.value)\n crc = zlib.crc32(msg) & 4294967295\n msg = struct.pack('>I', crc) + msg\n else:\n raise ProtocolError('Unexpected magic number: %d' % message.magic)\n return msg\n",
"role": "user"
},
{
"content": "encode_message",
"role": "assistant"
}
] |
def x(cls, client_id, correlation_id, payloads=None, acks=1, timeout=
DEFAULT_REPLICAS_ACK_TIMEOUT_MSECS):
if not isinstance(client_id, bytes):
raise TypeError('client_id={!r} should be bytes'.format(client_id))
payloads = [] if payloads is None else payloads
grouped_payloads = group_by_topic_and_partition(payloads)
message = cls._encode_message_header(client_id, correlation_id,
KafkaCodec.PRODUCE_KEY)
message += struct.pack('>hii', acks, timeout, len(grouped_payloads))
for topic, topic_payloads in grouped_payloads.items():
message += write_short_ascii(topic)
message += struct.pack('>i', len(topic_payloads))
for partition, payload in topic_payloads.items():
msg_set = KafkaCodec._encode_message_set(payload.messages)
message += struct.pack('>ii', partition, len(msg_set))
message += msg_set
return message
|
encode_produce_request
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, client_id, correlation_id, payloads=None, acks=1, timeout=\n DEFAULT_REPLICAS_ACK_TIMEOUT_MSECS):\n if not isinstance(client_id, bytes):\n raise TypeError('client_id={!r} should be bytes'.format(client_id))\n payloads = [] if payloads is None else payloads\n grouped_payloads = group_by_topic_and_partition(payloads)\n message = cls._encode_message_header(client_id, correlation_id,\n KafkaCodec.PRODUCE_KEY)\n message += struct.pack('>hii', acks, timeout, len(grouped_payloads))\n for topic, topic_payloads in grouped_payloads.items():\n message += write_short_ascii(topic)\n message += struct.pack('>i', len(topic_payloads))\n for partition, payload in topic_payloads.items():\n msg_set = KafkaCodec._encode_message_set(payload.messages)\n message += struct.pack('>ii', partition, len(msg_set))\n message += msg_set\n return message\n",
"role": "user"
},
{
"content": "encode_produce_request",
"role": "assistant"
}
] |
def x(cls, data):
(correlation_id, num_topics), cur = relative_unpack('>ii', data, 0)
for _i in range(num_topics):
topic, cur = read_short_ascii(data, cur)
(num_partitions,), cur = relative_unpack('>i', data, cur)
for _i in range(num_partitions):
(partition, error, offset), cur = relative_unpack('>ihq', data, cur
)
yield ProduceResponse(topic, partition, error, offset)
|
decode_produce_response
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, data):\n (correlation_id, num_topics), cur = relative_unpack('>ii', data, 0)\n for _i in range(num_topics):\n topic, cur = read_short_ascii(data, cur)\n (num_partitions,), cur = relative_unpack('>i', data, cur)\n for _i in range(num_partitions):\n (partition, error, offset), cur = relative_unpack('>ihq', data, cur\n )\n yield ProduceResponse(topic, partition, error, offset)\n",
"role": "user"
},
{
"content": "decode_produce_response",
"role": "assistant"
}
] |
def x(cls, client_id, correlation_id, payloads=None, max_wait_time=100,
min_bytes=4096):
payloads = [] if payloads is None else payloads
grouped_payloads = group_by_topic_and_partition(payloads)
message = cls._encode_message_header(client_id, correlation_id,
KafkaCodec.FETCH_KEY)
assert isinstance(max_wait_time, int)
message += struct.pack('>iiii', -1, max_wait_time, min_bytes, len(
grouped_payloads))
for topic, topic_payloads in grouped_payloads.items():
message += write_short_ascii(topic)
message += struct.pack('>i', len(topic_payloads))
for partition, payload in topic_payloads.items():
message += struct.pack('>iqi', partition, payload.offset,
payload.max_bytes)
return message
|
encode_fetch_request
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, client_id, correlation_id, payloads=None, max_wait_time=100,\n min_bytes=4096):\n payloads = [] if payloads is None else payloads\n grouped_payloads = group_by_topic_and_partition(payloads)\n message = cls._encode_message_header(client_id, correlation_id,\n KafkaCodec.FETCH_KEY)\n assert isinstance(max_wait_time, int)\n message += struct.pack('>iiii', -1, max_wait_time, min_bytes, len(\n grouped_payloads))\n for topic, topic_payloads in grouped_payloads.items():\n message += write_short_ascii(topic)\n message += struct.pack('>i', len(topic_payloads))\n for partition, payload in topic_payloads.items():\n message += struct.pack('>iqi', partition, payload.offset,\n payload.max_bytes)\n return message\n",
"role": "user"
},
{
"content": "encode_fetch_request",
"role": "assistant"
}
] |
def x(cls, data):
(correlation_id, num_topics), cur = relative_unpack('>ii', data, 0)
for _i in range(num_topics):
topic, cur = read_short_ascii(data, cur)
(num_partitions,), cur = relative_unpack('>i', data, cur)
for _i in range(num_partitions):
(partition, error, highwater_mark_offset), cur = relative_unpack(
'>ihq', data, cur)
message_set, cur = read_int_string(data, cur)
yield FetchResponse(topic, partition, error,
highwater_mark_offset, KafkaCodec._decode_message_set_iter(
message_set))
|
decode_fetch_response
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, data):\n (correlation_id, num_topics), cur = relative_unpack('>ii', data, 0)\n for _i in range(num_topics):\n topic, cur = read_short_ascii(data, cur)\n (num_partitions,), cur = relative_unpack('>i', data, cur)\n for _i in range(num_partitions):\n (partition, error, highwater_mark_offset), cur = relative_unpack(\n '>ihq', data, cur)\n message_set, cur = read_int_string(data, cur)\n yield FetchResponse(topic, partition, error,\n highwater_mark_offset, KafkaCodec._decode_message_set_iter(\n message_set))\n",
"role": "user"
},
{
"content": "decode_fetch_response",
"role": "assistant"
}
] |
def x(cls, data):
(correlation_id, num_topics), cur = relative_unpack('>ii', data, 0)
for _i in range(num_topics):
topic, cur = read_short_ascii(data, cur)
(num_partitions,), cur = relative_unpack('>i', data, cur)
for _i in range(num_partitions):
(partition, error, num_offsets), cur = relative_unpack('>ihi',
data, cur)
offsets = []
for _i in range(num_offsets):
(offset,), cur = relative_unpack('>q', data, cur)
offsets.append(offset)
yield OffsetResponse(topic, partition, error, tuple(offsets))
|
decode_offset_response
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, data):\n (correlation_id, num_topics), cur = relative_unpack('>ii', data, 0)\n for _i in range(num_topics):\n topic, cur = read_short_ascii(data, cur)\n (num_partitions,), cur = relative_unpack('>i', data, cur)\n for _i in range(num_partitions):\n (partition, error, num_offsets), cur = relative_unpack('>ihi',\n data, cur)\n offsets = []\n for _i in range(num_offsets):\n (offset,), cur = relative_unpack('>q', data, cur)\n offsets.append(offset)\n yield OffsetResponse(topic, partition, error, tuple(offsets))\n",
"role": "user"
},
{
"content": "decode_offset_response",
"role": "assistant"
}
] |
def x(cls, client_id, correlation_id, topics=None):
topics = [] if topics is None else topics
message = [cls._encode_message_header(client_id, correlation_id,
KafkaCodec.METADATA_KEY), struct.pack('>i', len(topics))]
for topic in topics:
message.append(write_short_ascii(topic))
return b''.join(message)
|
encode_metadata_request
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, client_id, correlation_id, topics=None):\n topics = [] if topics is None else topics\n message = [cls._encode_message_header(client_id, correlation_id,\n KafkaCodec.METADATA_KEY), struct.pack('>i', len(topics))]\n for topic in topics:\n message.append(write_short_ascii(topic))\n return b''.join(message)\n",
"role": "user"
},
{
"content": "encode_metadata_request",
"role": "assistant"
}
] |
def x(cls, data):
(correlation_id, numbrokers), cur = relative_unpack('>ii', data, 0)
if numbrokers > MAX_BROKERS:
raise InvalidMessageError('Brokers:{} exceeds max:{}'.format(
numbrokers, MAX_BROKERS))
brokers = {}
for _i in range(numbrokers):
(nodeId,), cur = relative_unpack('>i', data, cur)
host, cur = read_short_ascii(data, cur)
(port,), cur = relative_unpack('>i', data, cur)
brokers[nodeId] = BrokerMetadata(nodeId, nativeString(host), port)
(num_topics,), cur = relative_unpack('>i', data, cur)
topic_metadata = {}
for _i in range(num_topics):
(topic_error,), cur = relative_unpack('>h', data, cur)
topic_name, cur = read_short_ascii(data, cur)
(num_partitions,), cur = relative_unpack('>i', data, cur)
partition_metadata = {}
for _j in range(num_partitions):
(partition_error_code, partition, leader, numReplicas
), cur = relative_unpack('>hiii', data, cur)
replicas, cur = relative_unpack('>%di' % numReplicas, data, cur)
(num_isr,), cur = relative_unpack('>i', data, cur)
isr, cur = relative_unpack('>%di' % num_isr, data, cur)
partition_metadata[partition] = PartitionMetadata(topic_name,
partition, partition_error_code, leader, replicas, isr)
topic_metadata[topic_name] = TopicMetadata(topic_name, topic_error,
partition_metadata)
return brokers, topic_metadata
|
decode_metadata_response
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, data):\n (correlation_id, numbrokers), cur = relative_unpack('>ii', data, 0)\n if numbrokers > MAX_BROKERS:\n raise InvalidMessageError('Brokers:{} exceeds max:{}'.format(\n numbrokers, MAX_BROKERS))\n brokers = {}\n for _i in range(numbrokers):\n (nodeId,), cur = relative_unpack('>i', data, cur)\n host, cur = read_short_ascii(data, cur)\n (port,), cur = relative_unpack('>i', data, cur)\n brokers[nodeId] = BrokerMetadata(nodeId, nativeString(host), port)\n (num_topics,), cur = relative_unpack('>i', data, cur)\n topic_metadata = {}\n for _i in range(num_topics):\n (topic_error,), cur = relative_unpack('>h', data, cur)\n topic_name, cur = read_short_ascii(data, cur)\n (num_partitions,), cur = relative_unpack('>i', data, cur)\n partition_metadata = {}\n for _j in range(num_partitions):\n (partition_error_code, partition, leader, numReplicas\n ), cur = relative_unpack('>hiii', data, cur)\n replicas, cur = relative_unpack('>%di' % numReplicas, data, cur)\n (num_isr,), cur = relative_unpack('>i', data, cur)\n isr, cur = relative_unpack('>%di' % num_isr, data, cur)\n partition_metadata[partition] = PartitionMetadata(topic_name,\n partition, partition_error_code, leader, replicas, isr)\n topic_metadata[topic_name] = TopicMetadata(topic_name, topic_error,\n partition_metadata)\n return brokers, topic_metadata\n",
"role": "user"
},
{
"content": "decode_metadata_response",
"role": "assistant"
}
] |
def x(cls, client_id, correlation_id, consumer_group):
message = cls._encode_message_header(client_id, correlation_id,
KafkaCodec.CONSUMER_METADATA_KEY)
message += write_short_ascii(consumer_group)
return message
|
encode_consumermetadata_request
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, client_id, correlation_id, consumer_group):\n message = cls._encode_message_header(client_id, correlation_id,\n KafkaCodec.CONSUMER_METADATA_KEY)\n message += write_short_ascii(consumer_group)\n return message\n",
"role": "user"
},
{
"content": "encode_consumermetadata_request",
"role": "assistant"
}
] |
def x(cls, data):
(correlation_id, error_code, node_id), cur = relative_unpack('>ihi',
data, 0)
host, cur = read_short_ascii(data, cur)
(port,), cur = relative_unpack('>i', data, cur)
return ConsumerMetadataResponse(error_code, node_id, nativeString(host),
port)
|
decode_consumermetadata_response
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, data):\n (correlation_id, error_code, node_id), cur = relative_unpack('>ihi',\n data, 0)\n host, cur = read_short_ascii(data, cur)\n (port,), cur = relative_unpack('>i', data, cur)\n return ConsumerMetadataResponse(error_code, node_id, nativeString(host),\n port)\n",
"role": "user"
},
{
"content": "decode_consumermetadata_response",
"role": "assistant"
}
] |
def x(cls, client_id, correlation_id, group, group_generation_id,
consumer_id, payloads):
grouped_payloads = group_by_topic_and_partition(payloads)
message = cls._encode_message_header(client_id, correlation_id,
KafkaCodec.OFFSET_COMMIT_KEY, api_version=1)
message += write_short_ascii(group)
message += struct.pack('>i', group_generation_id)
message += write_short_ascii(consumer_id)
message += struct.pack('>i', len(grouped_payloads))
for topic, topic_payloads in grouped_payloads.items():
message += write_short_ascii(topic)
message += struct.pack('>i', len(topic_payloads))
for partition, payload in topic_payloads.items():
message += struct.pack('>iqq', partition, payload.offset,
payload.timestamp)
message += write_short_bytes(payload.metadata)
return message
|
encode_offset_commit_request
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, client_id, correlation_id, group, group_generation_id,\n consumer_id, payloads):\n grouped_payloads = group_by_topic_and_partition(payloads)\n message = cls._encode_message_header(client_id, correlation_id,\n KafkaCodec.OFFSET_COMMIT_KEY, api_version=1)\n message += write_short_ascii(group)\n message += struct.pack('>i', group_generation_id)\n message += write_short_ascii(consumer_id)\n message += struct.pack('>i', len(grouped_payloads))\n for topic, topic_payloads in grouped_payloads.items():\n message += write_short_ascii(topic)\n message += struct.pack('>i', len(topic_payloads))\n for partition, payload in topic_payloads.items():\n message += struct.pack('>iqq', partition, payload.offset,\n payload.timestamp)\n message += write_short_bytes(payload.metadata)\n return message\n",
"role": "user"
},
{
"content": "encode_offset_commit_request",
"role": "assistant"
}
] |
def x(cls, data):
(correlation_id,), cur = relative_unpack('>i', data, 0)
(num_topics,), cur = relative_unpack('>i', data, cur)
for _i in range(num_topics):
topic, cur = read_short_ascii(data, cur)
(num_partitions,), cur = relative_unpack('>i', data, cur)
for _i in range(num_partitions):
(partition, error), cur = relative_unpack('>ih', data, cur)
yield OffsetCommitResponse(topic, partition, error)
|
decode_offset_commit_response
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, data):\n (correlation_id,), cur = relative_unpack('>i', data, 0)\n (num_topics,), cur = relative_unpack('>i', data, cur)\n for _i in range(num_topics):\n topic, cur = read_short_ascii(data, cur)\n (num_partitions,), cur = relative_unpack('>i', data, cur)\n for _i in range(num_partitions):\n (partition, error), cur = relative_unpack('>ih', data, cur)\n yield OffsetCommitResponse(topic, partition, error)\n",
"role": "user"
},
{
"content": "decode_offset_commit_response",
"role": "assistant"
}
] |
def x(cls, client_id, correlation_id, group, payloads):
grouped_payloads = group_by_topic_and_partition(payloads)
message = cls._encode_message_header(client_id, correlation_id,
KafkaCodec.OFFSET_FETCH_KEY, api_version=1)
message += write_short_ascii(group)
message += struct.pack('>i', len(grouped_payloads))
for topic, topic_payloads in grouped_payloads.items():
message += write_short_ascii(topic)
message += struct.pack('>i', len(topic_payloads))
for partition in topic_payloads:
message += struct.pack('>i', partition)
return message
|
encode_offset_fetch_request
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, client_id, correlation_id, group, payloads):\n grouped_payloads = group_by_topic_and_partition(payloads)\n message = cls._encode_message_header(client_id, correlation_id,\n KafkaCodec.OFFSET_FETCH_KEY, api_version=1)\n message += write_short_ascii(group)\n message += struct.pack('>i', len(grouped_payloads))\n for topic, topic_payloads in grouped_payloads.items():\n message += write_short_ascii(topic)\n message += struct.pack('>i', len(topic_payloads))\n for partition in topic_payloads:\n message += struct.pack('>i', partition)\n return message\n",
"role": "user"
},
{
"content": "encode_offset_fetch_request",
"role": "assistant"
}
] |
def x(cls, data):
(correlation_id,), cur = relative_unpack('>i', data, 0)
(num_topics,), cur = relative_unpack('>i', data, cur)
for _i in range(num_topics):
topic, cur = read_short_ascii(data, cur)
(num_partitions,), cur = relative_unpack('>i', data, cur)
for _i in range(num_partitions):
(partition, offset), cur = relative_unpack('>iq', data, cur)
metadata, cur = read_short_bytes(data, cur)
(error,), cur = relative_unpack('>h', data, cur)
yield OffsetFetchResponse(topic, partition, offset, metadata, error
)
|
decode_offset_fetch_response
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, data):\n (correlation_id,), cur = relative_unpack('>i', data, 0)\n (num_topics,), cur = relative_unpack('>i', data, cur)\n for _i in range(num_topics):\n topic, cur = read_short_ascii(data, cur)\n (num_partitions,), cur = relative_unpack('>i', data, cur)\n for _i in range(num_partitions):\n (partition, offset), cur = relative_unpack('>iq', data, cur)\n metadata, cur = read_short_bytes(data, cur)\n (error,), cur = relative_unpack('>h', data, cur)\n yield OffsetFetchResponse(topic, partition, offset, metadata, error\n )\n",
"role": "user"
},
{
"content": "decode_offset_fetch_response",
"role": "assistant"
}
] |
def x(self, response):
correlation_id = response[0:4]
try:
d = self._pending.pop(correlation_id)
except KeyError:
self._log.warn(
'Response has unknown correlation ID {correlation_id!r}. Dropping connection to {peer}.'
, correlation_id=correlation_id, peer=self.transport.getPeer())
self.transport.loseConnection()
else:
d.callback(response)
|
stringReceived
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, response):\n correlation_id = response[0:4]\n try:\n d = self._pending.pop(correlation_id)\n except KeyError:\n self._log.warn(\n 'Response has unknown correlation ID {correlation_id!r}. Dropping connection to {peer}.'\n , correlation_id=correlation_id, peer=self.transport.getPeer())\n self.transport.loseConnection()\n else:\n d.callback(response)\n",
"role": "user"
},
{
"content": "stringReceived",
"role": "assistant"
}
] |
def x(self, reason=connectionDone):
self._failed = reason
pending, self._pending = self._pending, None
for d in pending.values():
d.errback(reason)
|
connectionLost
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, reason=connectionDone):\n self._failed = reason\n pending, self._pending = self._pending, None\n for d in pending.values():\n d.errback(reason)\n",
"role": "user"
},
{
"content": "connectionLost",
"role": "assistant"
}
] |
def x(self, request):
if self._failed is not None:
return fail(self._failed)
correlation_id = request[4:8]
assert correlation_id not in self._pending
d = Deferred()
self.sendString(request)
self._pending[correlation_id] = d
return d
|
request
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, request):\n if self._failed is not None:\n return fail(self._failed)\n correlation_id = request[4:8]\n assert correlation_id not in self._pending\n d = Deferred()\n self.sendString(request)\n self._pending[correlation_id] = d\n return d\n",
"role": "user"
},
{
"content": "request",
"role": "assistant"
}
] |
def x(self, topic, key=None, msgs=()):
try:
topic = _coerce_topic(topic)
if key is not None and not isinstance(key, bytes):
raise TypeError('key={!r} must be bytes or None'.format(key))
if not msgs:
raise ValueError('msgs must be a non-empty sequence')
msg_cnt = len(msgs)
byte_cnt = 0
for index, m in enumerate(msgs):
if m is None:
continue
if not isinstance(m, bytes):
raise TypeError(
'Message {} to topic {} ({!r:.100}) has type {}, but must have type {}'
.format(index, topic, m, type(m).__name__, type(bytes).
__name__))
byte_cnt += len(m)
except Exception:
return fail()
d = Deferred(self._cancel_send_messages)
self._batch_reqs.append(SendRequest(topic, key, msgs, d))
self._waitingMsgCount += msg_cnt
self._waitingByteCount += byte_cnt
self._outstanding.append(d)
d.addBoth(self._remove_from_outstanding, d)
self._check_send_batch()
return d
|
send_messages
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, topic, key=None, msgs=()):\n try:\n topic = _coerce_topic(topic)\n if key is not None and not isinstance(key, bytes):\n raise TypeError('key={!r} must be bytes or None'.format(key))\n if not msgs:\n raise ValueError('msgs must be a non-empty sequence')\n msg_cnt = len(msgs)\n byte_cnt = 0\n for index, m in enumerate(msgs):\n if m is None:\n continue\n if not isinstance(m, bytes):\n raise TypeError(\n 'Message {} to topic {} ({!r:.100}) has type {}, but must have type {}'\n .format(index, topic, m, type(m).__name__, type(bytes).\n __name__))\n byte_cnt += len(m)\n except Exception:\n return fail()\n d = Deferred(self._cancel_send_messages)\n self._batch_reqs.append(SendRequest(topic, key, msgs, d))\n self._waitingMsgCount += msg_cnt\n self._waitingByteCount += byte_cnt\n self._outstanding.append(d)\n d.addBoth(self._remove_from_outstanding, d)\n self._check_send_batch()\n return d\n",
"role": "user"
},
{
"content": "send_messages",
"role": "assistant"
}
] |
def x(self):
self.stopping = True
if self._batch_send_d:
self._batch_send_d.cancel()
if self.batch_every_t is not None:
if self._sendLooper is not None:
self._sendLooper.stop()
self._cancel_outstanding()
return self._sendLooperD or succeed(None)
|
stop
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n self.stopping = True\n if self._batch_send_d:\n self._batch_send_d.cancel()\n if self.batch_every_t is not None:\n if self._sendLooper is not None:\n self._sendLooper.stop()\n self._cancel_outstanding()\n return self._sendLooperD or succeed(None)\n",
"role": "user"
},
{
"content": "stop",
"role": "assistant"
}
] |
def x(self, fail):
log.warning('_send_timer_failed:%r: %s', fail, fail.getBriefTraceback())
self._sendLooperD = self._sendLooper.start(self.batch_every_t, now=False)
|
send_timer_failed
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, fail):\n log.warning('_send_timer_failed:%r: %s', fail, fail.getBriefTraceback())\n self._sendLooperD = self._sendLooper.start(self.batch_every_t, now=False)\n",
"role": "user"
},
{
"content": "send_timer_failed",
"role": "assistant"
}
] |
def x(self, lCall):
if self._sendLooper is not lCall:
log.warning('commitTimerStopped with wrong timer:%s not:%s', lCall,
self._sendLooper)
else:
self._sendLooper = None
self._sendLooperD = None
|
send_timer_stopped
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, lCall):\n if self._sendLooper is not lCall:\n log.warning('commitTimerStopped with wrong timer:%s not:%s', lCall,\n self._sendLooper)\n else:\n self._sendLooper = None\n self._sendLooperD = None\n",
"role": "user"
},
{
"content": "send_timer_stopped",
"role": "assistant"
}
] |
def x(self, topic, key=None):
while self.client.metadata_error_for_topic(topic):
if self._req_attempts >= self._max_attempts:
_check_error(self.client.metadata_error_for_topic(topic))
yield self.client.load_metadata_for_topics(topic)
if not self.client.metadata_error_for_topic(topic):
break
self._req_attempts += 1
d = Deferred()
self.client.reactor.callLater(self._retry_interval, d.callback, True)
self._retry_interval *= self.RETRY_INTERVAL_FACTOR
yield d
partitions = self.client.topic_partitions[topic]
if topic not in self.partitioners:
self.partitioners[topic] = self.partitioner_class(topic, partitions)
partition = self.partitioners[topic].partition(key, partitions)
returnValue(partition)
|
next_partition
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, topic, key=None):\n while self.client.metadata_error_for_topic(topic):\n if self._req_attempts >= self._max_attempts:\n _check_error(self.client.metadata_error_for_topic(topic))\n yield self.client.load_metadata_for_topics(topic)\n if not self.client.metadata_error_for_topic(topic):\n break\n self._req_attempts += 1\n d = Deferred()\n self.client.reactor.callLater(self._retry_interval, d.callback, True)\n self._retry_interval *= self.RETRY_INTERVAL_FACTOR\n yield d\n partitions = self.client.topic_partitions[topic]\n if topic not in self.partitioners:\n self.partitioners[topic] = self.partitioner_class(topic, partitions)\n partition = self.partitioners[topic].partition(key, partitions)\n returnValue(partition)\n",
"role": "user"
},
{
"content": "next_partition",
"role": "assistant"
}
] |
def x(self, parts_results, requests):
reqsByTopicPart = defaultdict(list)
payloadsByTopicPart = defaultdict(list)
deferredsByTopicPart = defaultdict(list)
for (success, part_or_failure), req in zip(parts_results, requests):
if req.deferred.called:
continue
if not success:
req.deferred.errback(part_or_failure)
continue
topicPart = TopicAndPartition(req.topic, part_or_failure)
reqsByTopicPart[topicPart].append(req)
deferredsByTopicPart[topicPart].append(req.deferred)
payloads = []
for (topic, partition), reqs in reqsByTopicPart.items():
msgSet = create_message_set(reqs, self.codec)
req = ProduceRequest(topic, partition, msgSet)
topicPart = TopicAndPartition(topic, partition)
payloads.append(req)
payloadsByTopicPart[topicPart] = req
if not payloads:
return
d = self.client.send_produce_request(payloads, acks=self.req_acks,
timeout=self.ack_timeout, fail_on_error=False)
self._req_attempts += 1
d.addBoth(self._handle_send_response, payloadsByTopicPart,
deferredsByTopicPart)
return d
|
send_requests
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, parts_results, requests):\n reqsByTopicPart = defaultdict(list)\n payloadsByTopicPart = defaultdict(list)\n deferredsByTopicPart = defaultdict(list)\n for (success, part_or_failure), req in zip(parts_results, requests):\n if req.deferred.called:\n continue\n if not success:\n req.deferred.errback(part_or_failure)\n continue\n topicPart = TopicAndPartition(req.topic, part_or_failure)\n reqsByTopicPart[topicPart].append(req)\n deferredsByTopicPart[topicPart].append(req.deferred)\n payloads = []\n for (topic, partition), reqs in reqsByTopicPart.items():\n msgSet = create_message_set(reqs, self.codec)\n req = ProduceRequest(topic, partition, msgSet)\n topicPart = TopicAndPartition(topic, partition)\n payloads.append(req)\n payloadsByTopicPart[topicPart] = req\n if not payloads:\n return\n d = self.client.send_produce_request(payloads, acks=self.req_acks,\n timeout=self.ack_timeout, fail_on_error=False)\n self._req_attempts += 1\n d.addBoth(self._handle_send_response, payloadsByTopicPart,\n deferredsByTopicPart)\n return d\n",
"role": "user"
},
{
"content": "send_requests",
"role": "assistant"
}
] |
def x(self, resp):
self._batch_send_d = None
self._req_attempts = 0
self._retry_interval = self._init_retry_interval
if isinstance(resp, Failure) and not resp.check(tid_CancelledError,
CancelledError):
log.error('Failure detected in _complete_batch_send: %r\n%r', resp,
resp.getTraceback())
return
|
complete_batch_send
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, resp):\n self._batch_send_d = None\n self._req_attempts = 0\n self._retry_interval = self._init_retry_interval\n if isinstance(resp, Failure) and not resp.check(tid_CancelledError,\n CancelledError):\n log.error('Failure detected in _complete_batch_send: %r\\n%r', resp,\n resp.getTraceback())\n return\n",
"role": "user"
},
{
"content": "complete_batch_send",
"role": "assistant"
}
] |
def x(self, result=None):
if (self.batch_every_n and self.batch_every_n <= self._waitingMsgCount or
self.batch_every_b and self.batch_every_b <= self._waitingByteCount):
self._send_batch()
return result
|
check_send_batch
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, result=None):\n if (self.batch_every_n and self.batch_every_n <= self._waitingMsgCount or\n self.batch_every_b and self.batch_every_b <= self._waitingByteCount):\n self._send_batch()\n return result\n",
"role": "user"
},
{
"content": "check_send_batch",
"role": "assistant"
}
] |
def x(self):
if not self._batch_reqs or self._batch_send_d:
return
requests, self._batch_reqs = self._batch_reqs, []
self._waitingByteCount = 0
self._waitingMsgCount = 0
d_list = []
for req in requests:
d_list.append(self._next_partition(req.topic, req.key))
d = self._batch_send_d = Deferred()
d.addCallback(lambda r: DeferredList(d_list, consumeErrors=True))
d.addCallback(self._send_requests, requests)
d.addBoth(self._complete_batch_send)
d.addBoth(self._check_send_batch)
d.callback(None)
|
send_batch
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if not self._batch_reqs or self._batch_send_d:\n return\n requests, self._batch_reqs = self._batch_reqs, []\n self._waitingByteCount = 0\n self._waitingMsgCount = 0\n d_list = []\n for req in requests:\n d_list.append(self._next_partition(req.topic, req.key))\n d = self._batch_send_d = Deferred()\n d.addCallback(lambda r: DeferredList(d_list, consumeErrors=True))\n d.addCallback(self._send_requests, requests)\n d.addBoth(self._complete_batch_send)\n d.addBoth(self._check_send_batch)\n d.callback(None)\n",
"role": "user"
},
{
"content": "send_batch",
"role": "assistant"
}
] |
def x(self, d):
for req in self._batch_reqs:
if req.deferred == d:
msgs = req.messages
self._waitingMsgCount -= len(msgs)
for m in (_m for _m in msgs if _m is not None):
self._waitingByteCount -= len(m)
self._batch_reqs.remove(req)
d.errback(CancelledError(request_sent=False))
return
d.errback(CancelledError(request_sent=self._batch_send_d is not None))
return
|
cancel_send_messages
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, d):\n for req in self._batch_reqs:\n if req.deferred == d:\n msgs = req.messages\n self._waitingMsgCount -= len(msgs)\n for m in (_m for _m in msgs if _m is not None):\n self._waitingByteCount -= len(m)\n self._batch_reqs.remove(req)\n d.errback(CancelledError(request_sent=False))\n return\n d.errback(CancelledError(request_sent=self._batch_send_d is not None))\n return\n",
"role": "user"
},
{
"content": "cancel_send_messages",
"role": "assistant"
}
] |
def x(self, result, payloadsByTopicPart, deferredsByTopicPart):
def x(d_list, result=None):
for d in d_list:
if not isinstance(d, Deferred):
_deliver_result(d, result)
elif not d.called:
d.callback(result)
def x(payloads):
d = self.client.send_produce_request(payloads, acks=self.req_acks,
timeout=self.ack_timeout, fail_on_error=False)
self._req_attempts += 1
d.addBoth(self._handle_send_response, payloadsByTopicPart,
deferredsByTopicPart)
return d
def x(failure, dc):
dc.cancel()
_deliver_result(deferredsByTopicPart.values(), failure)
return failure
def x(failed_payloads_with_errs):
if self._req_attempts >= self._max_attempts:
for p, f in failed_payloads_with_errs:
t_and_p = TopicAndPartition(p.topic, p.partition)
_deliver_result(deferredsByTopicPart[t_and_p], f)
return
d = Deferred()
dc = self.client.reactor.callLater(self._retry_interval, d.callback,
[p for p, f in failed_payloads])
self._retry_interval *= self.RETRY_INTERVAL_FACTOR
d.addErrback(_cancel_retry, dc)
reset_topics = set()
for payload, e in failed_payloads:
if isinstance(e, NotLeaderForPartitionError) or isinstance(e,
UnknownTopicOrPartitionError):
reset_topics.add(payload.topic)
if reset_topics:
self.client.reset_topic_metadata(*reset_topics)
d.addCallback(_do_retry)
return d
failed_payloads = []
if not result:
if self.req_acks == PRODUCER_ACK_NOT_REQUIRED:
result = None
else:
result = Failure(NoResponseError())
_deliver_result(deferredsByTopicPart.values(), result)
return
elif isinstance(result, Failure):
if not result.check(FailedPayloadsError):
if result.check(KafkaError):
failure, result = result, []
failed_payloads = [(p, failure) for p in
payloadsByTopicPart.values()]
else:
if not result.check(tid_CancelledError):
log.error('Unexpected failure: %r in _handle_send_response'
, result)
_deliver_result(deferredsByTopicPart.values(), result)
return
else:
failure = result
result = failure.value.args[0]
failed_payloads = failure.value.args[1]
for res in result:
t_and_p = TopicAndPartition(res.topic, res.partition)
t_and_p_err = _check_error(res, raiseException=False)
if not t_and_p_err:
d_list = deferredsByTopicPart[t_and_p]
_deliver_result(d_list, res)
else:
p = payloadsByTopicPart[t_and_p]
failed_payloads.append((p, t_and_p_err))
if failed_payloads:
return _check_retry_payloads(failed_payloads)
return
|
handle_send_response
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, result, payloadsByTopicPart, deferredsByTopicPart):\n\n def x(d_list, result=None):\n for d in d_list:\n if not isinstance(d, Deferred):\n _deliver_result(d, result)\n elif not d.called:\n d.callback(result)\n\n def x(payloads):\n d = self.client.send_produce_request(payloads, acks=self.req_acks,\n timeout=self.ack_timeout, fail_on_error=False)\n self._req_attempts += 1\n d.addBoth(self._handle_send_response, payloadsByTopicPart,\n deferredsByTopicPart)\n return d\n\n def x(failure, dc):\n dc.cancel()\n _deliver_result(deferredsByTopicPart.values(), failure)\n return failure\n\n def x(failed_payloads_with_errs):\n if self._req_attempts >= self._max_attempts:\n for p, f in failed_payloads_with_errs:\n t_and_p = TopicAndPartition(p.topic, p.partition)\n _deliver_result(deferredsByTopicPart[t_and_p], f)\n return\n d = Deferred()\n dc = self.client.reactor.callLater(self._retry_interval, d.callback,\n [p for p, f in failed_payloads])\n self._retry_interval *= self.RETRY_INTERVAL_FACTOR\n d.addErrback(_cancel_retry, dc)\n reset_topics = set()\n for payload, e in failed_payloads:\n if isinstance(e, NotLeaderForPartitionError) or isinstance(e,\n UnknownTopicOrPartitionError):\n reset_topics.add(payload.topic)\n if reset_topics:\n self.client.reset_topic_metadata(*reset_topics)\n d.addCallback(_do_retry)\n return d\n failed_payloads = []\n if not result:\n if self.req_acks == PRODUCER_ACK_NOT_REQUIRED:\n result = None\n else:\n result = Failure(NoResponseError())\n _deliver_result(deferredsByTopicPart.values(), result)\n return\n elif isinstance(result, Failure):\n if not result.check(FailedPayloadsError):\n if result.check(KafkaError):\n failure, result = result, []\n failed_payloads = [(p, failure) for p in\n payloadsByTopicPart.values()]\n else:\n if not result.check(tid_CancelledError):\n log.error('Unexpected failure: %r in _handle_send_response'\n , result)\n _deliver_result(deferredsByTopicPart.values(), result)\n return\n else:\n failure = result\n result = failure.value.args[0]\n failed_payloads = failure.value.args[1]\n for res in result:\n t_and_p = TopicAndPartition(res.topic, res.partition)\n t_and_p_err = _check_error(res, raiseException=False)\n if not t_and_p_err:\n d_list = deferredsByTopicPart[t_and_p]\n _deliver_result(d_list, res)\n else:\n p = payloadsByTopicPart[t_and_p]\n failed_payloads.append((p, t_and_p_err))\n if failed_payloads:\n return _check_retry_payloads(failed_payloads)\n return\n",
"role": "user"
},
{
"content": "handle_send_response",
"role": "assistant"
}
] |
def x(self):
for d in list(self._outstanding):
d.addErrback(lambda _: None)
d.cancel()
|
cancel_outstanding
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n for d in list(self._outstanding):\n d.addErrback(lambda _: None)\n d.cancel()\n",
"role": "user"
},
{
"content": "cancel_outstanding",
"role": "assistant"
}
] |
def x(hosts):
if isinstance(hosts, bytes):
hosts = hosts.split(b',')
elif isinstance(hosts, _unicode):
hosts = hosts.split(u',')
result = set()
for host_port in hosts:
res = nativeString(host_port).split(':')
host = res[0].strip()
port = int(res[1].strip()) if len(res) > 1 else DefaultKafkaPort
result.add((host, port))
return sorted(result)
|
normalize_hosts
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(hosts):\n if isinstance(hosts, bytes):\n hosts = hosts.split(b',')\n elif isinstance(hosts, _unicode):\n hosts = hosts.split(u',')\n result = set()\n for host_port in hosts:\n res = nativeString(host_port).split(':')\n host = res[0].strip()\n port = int(res[1].strip()) if len(res) > 1 else DefaultKafkaPort\n result.add((host, port))\n return sorted(result)\n",
"role": "user"
},
{
"content": "normalize_hosts",
"role": "assistant"
}
] |
def x(self, *groups):
groups = tuple(_coerce_consumer_group(g) for g in groups)
for group in groups:
if group in self.consumer_group_to_brokers:
del self.consumer_group_to_brokers[group]
|
reset_consumer_group_metadata
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, *groups):\n groups = tuple(_coerce_consumer_group(g) for g in groups)\n for group in groups:\n if group in self.consumer_group_to_brokers:\n del self.consumer_group_to_brokers[group]\n",
"role": "user"
},
{
"content": "reset_consumer_group_metadata",
"role": "assistant"
}
] |
def x(self):
self.topics_to_brokers.clear()
self.topic_partitions.clear()
self.topic_errors.clear()
self.consumer_group_to_brokers.clear()
|
reset_all_metadata
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n self.topics_to_brokers.clear()\n self.topic_partitions.clear()\n self.topic_errors.clear()\n self.consumer_group_to_brokers.clear()\n",
"role": "user"
},
{
"content": "reset_all_metadata",
"role": "assistant"
}
] |
def x(self, topic):
topic = _coerce_topic(topic)
if topic not in self.topic_partitions:
return False
if not self.topic_partitions[topic]:
return False
return all(self.partition_fully_replicated(TopicAndPartition(topic, p)) for
p in self.topic_partitions[topic])
|
topic_fully_replicated
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, topic):\n topic = _coerce_topic(topic)\n if topic not in self.topic_partitions:\n return False\n if not self.topic_partitions[topic]:\n return False\n return all(self.partition_fully_replicated(TopicAndPartition(topic, p)) for\n p in self.topic_partitions[topic])\n",
"role": "user"
},
{
"content": "topic_fully_replicated",
"role": "assistant"
}
] |
def x(self):
log.debug('%r: close', self)
self._closing = True
brokerclients, self.clients = self.clients, None
self._close_brokerclients(brokerclients.values())
self.reset_all_metadata()
return self.close_dlist or defer.succeed(None)
|
close
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n log.debug('%r: close', self)\n self._closing = True\n brokerclients, self.clients = self.clients, None\n self._close_brokerclients(brokerclients.values())\n self.reset_all_metadata()\n return self.close_dlist or defer.succeed(None)\n",
"role": "user"
},
{
"content": "close",
"role": "assistant"
}
] |
def x(self, *topics):
topics = tuple(_coerce_topic(t) for t in topics)
log.debug('%r: load_metadata_for_topics(%s)', self, ', '.join(repr(t) for
t in topics))
fetch_all_metadata = not topics
requestId = self._next_id()
request = KafkaCodec.encode_metadata_request(self._clientIdBytes,
requestId, topics)
def x(response):
brokers, topics = KafkaCodec.decode_metadata_response(response)
log.debug('%r: got metadata brokers=%r topics=%r', self, brokers,
topics)
if fetch_all_metadata:
self._brokers = brokers
self._topics = topics
ok_to_remove = fetch_all_metadata and len(brokers)
self._update_brokers(brokers.values(), remove=ok_to_remove)
for topic, topic_metadata in topics.items():
_, topic_error, partitions = topic_metadata
self.reset_topic_metadata(topic)
self.topic_errors[topic] = topic_error
if not partitions:
log.warning('No partitions for %s, Err:%d', topic, topic_error)
continue
self.topic_partitions[topic] = []
for partition, meta in partitions.items():
self.topic_partitions[topic].append(partition)
topic_part = TopicAndPartition(topic, partition)
self.partition_meta[topic_part] = meta
if meta.leader == -1:
log.warning('No leader for topic %s partition %s',
topic, partition)
self.topics_to_brokers[topic_part] = None
else:
self.topics_to_brokers[topic_part] = brokers[meta.leader]
self.topic_partitions[topic] = sorted(self.topic_partitions[topic])
return True
def x(err):
if err.check(t_CancelledError, CancelledError):
return None
log.error('Failed to retrieve metadata:%s', err)
raise KafkaUnavailableError(
'Unable to load metadata from configured hosts: {!r}'.format(err))
d = self._send_broker_unaware_request(requestId, request)
d.addCallbacks(_handleMetadataResponse, _handleMetadataErr)
return d
|
load_metadata_for_topics
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, *topics):\n topics = tuple(_coerce_topic(t) for t in topics)\n log.debug('%r: load_metadata_for_topics(%s)', self, ', '.join(repr(t) for\n t in topics))\n fetch_all_metadata = not topics\n requestId = self._next_id()\n request = KafkaCodec.encode_metadata_request(self._clientIdBytes,\n requestId, topics)\n\n def x(response):\n brokers, topics = KafkaCodec.decode_metadata_response(response)\n log.debug('%r: got metadata brokers=%r topics=%r', self, brokers,\n topics)\n if fetch_all_metadata:\n self._brokers = brokers\n self._topics = topics\n ok_to_remove = fetch_all_metadata and len(brokers)\n self._update_brokers(brokers.values(), remove=ok_to_remove)\n for topic, topic_metadata in topics.items():\n _, topic_error, partitions = topic_metadata\n self.reset_topic_metadata(topic)\n self.topic_errors[topic] = topic_error\n if not partitions:\n log.warning('No partitions for %s, Err:%d', topic, topic_error)\n continue\n self.topic_partitions[topic] = []\n for partition, meta in partitions.items():\n self.topic_partitions[topic].append(partition)\n topic_part = TopicAndPartition(topic, partition)\n self.partition_meta[topic_part] = meta\n if meta.leader == -1:\n log.warning('No leader for topic %s partition %s',\n topic, partition)\n self.topics_to_brokers[topic_part] = None\n else:\n self.topics_to_brokers[topic_part] = brokers[meta.leader]\n self.topic_partitions[topic] = sorted(self.topic_partitions[topic])\n return True\n\n def x(err):\n if err.check(t_CancelledError, CancelledError):\n return None\n log.error('Failed to retrieve metadata:%s', err)\n raise KafkaUnavailableError(\n 'Unable to load metadata from configured hosts: {!r}'.format(err))\n d = self._send_broker_unaware_request(requestId, request)\n d.addCallbacks(_handleMetadataResponse, _handleMetadataErr)\n return d\n",
"role": "user"
},
{
"content": "load_metadata_for_topics",
"role": "assistant"
}
] |
def x(self, group):
group = _coerce_consumer_group(group)
log.debug('%r: load_consumer_metadata_for_group(%r)', self, group)
if group in self.coordinator_fetches:
d = defer.Deferred()
self.coordinator_fetches[group][1].append(d)
return d
requestId = self._next_id()
request = KafkaCodec.encode_consumermetadata_request(self.
_clientIdBytes, requestId, group)
def x(response_bytes):
response = KafkaCodec.decode_consumermetadata_response(response_bytes)
log.debug('%r: load_consumer_metadata_for_group(%r) -> %r', self,
group, response)
if response.error:
raise BrokerResponseError.errnos.get(response.error, UnknownError)(
response)
bm = BrokerMetadata(response.node_id, response.host, response.port)
self.consumer_group_to_brokers[group] = bm
self._update_brokers([bm])
return True
def x(err):
log.error('Failed to retrieve consumer metadata for group %r',
group, exc_info=(err.type, err.value, err.getTracebackObject()))
self.reset_consumer_group_metadata(group)
raise ConsumerCoordinatorNotAvailableError(
'Coordinator for group {!r} not available'.format(group))
def x(result):
[_, ds] = self.coordinator_fetches.pop(group, None)
for d in ds:
d.callback(result)
request_d = self._send_broker_unaware_request(requestId, request)
d = defer.Deferred()
self.coordinator_fetches[group] = request_d, [d]
request_d.addCallback(_handleConsumerMetadataResponse)
request_d.addErrback(_handleConsumerMetadataErr)
request_d.addBoth(_propagate)
return d
|
load_consumer_metadata_for_group
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, group):\n group = _coerce_consumer_group(group)\n log.debug('%r: load_consumer_metadata_for_group(%r)', self, group)\n if group in self.coordinator_fetches:\n d = defer.Deferred()\n self.coordinator_fetches[group][1].append(d)\n return d\n requestId = self._next_id()\n request = KafkaCodec.encode_consumermetadata_request(self.\n _clientIdBytes, requestId, group)\n\n def x(response_bytes):\n response = KafkaCodec.decode_consumermetadata_response(response_bytes)\n log.debug('%r: load_consumer_metadata_for_group(%r) -> %r', self,\n group, response)\n if response.error:\n raise BrokerResponseError.errnos.get(response.error, UnknownError)(\n response)\n bm = BrokerMetadata(response.node_id, response.host, response.port)\n self.consumer_group_to_brokers[group] = bm\n self._update_brokers([bm])\n return True\n\n def x(err):\n log.error('Failed to retrieve consumer metadata for group %r',\n group, exc_info=(err.type, err.value, err.getTracebackObject()))\n self.reset_consumer_group_metadata(group)\n raise ConsumerCoordinatorNotAvailableError(\n 'Coordinator for group {!r} not available'.format(group))\n\n def x(result):\n [_, ds] = self.coordinator_fetches.pop(group, None)\n for d in ds:\n d.callback(result)\n request_d = self._send_broker_unaware_request(requestId, request)\n d = defer.Deferred()\n self.coordinator_fetches[group] = request_d, [d]\n request_d.addCallback(_handleConsumerMetadataResponse)\n request_d.addErrback(_handleConsumerMetadataErr)\n request_d.addBoth(_propagate)\n return d\n",
"role": "user"
},
{
"content": "load_consumer_metadata_for_group",
"role": "assistant"
}
] |
def x(self, payloads=None, acks=1, timeout=DEFAULT_REPLICAS_ACK_MSECS,
fail_on_error=True, callback=None):
encoder = partial(KafkaCodec.encode_produce_request, acks=acks, timeout
=timeout)
if acks == 0:
decoder = None
else:
decoder = KafkaCodec.decode_produce_response
resps = yield self._send_broker_aware_request(payloads, encoder, decoder)
returnValue(self._handle_responses(resps, fail_on_error, callback))
|
send_produce_request
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, payloads=None, acks=1, timeout=DEFAULT_REPLICAS_ACK_MSECS,\n fail_on_error=True, callback=None):\n encoder = partial(KafkaCodec.encode_produce_request, acks=acks, timeout\n =timeout)\n if acks == 0:\n decoder = None\n else:\n decoder = KafkaCodec.decode_produce_response\n resps = yield self._send_broker_aware_request(payloads, encoder, decoder)\n returnValue(self._handle_responses(resps, fail_on_error, callback))\n",
"role": "user"
},
{
"content": "send_produce_request",
"role": "assistant"
}
] |
def x(self, payloads=None, fail_on_error=True, callback=None, max_wait_time
=DEFAULT_FETCH_SERVER_WAIT_MSECS, min_bytes=DEFAULT_FETCH_MIN_BYTES):
if max_wait_time / 1000 > self.timeout - 0.1:
raise ValueError(
'%r: max_wait_time: %d must be less than client.timeout by at least 100 milliseconds.'
, self, max_wait_time)
encoder = partial(KafkaCodec.encode_fetch_request, max_wait_time=
max_wait_time, min_bytes=min_bytes)
resps = yield self._send_broker_aware_request(payloads, encoder,
KafkaCodec.decode_fetch_response)
returnValue(self._handle_responses(resps, fail_on_error, callback))
|
send_fetch_request
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, payloads=None, fail_on_error=True, callback=None, max_wait_time\n =DEFAULT_FETCH_SERVER_WAIT_MSECS, min_bytes=DEFAULT_FETCH_MIN_BYTES):\n if max_wait_time / 1000 > self.timeout - 0.1:\n raise ValueError(\n '%r: max_wait_time: %d must be less than client.timeout by at least 100 milliseconds.'\n , self, max_wait_time)\n encoder = partial(KafkaCodec.encode_fetch_request, max_wait_time=\n max_wait_time, min_bytes=min_bytes)\n resps = yield self._send_broker_aware_request(payloads, encoder,\n KafkaCodec.decode_fetch_response)\n returnValue(self._handle_responses(resps, fail_on_error, callback))\n",
"role": "user"
},
{
"content": "send_fetch_request",
"role": "assistant"
}
] |
def x(self, group, payloads=None, fail_on_error=True, callback=None):
encoder = partial(KafkaCodec.encode_offset_fetch_request, group=group)
decoder = KafkaCodec.decode_offset_fetch_response
resps = yield self._send_broker_aware_request(payloads, encoder,
decoder, consumer_group=group)
returnValue(self._handle_responses(resps, fail_on_error, callback, group))
|
send_offset_fetch_request
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, group, payloads=None, fail_on_error=True, callback=None):\n encoder = partial(KafkaCodec.encode_offset_fetch_request, group=group)\n decoder = KafkaCodec.decode_offset_fetch_response\n resps = yield self._send_broker_aware_request(payloads, encoder,\n decoder, consumer_group=group)\n returnValue(self._handle_responses(resps, fail_on_error, callback, group))\n",
"role": "user"
},
{
"content": "send_offset_fetch_request",
"role": "assistant"
}
] |
def x(self, group, payloads=None, fail_on_error=True, callback=None,
group_generation_id=-1, consumer_id=''):
group = _coerce_consumer_group(group)
encoder = partial(KafkaCodec.encode_offset_commit_request, group=group,
group_generation_id=group_generation_id, consumer_id=consumer_id)
decoder = KafkaCodec.decode_offset_commit_response
resps = yield self._send_broker_aware_request(payloads, encoder,
decoder, consumer_group=group)
returnValue(self._handle_responses(resps, fail_on_error, callback, group))
|
send_offset_commit_request
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, group, payloads=None, fail_on_error=True, callback=None,\n group_generation_id=-1, consumer_id=''):\n group = _coerce_consumer_group(group)\n encoder = partial(KafkaCodec.encode_offset_commit_request, group=group,\n group_generation_id=group_generation_id, consumer_id=consumer_id)\n decoder = KafkaCodec.decode_offset_commit_response\n resps = yield self._send_broker_aware_request(payloads, encoder,\n decoder, consumer_group=group)\n returnValue(self._handle_responses(resps, fail_on_error, callback, group))\n",
"role": "user"
},
{
"content": "send_offset_commit_request",
"role": "assistant"
}
] |
def x(self, node_id):
if self._closing:
raise ClientError(
'Cannot get broker client for node_id={}: {} has been closed'.
format(node_id, self))
if node_id not in self.clients:
broker_metadata = self._brokers[node_id]
log.debug('%r: creating client for %s', self, broker_metadata)
self.clients[node_id] = _KafkaBrokerClient(self.reactor, self.
_endpoint_factory, broker_metadata, self.clientId, self.
_retry_policy)
return self.clients[node_id]
|
get_brokerclient
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, node_id):\n if self._closing:\n raise ClientError(\n 'Cannot get broker client for node_id={}: {} has been closed'.\n format(node_id, self))\n if node_id not in self.clients:\n broker_metadata = self._brokers[node_id]\n log.debug('%r: creating client for %s', self, broker_metadata)\n self.clients[node_id] = _KafkaBrokerClient(self.reactor, self.\n _endpoint_factory, broker_metadata, self.clientId, self.\n _retry_policy)\n return self.clients[node_id]\n",
"role": "user"
},
{
"content": "get_brokerclient",
"role": "assistant"
}
] |
def x(self, clients):
def x(failure, brokerclient):
log.debug('BrokerClient: %s close result: %s: %s', brokerclient,
failure.type.__name__, failure.getErrorMessage())
def x(result, close_dlist):
if close_dlist == self.close_dlist:
self.close_dlist = None
if not self.close_dlist:
dList = []
else:
log.debug('%r: _close_brokerclients has nested deferredlist: %r',
self, self.close_dlist)
dList = [self.close_dlist]
for brokerClient in clients:
log.debug('Calling close on: %r', brokerClient)
d = brokerClient.close().addErrback(_log_close_failure, brokerClient)
dList.append(d)
self.close_dlist = DeferredList(dList)
self.close_dlist.addBoth(_clean_close_dlist, self.close_dlist)
|
close_brokerclients
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, clients):\n\n def x(failure, brokerclient):\n log.debug('BrokerClient: %s close result: %s: %s', brokerclient,\n failure.type.__name__, failure.getErrorMessage())\n\n def x(result, close_dlist):\n if close_dlist == self.close_dlist:\n self.close_dlist = None\n if not self.close_dlist:\n dList = []\n else:\n log.debug('%r: _close_brokerclients has nested deferredlist: %r',\n self, self.close_dlist)\n dList = [self.close_dlist]\n for brokerClient in clients:\n log.debug('Calling close on: %r', brokerClient)\n d = brokerClient.close().addErrback(_log_close_failure, brokerClient)\n dList.append(d)\n self.close_dlist = DeferredList(dList)\n self.close_dlist.addBoth(_clean_close_dlist, self.close_dlist)\n",
"role": "user"
},
{
"content": "close_brokerclients",
"role": "assistant"
}
] |
def x(self, brokers, remove=False):
log.debug('%r: _update_brokers(%r, remove=%r)', self, brokers, remove)
brokers_by_id = {bm.node_id: bm for bm in brokers}
self._brokers.update(brokers_by_id)
for node_id, broker_meta in brokers_by_id.items():
if node_id not in self.clients:
continue
self.clients[node_id].updateMetadata(broker_meta)
if remove:
to_close = [self.clients.pop(node_id) for node_id in set(self.
clients) - set(brokers_by_id)]
if to_close:
self._close_brokerclients(to_close)
|
update_brokers
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, brokers, remove=False):\n log.debug('%r: _update_brokers(%r, remove=%r)', self, brokers, remove)\n brokers_by_id = {bm.node_id: bm for bm in brokers}\n self._brokers.update(brokers_by_id)\n for node_id, broker_meta in brokers_by_id.items():\n if node_id not in self.clients:\n continue\n self.clients[node_id].updateMetadata(broker_meta)\n if remove:\n to_close = [self.clients.pop(node_id) for node_id in set(self.\n clients) - set(brokers_by_id)]\n if to_close:\n self._close_brokerclients(to_close)\n",
"role": "user"
},
{
"content": "update_brokers",
"role": "assistant"
}
] |
def x(self, topic, partition):
key = TopicAndPartition(topic, partition)
if self.topics_to_brokers.get(key) is None:
yield self.load_metadata_for_topics(topic)
if key not in self.topics_to_brokers:
raise PartitionUnavailableError('%s not available' % str(key))
returnValue(self.topics_to_brokers[key])
|
get_leader_for_partition
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, topic, partition):\n key = TopicAndPartition(topic, partition)\n if self.topics_to_brokers.get(key) is None:\n yield self.load_metadata_for_topics(topic)\n if key not in self.topics_to_brokers:\n raise PartitionUnavailableError('%s not available' % str(key))\n returnValue(self.topics_to_brokers[key])\n",
"role": "user"
},
{
"content": "get_leader_for_partition",
"role": "assistant"
}
] |
def x(self, consumer_group):
if self.consumer_group_to_brokers.get(consumer_group) is None:
yield self.load_consumer_metadata_for_group(consumer_group)
returnValue(self.consumer_group_to_brokers.get(consumer_group))
|
get_coordinator_for_group
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, consumer_group):\n if self.consumer_group_to_brokers.get(consumer_group) is None:\n yield self.load_consumer_metadata_for_group(consumer_group)\n returnValue(self.consumer_group_to_brokers.get(consumer_group))\n",
"role": "user"
},
{
"content": "get_coordinator_for_group",
"role": "assistant"
}
] |
def x(self, broker, requestId, request, **kwArgs):
def x(broker, requestId):
try:
broker.cancelRequest(requestId, reason=RequestTimedOutError(
'Request: {} cancelled due to timeout'.format(requestId)))
except KeyError:
log.exception(
'ERROR: Failed to find key for timed-out request. Broker: %r Req: %d'
, broker, requestId)
raise
if self._disconnect_on_timeout:
broker.disconnect()
def x(timeout, start):
now = self.reactor.seconds()
if now >= start + timeout:
log.warning('Reactor was starved for %r seconds', now - start)
def x(result, dc):
if dc.active():
dc.cancel()
return result
log.debug('_mrtb: sending request: %d to broker: %r', requestId, broker)
d = broker.makeRequest(requestId, request, **kwArgs)
dc = self.reactor.callLater(self.timeout, _timeout_request, broker,
requestId)
rc = self.reactor.callLater(self.timeout * 0.9, _alert_blocked_reactor,
self.timeout, self.reactor.seconds())
d.addBoth(_cancel_timeout, dc)
d.addBoth(_cancel_timeout, rc)
return d
|
make_request_to_broker
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, broker, requestId, request, **kwArgs):\n\n def x(broker, requestId):\n try:\n broker.cancelRequest(requestId, reason=RequestTimedOutError(\n 'Request: {} cancelled due to timeout'.format(requestId)))\n except KeyError:\n log.exception(\n 'ERROR: Failed to find key for timed-out request. Broker: %r Req: %d'\n , broker, requestId)\n raise\n if self._disconnect_on_timeout:\n broker.disconnect()\n\n def x(timeout, start):\n now = self.reactor.seconds()\n if now >= start + timeout:\n log.warning('Reactor was starved for %r seconds', now - start)\n\n def x(result, dc):\n if dc.active():\n dc.cancel()\n return result\n log.debug('_mrtb: sending request: %d to broker: %r', requestId, broker)\n d = broker.makeRequest(requestId, request, **kwArgs)\n dc = self.reactor.callLater(self.timeout, _timeout_request, broker,\n requestId)\n rc = self.reactor.callLater(self.timeout * 0.9, _alert_blocked_reactor,\n self.timeout, self.reactor.seconds())\n d.addBoth(_cancel_timeout, dc)\n d.addBoth(_cancel_timeout, rc)\n return d\n",
"role": "user"
},
{
"content": "make_request_to_broker",
"role": "assistant"
}
] |
def x(self, requestId, request):
node_ids = list(self._brokers.keys())
random.shuffle(node_ids)
def x(node_id):
try:
return self.clients[node_id].connected()
except KeyError:
return False
node_ids.sort(reverse=True, key=connected)
for node_id in node_ids:
broker = self._get_brokerclient(node_id)
try:
log.debug('_sbur: sending request %d to broker %r', requestId,
broker)
d = self._make_request_to_broker(broker, requestId, request)
resp = yield d
returnValue(resp)
except KafkaError as e:
log.warning(
'Will try next server after request with correlationId=%d failed against server %s:%i. Error: %s'
, requestId, broker.host, broker.port, e)
returnValue((yield self._send_bootstrap_request(request)))
|
send_broker_unaware_request
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, requestId, request):\n node_ids = list(self._brokers.keys())\n random.shuffle(node_ids)\n\n def x(node_id):\n try:\n return self.clients[node_id].connected()\n except KeyError:\n return False\n node_ids.sort(reverse=True, key=connected)\n for node_id in node_ids:\n broker = self._get_brokerclient(node_id)\n try:\n log.debug('_sbur: sending request %d to broker %r', requestId,\n broker)\n d = self._make_request_to_broker(broker, requestId, request)\n resp = yield d\n returnValue(resp)\n except KafkaError as e:\n log.warning(\n 'Will try next server after request with correlationId=%d failed against server %s:%i. Error: %s'\n , requestId, broker.host, broker.port, e)\n returnValue((yield self._send_bootstrap_request(request)))\n",
"role": "user"
},
{
"content": "send_broker_unaware_request",
"role": "assistant"
}
] |
def x(self, request):
hostports = list(self._bootstrap_hosts)
random.shuffle(hostports)
for host, port in hostports:
ep = self._endpoint_factory(self.reactor, host, port)
try:
protocol = yield ep.connect(_bootstrapFactory)
except Exception as e:
log.debug('%s: bootstrap connect to %s:%s -> %s', self, host,
port, e)
continue
try:
response = yield protocol.request(request).addTimeout(self.
timeout, self.reactor)
except Exception:
log.debug('%s: bootstrap request to %s:%s failed', self, host,
port, exc_info=True)
else:
returnValue(response)
finally:
protocol.transport.loseConnection()
raise KafkaUnavailableError('Failed to bootstrap from hosts {}'.format(
hostports))
|
send_bootstrap_request
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, request):\n hostports = list(self._bootstrap_hosts)\n random.shuffle(hostports)\n for host, port in hostports:\n ep = self._endpoint_factory(self.reactor, host, port)\n try:\n protocol = yield ep.connect(_bootstrapFactory)\n except Exception as e:\n log.debug('%s: bootstrap connect to %s:%s -> %s', self, host,\n port, e)\n continue\n try:\n response = yield protocol.request(request).addTimeout(self.\n timeout, self.reactor)\n except Exception:\n log.debug('%s: bootstrap request to %s:%s failed', self, host,\n port, exc_info=True)\n else:\n returnValue(response)\n finally:\n protocol.transport.loseConnection()\n raise KafkaUnavailableError('Failed to bootstrap from hosts {}'.format(\n hostports))\n",
"role": "user"
},
{
"content": "send_bootstrap_request",
"role": "assistant"
}
] |
def x(self, payloads, encoder_fn, decode_fn, consumer_group=None):
if not payloads:
raise ValueError('Payloads parameter is empty')
original_keys = []
payloads_by_broker = collections.defaultdict(list)
for payload in payloads:
if consumer_group is None:
leader = yield self._get_leader_for_partition(payload.topic,
payload.partition)
if leader is None:
raise LeaderUnavailableError(
'Leader not available for topic %s partition %s' % (
payload.topic, payload.partition))
else:
leader = yield self._get_coordinator_for_group(consumer_group)
if leader is None:
raise ConsumerCoordinatorNotAvailableError(
'Coordinator not available for group: %s' % consumer_group)
payloads_by_broker[leader].append(payload)
original_keys.append((payload.topic, payload.partition))
acc = {}
expectResponse = decode_fn is not None
failed_payloads = []
inFlight = []
payloadsList = []
for broker_meta, payloads in payloads_by_broker.items():
broker = self._get_brokerclient(broker_meta.node_id)
requestId = self._next_id()
request = encoder_fn(client_id=self._clientIdBytes, correlation_id=
requestId, payloads=payloads)
d = self._make_request_to_broker(broker, requestId, request,
expectResponse=expectResponse)
inFlight.append(d)
payloadsList.append(payloads)
results = yield DeferredList(inFlight, consumeErrors=True)
for (success, response), payloads in zip(results, payloadsList):
if not success:
log.debug('%r: request:%r to broker failed: %r', self, payloads,
response)
failed_payloads.extend([(p, response) for p in payloads])
continue
if not expectResponse:
continue
for response in decode_fn(response):
acc[response.topic, response.partition] = response
responses = [acc[k] for k in original_keys if k in acc] if acc else []
if failed_payloads:
self.reset_all_metadata()
raise FailedPayloadsError(responses, failed_payloads)
returnValue(responses)
|
send_broker_aware_request
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, payloads, encoder_fn, decode_fn, consumer_group=None):\n if not payloads:\n raise ValueError('Payloads parameter is empty')\n original_keys = []\n payloads_by_broker = collections.defaultdict(list)\n for payload in payloads:\n if consumer_group is None:\n leader = yield self._get_leader_for_partition(payload.topic,\n payload.partition)\n if leader is None:\n raise LeaderUnavailableError(\n 'Leader not available for topic %s partition %s' % (\n payload.topic, payload.partition))\n else:\n leader = yield self._get_coordinator_for_group(consumer_group)\n if leader is None:\n raise ConsumerCoordinatorNotAvailableError(\n 'Coordinator not available for group: %s' % consumer_group)\n payloads_by_broker[leader].append(payload)\n original_keys.append((payload.topic, payload.partition))\n acc = {}\n expectResponse = decode_fn is not None\n failed_payloads = []\n inFlight = []\n payloadsList = []\n for broker_meta, payloads in payloads_by_broker.items():\n broker = self._get_brokerclient(broker_meta.node_id)\n requestId = self._next_id()\n request = encoder_fn(client_id=self._clientIdBytes, correlation_id=\n requestId, payloads=payloads)\n d = self._make_request_to_broker(broker, requestId, request,\n expectResponse=expectResponse)\n inFlight.append(d)\n payloadsList.append(payloads)\n results = yield DeferredList(inFlight, consumeErrors=True)\n for (success, response), payloads in zip(results, payloadsList):\n if not success:\n log.debug('%r: request:%r to broker failed: %r', self, payloads,\n response)\n failed_payloads.extend([(p, response) for p in payloads])\n continue\n if not expectResponse:\n continue\n for response in decode_fn(response):\n acc[response.topic, response.partition] = response\n responses = [acc[k] for k in original_keys if k in acc] if acc else []\n if failed_payloads:\n self.reset_all_metadata()\n raise FailedPayloadsError(responses, failed_payloads)\n returnValue(responses)\n",
"role": "user"
},
{
"content": "send_broker_aware_request",
"role": "assistant"
}
] |
def x(expression, options=None):
descripter = ExpressionDescriptor(expression, options)
return descripter.get_description(DescriptionTypeEnum.FULL)
|
get_description
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(expression, options=None):\n descripter = ExpressionDescriptor(expression, options)\n return descripter.get_description(DescriptionTypeEnum.FULL)\n",
"role": "user"
},
{
"content": "get_description",
"role": "assistant"
}
] |
def x(self, description_type=DescriptionTypeEnum.FULL):
try:
if self._parsed is False:
parser = ExpressionParser(self._expression, self._options)
self._expression_parts = parser.parse()
self._parsed = True
choices = {DescriptionTypeEnum.FULL: self.get_full_description,
DescriptionTypeEnum.TIMEOFDAY: self.get_time_of_day_description,
DescriptionTypeEnum.HOURS: self.get_hours_description,
DescriptionTypeEnum.MINUTES: self.get_minutes_description,
DescriptionTypeEnum.SECONDS: self.get_seconds_description,
DescriptionTypeEnum.DAYOFMONTH: self.
get_day_of_month_description, DescriptionTypeEnum.MONTH: self.
get_month_description, DescriptionTypeEnum.DAYOFWEEK: self.
get_day_of_week_description, DescriptionTypeEnum.YEAR: self.
get_year_description}
description = choices.get(description_type, self.
get_seconds_description)()
except Exception as ex:
if self._options.throw_exception_on_parse_error:
raise
else:
description = str(ex)
return description
|
get_description
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, description_type=DescriptionTypeEnum.FULL):\n try:\n if self._parsed is False:\n parser = ExpressionParser(self._expression, self._options)\n self._expression_parts = parser.parse()\n self._parsed = True\n choices = {DescriptionTypeEnum.FULL: self.get_full_description,\n DescriptionTypeEnum.TIMEOFDAY: self.get_time_of_day_description,\n DescriptionTypeEnum.HOURS: self.get_hours_description,\n DescriptionTypeEnum.MINUTES: self.get_minutes_description,\n DescriptionTypeEnum.SECONDS: self.get_seconds_description,\n DescriptionTypeEnum.DAYOFMONTH: self.\n get_day_of_month_description, DescriptionTypeEnum.MONTH: self.\n get_month_description, DescriptionTypeEnum.DAYOFWEEK: self.\n get_day_of_week_description, DescriptionTypeEnum.YEAR: self.\n get_year_description}\n description = choices.get(description_type, self.\n get_seconds_description)()\n except Exception as ex:\n if self._options.throw_exception_on_parse_error:\n raise\n else:\n description = str(ex)\n return description\n",
"role": "user"
},
{
"content": "get_description",
"role": "assistant"
}
] |
def x(self):
try:
time_segment = self.get_time_of_day_description()
day_of_month_desc = self.get_day_of_month_description()
month_desc = self.get_month_description()
day_of_week_desc = self.get_day_of_week_description()
year_desc = self.get_year_description()
description = '{0}{1}{2}{3}{4}'.format(time_segment,
day_of_month_desc, day_of_week_desc, month_desc, year_desc)
description = self.transform_verbosity(description, self._options.
verbose)
description = self.transform_case(description, self._options.
casing_type)
except Exception:
description = _(
'An error occured when generating the expression description. Check the cron expression syntax.'
)
if self._options.throw_exception_on_parse_error:
raise FormatException(description)
return description
|
get_full_description
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n try:\n time_segment = self.get_time_of_day_description()\n day_of_month_desc = self.get_day_of_month_description()\n month_desc = self.get_month_description()\n day_of_week_desc = self.get_day_of_week_description()\n year_desc = self.get_year_description()\n description = '{0}{1}{2}{3}{4}'.format(time_segment,\n day_of_month_desc, day_of_week_desc, month_desc, year_desc)\n description = self.transform_verbosity(description, self._options.\n verbose)\n description = self.transform_case(description, self._options.\n casing_type)\n except Exception:\n description = _(\n 'An error occured when generating the expression description. Check the cron expression syntax.'\n )\n if self._options.throw_exception_on_parse_error:\n raise FormatException(description)\n return description\n",
"role": "user"
},
{
"content": "get_full_description",
"role": "assistant"
}
] |
def x(self):
seconds_expression = self._expression_parts[0]
minute_expression = self._expression_parts[1]
hour_expression = self._expression_parts[2]
description = StringBuilder()
if any(exp in minute_expression for exp in self._special_characters
) is False and any(exp in hour_expression for exp in self.
_special_characters) is False and any(exp in seconds_expression for
exp in self._special_characters) is False:
description.append(_('At '))
description.append(self.format_time(hour_expression,
minute_expression, seconds_expression))
elif '-' in minute_expression and ',' not in minute_expression and any(
exp in hour_expression for exp in self._special_characters) is False:
minute_parts = minute_expression.split('-')
description.append(_('Every minute between {0} and {1}').format(
self.format_time(hour_expression, minute_parts[0]), self.
format_time(hour_expression, minute_parts[1])))
elif ',' in hour_expression and '-' not in hour_expression and any(exp in
minute_expression for exp in self._special_characters) is False:
hour_parts = hour_expression.split(',')
description.append(_('At'))
for i, hour_part in enumerate(hour_parts):
description.append(' ')
description.append(self.format_time(hour_part, minute_expression))
if i < len(hour_parts) - 2:
description.append(',')
if i == len(hour_parts) - 2:
description.append(_(' and'))
else:
seconds_description = self.get_seconds_description()
minutes_description = self.get_minutes_description()
hours_description = self.get_hours_description()
description.append(seconds_description)
if description:
description.append(', ')
description.append(minutes_description)
if description:
description.append(', ')
description.append(hours_description)
return str(description)
|
get_time_of_day_description
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n seconds_expression = self._expression_parts[0]\n minute_expression = self._expression_parts[1]\n hour_expression = self._expression_parts[2]\n description = StringBuilder()\n if any(exp in minute_expression for exp in self._special_characters\n ) is False and any(exp in hour_expression for exp in self.\n _special_characters) is False and any(exp in seconds_expression for\n exp in self._special_characters) is False:\n description.append(_('At '))\n description.append(self.format_time(hour_expression,\n minute_expression, seconds_expression))\n elif '-' in minute_expression and ',' not in minute_expression and any(\n exp in hour_expression for exp in self._special_characters) is False:\n minute_parts = minute_expression.split('-')\n description.append(_('Every minute between {0} and {1}').format(\n self.format_time(hour_expression, minute_parts[0]), self.\n format_time(hour_expression, minute_parts[1])))\n elif ',' in hour_expression and '-' not in hour_expression and any(exp in\n minute_expression for exp in self._special_characters) is False:\n hour_parts = hour_expression.split(',')\n description.append(_('At'))\n for i, hour_part in enumerate(hour_parts):\n description.append(' ')\n description.append(self.format_time(hour_part, minute_expression))\n if i < len(hour_parts) - 2:\n description.append(',')\n if i == len(hour_parts) - 2:\n description.append(_(' and'))\n else:\n seconds_description = self.get_seconds_description()\n minutes_description = self.get_minutes_description()\n hours_description = self.get_hours_description()\n description.append(seconds_description)\n if description:\n description.append(', ')\n description.append(minutes_description)\n if description:\n description.append(', ')\n description.append(hours_description)\n return str(description)\n",
"role": "user"
},
{
"content": "get_time_of_day_description",
"role": "assistant"
}
] |
def x(self):
return self.get_segment_description(self._expression_parts[0], _(
'every second'), lambda s: s, lambda s: _('every {0} seconds').
format(s), lambda s: _('seconds {0} through {1} past the minute'),
lambda s: _('at {0} seconds past the minute'))
|
get_seconds_description
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n return self.get_segment_description(self._expression_parts[0], _(\n 'every second'), lambda s: s, lambda s: _('every {0} seconds').\n format(s), lambda s: _('seconds {0} through {1} past the minute'), \n lambda s: _('at {0} seconds past the minute'))\n",
"role": "user"
},
{
"content": "get_seconds_description",
"role": "assistant"
}
] |
def x(self):
return self.get_segment_description(self._expression_parts[1], _(
'every minute'), lambda s: s, lambda s: _('every {0} minutes').
format(s), lambda s: _('minutes {0} through {1} past the hour'), lambda
s: '' if s == '0' else _('at {0} minutes past the hour'))
|
get_minutes_description
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n return self.get_segment_description(self._expression_parts[1], _(\n 'every minute'), lambda s: s, lambda s: _('every {0} minutes').\n format(s), lambda s: _('minutes {0} through {1} past the hour'), lambda\n s: '' if s == '0' else _('at {0} minutes past the hour'))\n",
"role": "user"
},
{
"content": "get_minutes_description",
"role": "assistant"
}
] |
def x(self):
expression = self._expression_parts[2]
return self.get_segment_description(expression, _('every hour'), lambda
s: self.format_time(s, '0'), lambda s: _('every {0} hours').format(
s), lambda s: _('between {0} and {1}'), lambda s: _('at {0}'))
|
get_hours_description
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n expression = self._expression_parts[2]\n return self.get_segment_description(expression, _('every hour'), lambda\n s: self.format_time(s, '0'), lambda s: _('every {0} hours').format(\n s), lambda s: _('between {0} and {1}'), lambda s: _('at {0}'))\n",
"role": "user"
},
{
"content": "get_hours_description",
"role": "assistant"
}
] |
def x(self):
if self._expression_parts[5] == '*' and self._expression_parts[3] != '*':
return ''
def x(s):
exp = s
if '#' in s:
exp, useless = s.split('#', 2)
elif 'L' in s:
exp = exp.replace('L', '')
return self.number_to_day(int(exp))
def x(s):
if '#' in s:
day_of_week_of_month = s[s.find('#') + 1:]
try:
day_of_week_of_month_number = int(day_of_week_of_month)
choices = {(1): _('first'), (2): _('second'), (3): _(
'third'), (4): _('forth'), (5): _('fifth')}
day_of_week_of_month_description = choices.get(
day_of_week_of_month_number, '')
except ValueError:
day_of_week_of_month_description = ''
formated = '{}{}{}'.format(_(', on the '),
day_of_week_of_month_description, _(' {0} of the month'))
elif 'L' in s:
formated = _(', on the last {0} of the month')
else:
formated = _(', only on {0}')
return formated
return self.get_segment_description(self._expression_parts[5], _(
', every day'), lambda s: get_day_name(s), lambda s: _(
', every {0} days of the week').format(s), lambda s: _(
', {0} through {1}'), lambda s: get_format(s))
|
get_day_of_week_description
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self._expression_parts[5] == '*' and self._expression_parts[3] != '*':\n return ''\n\n def x(s):\n exp = s\n if '#' in s:\n exp, useless = s.split('#', 2)\n elif 'L' in s:\n exp = exp.replace('L', '')\n return self.number_to_day(int(exp))\n\n def x(s):\n if '#' in s:\n day_of_week_of_month = s[s.find('#') + 1:]\n try:\n day_of_week_of_month_number = int(day_of_week_of_month)\n choices = {(1): _('first'), (2): _('second'), (3): _(\n 'third'), (4): _('forth'), (5): _('fifth')}\n day_of_week_of_month_description = choices.get(\n day_of_week_of_month_number, '')\n except ValueError:\n day_of_week_of_month_description = ''\n formated = '{}{}{}'.format(_(', on the '),\n day_of_week_of_month_description, _(' {0} of the month'))\n elif 'L' in s:\n formated = _(', on the last {0} of the month')\n else:\n formated = _(', only on {0}')\n return formated\n return self.get_segment_description(self._expression_parts[5], _(\n ', every day'), lambda s: get_day_name(s), lambda s: _(\n ', every {0} days of the week').format(s), lambda s: _(\n ', {0} through {1}'), lambda s: get_format(s))\n",
"role": "user"
},
{
"content": "get_day_of_week_description",
"role": "assistant"
}
] |
def x(self):
return self.get_segment_description(self._expression_parts[4], '', lambda
s: datetime.date(datetime.date.today().year, int(s), 1).strftime(
'%B'), lambda s: _(', every {0} months').format(s), lambda s: _(
', {0} through {1}'), lambda s: _(', only in {0}'))
|
get_month_description
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n return self.get_segment_description(self._expression_parts[4], '', lambda\n s: datetime.date(datetime.date.today().year, int(s), 1).strftime(\n '%B'), lambda s: _(', every {0} months').format(s), lambda s: _(\n ', {0} through {1}'), lambda s: _(', only in {0}'))\n",
"role": "user"
},
{
"content": "get_month_description",
"role": "assistant"
}
] |
def x(self):
expression = self._expression_parts[3]
expression = expression.replace('?', '*')
if expression == 'L':
description = _(', on the last day of the month')
elif expression == 'LW' or expression == 'WL':
description = _(', on the last weekday of the month')
else:
regex = re.compile('(\\d{1,2}W)|(W\\d{1,2})')
if regex.match(expression):
m = regex.match(expression)
day_number = int(m.group().replace('W', ''))
day_string = _('first weekday') if day_number == 1 else _(
'weekday nearest day {0}').format(day_number)
description = _(', on the {0} of the month').format(day_string)
else:
description = self.get_segment_description(expression, _(
', every day'), lambda s: s, lambda s: _(', every day') if
s == '1' else _(', every {0} days'), lambda s: _(
', between day {0} and {1} of the month'), lambda s: _(
', on day {0} of the month'))
return description
|
get_day_of_month_description
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n expression = self._expression_parts[3]\n expression = expression.replace('?', '*')\n if expression == 'L':\n description = _(', on the last day of the month')\n elif expression == 'LW' or expression == 'WL':\n description = _(', on the last weekday of the month')\n else:\n regex = re.compile('(\\\\d{1,2}W)|(W\\\\d{1,2})')\n if regex.match(expression):\n m = regex.match(expression)\n day_number = int(m.group().replace('W', ''))\n day_string = _('first weekday') if day_number == 1 else _(\n 'weekday nearest day {0}').format(day_number)\n description = _(', on the {0} of the month').format(day_string)\n else:\n description = self.get_segment_description(expression, _(\n ', every day'), lambda s: s, lambda s: _(', every day') if \n s == '1' else _(', every {0} days'), lambda s: _(\n ', between day {0} and {1} of the month'), lambda s: _(\n ', on day {0} of the month'))\n return description\n",
"role": "user"
},
{
"content": "get_day_of_month_description",
"role": "assistant"
}
] |
def x(self):
def x(s):
regex = re.compile('^\\d+$')
if regex.match(s):
year_int = int(s)
if year_int < 1900:
return year_int
return datetime.date(year_int, 1, 1).strftime('%Y')
else:
return s
return self.get_segment_description(self._expression_parts[6], '', lambda
s: format_year(s), lambda s: _(', every {0} years').format(s), lambda
s: _(', {0} through {1}'), lambda s: _(', only in {0}'))
|
get_year_description
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n\n def x(s):\n regex = re.compile('^\\\\d+$')\n if regex.match(s):\n year_int = int(s)\n if year_int < 1900:\n return year_int\n return datetime.date(year_int, 1, 1).strftime('%Y')\n else:\n return s\n return self.get_segment_description(self._expression_parts[6], '', lambda\n s: format_year(s), lambda s: _(', every {0} years').format(s), lambda\n s: _(', {0} through {1}'), lambda s: _(', only in {0}'))\n",
"role": "user"
},
{
"content": "get_year_description",
"role": "assistant"
}
] |
def x(self, expression, all_description, get_single_item_description,
get_interval_description_format, get_between_description_format,
get_description_format):
description = None
if expression is None or expression == '':
description = ''
elif expression == '*':
description = all_description
elif any(ext in expression for ext in ['/', '-', ',']) is False:
description = get_description_format(expression).format(
get_single_item_description(expression))
elif '/' in expression:
segments = expression.split('/')
description = get_interval_description_format(segments[1]).format(
get_single_item_description(segments[1]))
if '-' in segments[0]:
between_segment_description = (self.
generate_between_segment_description(segments[0],
get_between_description_format, get_single_item_description))
if not between_segment_description.startswith(', '):
description += ', '
description += between_segment_description
elif any(ext in segments[0] for ext in ['*', ',']) is False:
range_item_description = get_description_format(segments[0]
).format(get_single_item_description(segments[0]))
range_item_description = range_item_description.replace(', ', '')
description += _(', starting {0}').format(range_item_description)
elif ',' in expression:
segments = expression.split(',')
description_content = ''
for i, segment in enumerate(segments):
if i > 0 and len(segments) > 2:
description_content += ','
if i < len(segments) - 1:
description_content += ' '
if i > 0 and len(segments) > 1 and (i == len(segments) - 1 or
len(segments) == 2):
description_content += _(' and ')
if '-' in segment:
between_description = (self.
generate_between_segment_description(segment, lambda s:
_(', {0} through {1}'), get_single_item_description))
between_description = between_description.replace(', ', '')
description_content += between_description
else:
description_content += get_single_item_description(segment)
description = get_description_format(expression).format(
description_content)
elif '-' in expression:
description = self.generate_between_segment_description(expression,
get_between_description_format, get_single_item_description)
return description
|
get_segment_description
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, expression, all_description, get_single_item_description,\n get_interval_description_format, get_between_description_format,\n get_description_format):\n description = None\n if expression is None or expression == '':\n description = ''\n elif expression == '*':\n description = all_description\n elif any(ext in expression for ext in ['/', '-', ',']) is False:\n description = get_description_format(expression).format(\n get_single_item_description(expression))\n elif '/' in expression:\n segments = expression.split('/')\n description = get_interval_description_format(segments[1]).format(\n get_single_item_description(segments[1]))\n if '-' in segments[0]:\n between_segment_description = (self.\n generate_between_segment_description(segments[0],\n get_between_description_format, get_single_item_description))\n if not between_segment_description.startswith(', '):\n description += ', '\n description += between_segment_description\n elif any(ext in segments[0] for ext in ['*', ',']) is False:\n range_item_description = get_description_format(segments[0]\n ).format(get_single_item_description(segments[0]))\n range_item_description = range_item_description.replace(', ', '')\n description += _(', starting {0}').format(range_item_description)\n elif ',' in expression:\n segments = expression.split(',')\n description_content = ''\n for i, segment in enumerate(segments):\n if i > 0 and len(segments) > 2:\n description_content += ','\n if i < len(segments) - 1:\n description_content += ' '\n if i > 0 and len(segments) > 1 and (i == len(segments) - 1 or \n len(segments) == 2):\n description_content += _(' and ')\n if '-' in segment:\n between_description = (self.\n generate_between_segment_description(segment, lambda s:\n _(', {0} through {1}'), get_single_item_description))\n between_description = between_description.replace(', ', '')\n description_content += between_description\n else:\n description_content += get_single_item_description(segment)\n description = get_description_format(expression).format(\n description_content)\n elif '-' in expression:\n description = self.generate_between_segment_description(expression,\n get_between_description_format, get_single_item_description)\n return description\n",
"role": "user"
},
{
"content": "get_segment_description",
"role": "assistant"
}
] |
def x(self, between_expression, get_between_description_format,
get_single_item_description):
description = ''
between_segments = between_expression.split('-')
between_segment_1_description = get_single_item_description(
between_segments[0])
between_segment_2_description = get_single_item_description(
between_segments[1])
between_segment_2_description = between_segment_2_description.replace(':00'
, ':59')
between_description_format = get_between_description_format(
between_expression)
description += between_description_format.format(
between_segment_1_description, between_segment_2_description)
return description
|
generate_between_segment_description
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, between_expression, get_between_description_format,\n get_single_item_description):\n description = ''\n between_segments = between_expression.split('-')\n between_segment_1_description = get_single_item_description(\n between_segments[0])\n between_segment_2_description = get_single_item_description(\n between_segments[1])\n between_segment_2_description = between_segment_2_description.replace(':00'\n , ':59')\n between_description_format = get_between_description_format(\n between_expression)\n description += between_description_format.format(\n between_segment_1_description, between_segment_2_description)\n return description\n",
"role": "user"
},
{
"content": "generate_between_segment_description",
"role": "assistant"
}
] |
def x(self, hour_expression, minute_expression, second_expression=''):
hour = int(hour_expression)
period = ''
if self._options.use_24hour_time_format is False:
period = ' PM' if hour >= 12 else ' AM'
if hour > 12:
hour -= 12
minute = str(int(minute_expression))
second = ''
if second_expression is not None and second_expression:
second = '{}{}'.format(':', str(int(second_expression)).zfill(2))
return '{0}:{1}{2}{3}'.format(str(hour).zfill(2), minute.zfill(2),
second, period)
|
format_time
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, hour_expression, minute_expression, second_expression=''):\n hour = int(hour_expression)\n period = ''\n if self._options.use_24hour_time_format is False:\n period = ' PM' if hour >= 12 else ' AM'\n if hour > 12:\n hour -= 12\n minute = str(int(minute_expression))\n second = ''\n if second_expression is not None and second_expression:\n second = '{}{}'.format(':', str(int(second_expression)).zfill(2))\n return '{0}:{1}{2}{3}'.format(str(hour).zfill(2), minute.zfill(2),\n second, period)\n",
"role": "user"
},
{
"content": "format_time",
"role": "assistant"
}
] |
def x(self, description, use_verbose_format):
if use_verbose_format is False:
description = description.replace(_(', every minute'), '')
description = description.replace(_(', every hour'), '')
description = description.replace(_(', every day'), '')
return description
|
transform_verbosity
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, description, use_verbose_format):\n if use_verbose_format is False:\n description = description.replace(_(', every minute'), '')\n description = description.replace(_(', every hour'), '')\n description = description.replace(_(', every day'), '')\n return description\n",
"role": "user"
},
{
"content": "transform_verbosity",
"role": "assistant"
}
] |
def x(self, description, case_type):
if case_type == CasingTypeEnum.Sentence:
description = '{}{}'.format(description[0].upper(), description[1:])
elif case_type == CasingTypeEnum.Title:
description = description.title()
else:
description = description.lower()
return description
|
transform_case
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, description, case_type):\n if case_type == CasingTypeEnum.Sentence:\n description = '{}{}'.format(description[0].upper(), description[1:])\n elif case_type == CasingTypeEnum.Title:\n description = description.title()\n else:\n description = description.lower()\n return description\n",
"role": "user"
},
{
"content": "transform_case",
"role": "assistant"
}
] |
def x(self, day_number):
return [calendar.day_name[6], calendar.day_name[0], calendar.day_name[1
], calendar.day_name[2], calendar.day_name[3], calendar.day_name[4],
calendar.day_name[5]][day_number]
|
number_to_day
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, day_number):\n return [calendar.day_name[6], calendar.day_name[0], calendar.day_name[1\n ], calendar.day_name[2], calendar.day_name[3], calendar.day_name[4],\n calendar.day_name[5]][day_number]\n",
"role": "user"
},
{
"content": "number_to_day",
"role": "assistant"
}
] |
def x(self, start_offset):
if self._start_d is not None:
raise RestartError('Start called on already-started consumer')
self._state = '[started]'
start_d = self._start_d = Deferred()
self._fetch_offset = start_offset
self._do_fetch()
if self.consumer_group and self.auto_commit_every_s:
self._commit_looper = LoopingCall(self._auto_commit)
self._commit_looper.clock = self.client.reactor
self._commit_looper_d = self._commit_looper.start(self.
auto_commit_every_s, now=False)
self._commit_looper_d.addCallbacks(self._commit_timer_stopped, self
._commit_timer_failed)
return start_d
|
start
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, start_offset):\n if self._start_d is not None:\n raise RestartError('Start called on already-started consumer')\n self._state = '[started]'\n start_d = self._start_d = Deferred()\n self._fetch_offset = start_offset\n self._do_fetch()\n if self.consumer_group and self.auto_commit_every_s:\n self._commit_looper = LoopingCall(self._auto_commit)\n self._commit_looper.clock = self.client.reactor\n self._commit_looper_d = self._commit_looper.start(self.\n auto_commit_every_s, now=False)\n self._commit_looper_d.addCallbacks(self._commit_timer_stopped, self\n ._commit_timer_failed)\n return start_d\n",
"role": "user"
},
{
"content": "start",
"role": "assistant"
}
] |
def x(self):
def x(result):
self._shutdown_d, d = None, self._shutdown_d
self.stop()
self._shuttingdown = False
d.callback((self._last_processed_offset, self._last_committed_offset))
def x(failure):
if failure.check(OperationInProgress):
failure.value.deferred.addCallback(_commit_and_stop)
return
self._shutdown_d, d = None, self._shutdown_d
self.stop()
self._shuttingdown = False
d.errback(failure)
def x(result):
if not self.consumer_group:
return _handle_shutdown_commit_success(None)
self.commit().addCallbacks(_handle_shutdown_commit_success,
_handle_shutdown_commit_failure)
if self._start_d is None:
return fail(Failure(RestopError(
'Shutdown called on non-running consumer')))
if self._shutdown_d:
return fail(Failure(RestopError('Shutdown called more than once.')))
self._shuttingdown = True
self._state = '[shutting down]'
self._shutdown_d = d = Deferred()
if self._processor_d:
self._processor_d.addCallback(_commit_and_stop)
else:
_commit_and_stop(None)
return d
|
shutdown
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n\n def x(result):\n self._shutdown_d, d = None, self._shutdown_d\n self.stop()\n self._shuttingdown = False\n d.callback((self._last_processed_offset, self._last_committed_offset))\n\n def x(failure):\n if failure.check(OperationInProgress):\n failure.value.deferred.addCallback(_commit_and_stop)\n return\n self._shutdown_d, d = None, self._shutdown_d\n self.stop()\n self._shuttingdown = False\n d.errback(failure)\n\n def x(result):\n if not self.consumer_group:\n return _handle_shutdown_commit_success(None)\n self.commit().addCallbacks(_handle_shutdown_commit_success,\n _handle_shutdown_commit_failure)\n if self._start_d is None:\n return fail(Failure(RestopError(\n 'Shutdown called on non-running consumer')))\n if self._shutdown_d:\n return fail(Failure(RestopError('Shutdown called more than once.')))\n self._shuttingdown = True\n self._state = '[shutting down]'\n self._shutdown_d = d = Deferred()\n if self._processor_d:\n self._processor_d.addCallback(_commit_and_stop)\n else:\n _commit_and_stop(None)\n return d\n",
"role": "user"
},
{
"content": "shutdown",
"role": "assistant"
}
] |
def x(self):
if self._start_d is None:
raise RestopError('Stop called on non-running consumer')
self._stopping = True
self._state = '[stopping]'
if self._request_d:
self._request_d.cancel()
if self._msg_block_d:
_msg_block_d, self._msg_block_d = self._msg_block_d, None
_msg_block_d.addErrback(lambda fail: fail.trap(CancelledError))
_msg_block_d.cancel()
if self._processor_d:
self._processor_d.cancel()
if self._retry_call:
self._retry_call.cancel()
if self._commit_ds:
while self._commit_ds:
d = self._commit_ds.pop()
d.cancel()
if self._commit_req:
self._commit_req.cancel()
if self._commit_call:
self._commit_call.cancel()
if self._commit_looper is not None:
self._commit_looper.stop()
self._stopping = False
self._state = '[stopped]'
self._start_d, d = None, self._start_d
if not d.called:
d.callback((self._last_processed_offset, self._last_committed_offset))
return self._last_processed_offset
|
stop
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self._start_d is None:\n raise RestopError('Stop called on non-running consumer')\n self._stopping = True\n self._state = '[stopping]'\n if self._request_d:\n self._request_d.cancel()\n if self._msg_block_d:\n _msg_block_d, self._msg_block_d = self._msg_block_d, None\n _msg_block_d.addErrback(lambda fail: fail.trap(CancelledError))\n _msg_block_d.cancel()\n if self._processor_d:\n self._processor_d.cancel()\n if self._retry_call:\n self._retry_call.cancel()\n if self._commit_ds:\n while self._commit_ds:\n d = self._commit_ds.pop()\n d.cancel()\n if self._commit_req:\n self._commit_req.cancel()\n if self._commit_call:\n self._commit_call.cancel()\n if self._commit_looper is not None:\n self._commit_looper.stop()\n self._stopping = False\n self._state = '[stopped]'\n self._start_d, d = None, self._start_d\n if not d.called:\n d.callback((self._last_processed_offset, self._last_committed_offset))\n return self._last_processed_offset\n",
"role": "user"
},
{
"content": "stop",
"role": "assistant"
}
] |
def x(self):
if not self.consumer_group:
return fail(Failure(InvalidConsumerGroupError('Bad Group_id:{0!r}'.
format(self.consumer_group))))
if (self._last_processed_offset is None or self._last_processed_offset ==
self._last_committed_offset):
return succeed(self._last_committed_offset)
if self._commit_ds:
d = Deferred()
self._commit_ds.append(d)
return fail(OperationInProgress(d))
d = Deferred()
self._commit_ds.append(d)
self._send_commit_request()
if self._commit_looper is not None:
self._commit_looper.reset()
return d
|
commit
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if not self.consumer_group:\n return fail(Failure(InvalidConsumerGroupError('Bad Group_id:{0!r}'.\n format(self.consumer_group))))\n if (self._last_processed_offset is None or self._last_processed_offset ==\n self._last_committed_offset):\n return succeed(self._last_committed_offset)\n if self._commit_ds:\n d = Deferred()\n self._commit_ds.append(d)\n return fail(OperationInProgress(d))\n d = Deferred()\n self._commit_ds.append(d)\n self._send_commit_request()\n if self._commit_looper is not None:\n self._commit_looper.reset()\n return d\n",
"role": "user"
},
{
"content": "commit",
"role": "assistant"
}
] |
def x(self, by_count=False):
if (self._stopping or self._shuttingdown or not self._start_d or self.
_last_processed_offset is None or not self.consumer_group or
by_count and not self.auto_commit_every_n):
return
if (not by_count or self._last_committed_offset is None or self.
_last_processed_offset - self._last_committed_offset >= self.
auto_commit_every_n):
if not self._commit_ds:
commit_d = self.commit()
commit_d.addErrback(self._handle_auto_commit_error)
else:
d = Deferred()
d.addCallback(self._retry_auto_commit, by_count)
self._commit_ds.append(d)
|
auto_commit
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, by_count=False):\n if (self._stopping or self._shuttingdown or not self._start_d or self.\n _last_processed_offset is None or not self.consumer_group or \n by_count and not self.auto_commit_every_n):\n return\n if (not by_count or self._last_committed_offset is None or self.\n _last_processed_offset - self._last_committed_offset >= self.\n auto_commit_every_n):\n if not self._commit_ds:\n commit_d = self.commit()\n commit_d.addErrback(self._handle_auto_commit_error)\n else:\n d = Deferred()\n d.addCallback(self._retry_auto_commit, by_count)\n self._commit_ds.append(d)\n",
"role": "user"
},
{
"content": "auto_commit",
"role": "assistant"
}
] |
def x(self, after=None):
if self._stopping or self._shuttingdown or self._start_d is None:
return
if self._retry_call is None:
if after is None:
after = self.retry_delay
self.retry_delay = min(self.retry_delay * REQUEST_RETRY_FACTOR,
self.retry_max_delay)
self._fetch_attempt_count += 1
self._retry_call = self.client.reactor.callLater(after, self._do_fetch)
|
retry_fetch
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, after=None):\n if self._stopping or self._shuttingdown or self._start_d is None:\n return\n if self._retry_call is None:\n if after is None:\n after = self.retry_delay\n self.retry_delay = min(self.retry_delay * REQUEST_RETRY_FACTOR,\n self.retry_max_delay)\n self._fetch_attempt_count += 1\n self._retry_call = self.client.reactor.callLater(after, self._do_fetch)\n",
"role": "user"
},
{
"content": "retry_fetch",
"role": "assistant"
}
] |
def x(self, response):
self._request_d = None
self.retry_delay = self.retry_init_delay
self._fetch_attempt_count = 1
response = response[0]
if hasattr(response, 'offsets'):
self._fetch_offset = response.offsets[0]
elif response.offset == OFFSET_NOT_COMMITTED:
self._fetch_offset = OFFSET_EARLIEST
else:
self._fetch_offset = response.offset + 1
self._last_committed_offset = response.offset
self._do_fetch()
|
handle_offset_response
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, response):\n self._request_d = None\n self.retry_delay = self.retry_init_delay\n self._fetch_attempt_count = 1\n response = response[0]\n if hasattr(response, 'offsets'):\n self._fetch_offset = response.offsets[0]\n elif response.offset == OFFSET_NOT_COMMITTED:\n self._fetch_offset = OFFSET_EARLIEST\n else:\n self._fetch_offset = response.offset + 1\n self._last_committed_offset = response.offset\n self._do_fetch()\n",
"role": "user"
},
{
"content": "handle_offset_response",
"role": "assistant"
}
] |
def x(self, failure):
self._request_d = None
if self._stopping and failure.check(CancelledError):
return
if (self.request_retry_max_attempts != 0 and self._fetch_attempt_count >=
self.request_retry_max_attempts):
log.debug('%r: Exhausted attempts: %d fetching offset from kafka: %r',
self, self.request_retry_max_attempts, failure)
self._start_d.errback(failure)
return
if (self.retry_delay < self.retry_max_delay or 0 == self.
_fetch_attempt_count % 2):
log.debug('%r: Failure fetching offset from kafka: %r', self, failure)
else:
log.warning('%r: Still failing fetching offset from kafka: %r',
self, failure)
self._retry_fetch()
|
handle_offset_error
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, failure):\n self._request_d = None\n if self._stopping and failure.check(CancelledError):\n return\n if (self.request_retry_max_attempts != 0 and self._fetch_attempt_count >=\n self.request_retry_max_attempts):\n log.debug('%r: Exhausted attempts: %d fetching offset from kafka: %r',\n self, self.request_retry_max_attempts, failure)\n self._start_d.errback(failure)\n return\n if (self.retry_delay < self.retry_max_delay or 0 == self.\n _fetch_attempt_count % 2):\n log.debug('%r: Failure fetching offset from kafka: %r', self, failure)\n else:\n log.warning('%r: Still failing fetching offset from kafka: %r',\n self, failure)\n self._retry_fetch()\n",
"role": "user"
},
{
"content": "handle_offset_error",
"role": "assistant"
}
] |
def x(self, retry_delay=None, attempt=None):
if self._commit_call and not self._commit_call.active():
self._commit_call = None
if self._commit_req is not None:
raise OperationInProgress(self._commit_req)
if retry_delay is None:
retry_delay = self.retry_init_delay
if attempt is None:
attempt = 1
commit_offset = self._last_processed_offset
commit_request = OffsetCommitRequest(self.topic, self.partition,
commit_offset, TIMESTAMP_INVALID, self.commit_metadata)
log.debug('Committing off=%d grp=%s tpc=%s part=%s req=%r', self.
_last_processed_offset, self.consumer_group, self.topic, self.
partition, commit_request)
self._commit_req = d = self.client.send_offset_commit_request(self.
consumer_group, [commit_request])
d.addBoth(self._clear_commit_req)
d.addCallbacks(self._update_committed_offset, self._handle_commit_error,
callbackArgs=(commit_offset,), errbackArgs=(retry_delay, attempt))
|
send_commit_request
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, retry_delay=None, attempt=None):\n if self._commit_call and not self._commit_call.active():\n self._commit_call = None\n if self._commit_req is not None:\n raise OperationInProgress(self._commit_req)\n if retry_delay is None:\n retry_delay = self.retry_init_delay\n if attempt is None:\n attempt = 1\n commit_offset = self._last_processed_offset\n commit_request = OffsetCommitRequest(self.topic, self.partition,\n commit_offset, TIMESTAMP_INVALID, self.commit_metadata)\n log.debug('Committing off=%d grp=%s tpc=%s part=%s req=%r', self.\n _last_processed_offset, self.consumer_group, self.topic, self.\n partition, commit_request)\n self._commit_req = d = self.client.send_offset_commit_request(self.\n consumer_group, [commit_request])\n d.addBoth(self._clear_commit_req)\n d.addCallbacks(self._update_committed_offset, self._handle_commit_error,\n callbackArgs=(commit_offset,), errbackArgs=(retry_delay, attempt))\n",
"role": "user"
},
{
"content": "send_commit_request",
"role": "assistant"
}
] |
def x(self, failure, retry_delay, attempt):
if self._stopping and failure.check(CancelledError):
return self._deliver_commit_result(self._last_committed_offset)
if not failure.check(KafkaError):
log.error('Unhandleable failure during commit attempt: %r\n\t%r',
failure, failure.getBriefTraceback())
return self._deliver_commit_result(failure)
if (self.request_retry_max_attempts != 0 and attempt >= self.
request_retry_max_attempts):
log.debug('%r: Exhausted attempts: %d to commit offset: %r', self,
self.request_retry_max_attempts, failure)
return self._deliver_commit_result(failure)
if retry_delay < self.retry_max_delay or 0 == attempt % 2:
log.debug('%r: Failure committing offset to kafka: %r', self, failure)
else:
log.warning('%r: Still failing committing offset to kafka: %r',
self, failure)
retry_delay = min(retry_delay * REQUEST_RETRY_FACTOR, self.retry_max_delay)
self._commit_call = self.client.reactor.callLater(retry_delay, self.
_send_commit_request, retry_delay, attempt + 1)
|
handle_commit_error
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, failure, retry_delay, attempt):\n if self._stopping and failure.check(CancelledError):\n return self._deliver_commit_result(self._last_committed_offset)\n if not failure.check(KafkaError):\n log.error('Unhandleable failure during commit attempt: %r\\n\\t%r',\n failure, failure.getBriefTraceback())\n return self._deliver_commit_result(failure)\n if (self.request_retry_max_attempts != 0 and attempt >= self.\n request_retry_max_attempts):\n log.debug('%r: Exhausted attempts: %d to commit offset: %r', self,\n self.request_retry_max_attempts, failure)\n return self._deliver_commit_result(failure)\n if retry_delay < self.retry_max_delay or 0 == attempt % 2:\n log.debug('%r: Failure committing offset to kafka: %r', self, failure)\n else:\n log.warning('%r: Still failing committing offset to kafka: %r',\n self, failure)\n retry_delay = min(retry_delay * REQUEST_RETRY_FACTOR, self.retry_max_delay)\n self._commit_call = self.client.reactor.callLater(retry_delay, self.\n _send_commit_request, retry_delay, attempt + 1)\n",
"role": "user"
},
{
"content": "handle_commit_error",
"role": "assistant"
}
] |
def x(self, failure):
if not (self._stopping and failure.check(CancelledError)):
if self._start_d:
self._start_d.errback(failure)
|
handle_processor_error
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, failure):\n if not (self._stopping and failure.check(CancelledError)):\n if self._start_d:\n self._start_d.errback(failure)\n",
"role": "user"
},
{
"content": "handle_processor_error",
"role": "assistant"
}
] |
def x(self, failure):
self._request_d = None
if failure.check(OffsetOutOfRangeError):
if self.auto_offset_reset is None:
self._start_d.errback(failure)
return
self._fetch_offset = self.auto_offset_reset
if self._stopping and failure.check(CancelledError):
return
if (self.request_retry_max_attempts != 0 and self._fetch_attempt_count >=
self.request_retry_max_attempts):
log.debug('%r: Exhausted attempts: %d fetching messages from kafka: %r'
, self, self.request_retry_max_attempts, failure)
self._start_d.errback(failure)
return
if (self.retry_delay < self.retry_max_delay or 0 == self.
_fetch_attempt_count % 2):
log.debug('%r: Failure fetching messages from kafka: %r', self, failure
)
else:
log.warning('%r: Still failing fetching messages from kafka: %r',
self, failure)
self._retry_fetch()
|
handle_fetch_error
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, failure):\n self._request_d = None\n if failure.check(OffsetOutOfRangeError):\n if self.auto_offset_reset is None:\n self._start_d.errback(failure)\n return\n self._fetch_offset = self.auto_offset_reset\n if self._stopping and failure.check(CancelledError):\n return\n if (self.request_retry_max_attempts != 0 and self._fetch_attempt_count >=\n self.request_retry_max_attempts):\n log.debug('%r: Exhausted attempts: %d fetching messages from kafka: %r'\n , self, self.request_retry_max_attempts, failure)\n self._start_d.errback(failure)\n return\n if (self.retry_delay < self.retry_max_delay or 0 == self.\n _fetch_attempt_count % 2):\n log.debug('%r: Failure fetching messages from kafka: %r', self, failure\n )\n else:\n log.warning('%r: Still failing fetching messages from kafka: %r',\n self, failure)\n self._retry_fetch()\n",
"role": "user"
},
{
"content": "handle_fetch_error",
"role": "assistant"
}
] |
def x(self, responses):
self.retry_delay = self.retry_init_delay
self._fetch_attempt_count = 1
if self._msg_block_d:
self._msg_block_d.addCallback(lambda _: self._handle_fetch_response
(responses))
return
self._request_d = None
messages = []
try:
for resp in responses:
if resp.partition != self.partition:
log.warning(
'%r: Got response with partition: %r not our own: %r',
self, resp.partition, self.partition)
continue
for message in resp.messages:
if message.offset < self._fetch_offset:
log.debug(
'Skipping message at offset: %d, because its offset is less that our fetch offset: %d.'
, message.offset, self._fetch_offset)
continue
messages.append(SourcedMessage(message=message.message,
offset=message.offset, topic=self.topic, partition=self
.partition))
self._fetch_offset = message.offset + 1
except ConsumerFetchSizeTooSmall:
factor = 2
if self.buffer_size <= 2 ** 20:
factor = 16
if self.max_buffer_size is None:
self.buffer_size *= factor
elif self.max_buffer_size is not None and self.buffer_size < self.max_buffer_size:
self.buffer_size = min(self.buffer_size * factor, self.
max_buffer_size)
else:
log.error('Max fetch size %d too small', self.max_buffer_size)
failure = Failure(ConsumerFetchSizeTooSmall(
'Max buffer size:%d too small for message', self.
max_buffer_size))
self._start_d.errback(failure)
return
log.debug(
'Next message larger than fetch size, increasing to %d (~2x) and retrying'
, self.buffer_size)
finally:
if messages:
self._msg_block_d = Deferred()
self._process_messages(messages)
self._retry_fetch(0)
|
handle_fetch_response
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, responses):\n self.retry_delay = self.retry_init_delay\n self._fetch_attempt_count = 1\n if self._msg_block_d:\n self._msg_block_d.addCallback(lambda _: self._handle_fetch_response\n (responses))\n return\n self._request_d = None\n messages = []\n try:\n for resp in responses:\n if resp.partition != self.partition:\n log.warning(\n '%r: Got response with partition: %r not our own: %r',\n self, resp.partition, self.partition)\n continue\n for message in resp.messages:\n if message.offset < self._fetch_offset:\n log.debug(\n 'Skipping message at offset: %d, because its offset is less that our fetch offset: %d.'\n , message.offset, self._fetch_offset)\n continue\n messages.append(SourcedMessage(message=message.message,\n offset=message.offset, topic=self.topic, partition=self\n .partition))\n self._fetch_offset = message.offset + 1\n except ConsumerFetchSizeTooSmall:\n factor = 2\n if self.buffer_size <= 2 ** 20:\n factor = 16\n if self.max_buffer_size is None:\n self.buffer_size *= factor\n elif self.max_buffer_size is not None and self.buffer_size < self.max_buffer_size:\n self.buffer_size = min(self.buffer_size * factor, self.\n max_buffer_size)\n else:\n log.error('Max fetch size %d too small', self.max_buffer_size)\n failure = Failure(ConsumerFetchSizeTooSmall(\n 'Max buffer size:%d too small for message', self.\n max_buffer_size))\n self._start_d.errback(failure)\n return\n log.debug(\n 'Next message larger than fetch size, increasing to %d (~2x) and retrying'\n , self.buffer_size)\n finally:\n if messages:\n self._msg_block_d = Deferred()\n self._process_messages(messages)\n self._retry_fetch(0)\n",
"role": "user"
},
{
"content": "handle_fetch_response",
"role": "assistant"
}
] |
def x(self, messages):
if self._shuttingdown:
return
if not messages:
if self._msg_block_d:
_msg_block_d, self._msg_block_d = self._msg_block_d, None
_msg_block_d.callback(True)
return
proc_block_size = sys.maxsize
if self.auto_commit_every_n:
proc_block_size = self.auto_commit_every_n
msgs_to_proc = messages[:proc_block_size]
msgs_remainder = messages[proc_block_size:]
last_offset = msgs_to_proc[-1].offset
self._processor_d = d = maybeDeferred(self.processor, self, msgs_to_proc)
log.debug('self.processor return: %r, last_offset: %r', d, last_offset)
d.addBoth(self._clear_processor_deferred)
d.addCallback(self._update_processed_offset, last_offset)
if self._stopping or self._start_d is None:
d.cancel()
else:
d.addCallback(lambda _: self._process_messages(msgs_remainder))
d.addErrback(self._handle_processor_error)
|
process_messages
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, messages):\n if self._shuttingdown:\n return\n if not messages:\n if self._msg_block_d:\n _msg_block_d, self._msg_block_d = self._msg_block_d, None\n _msg_block_d.callback(True)\n return\n proc_block_size = sys.maxsize\n if self.auto_commit_every_n:\n proc_block_size = self.auto_commit_every_n\n msgs_to_proc = messages[:proc_block_size]\n msgs_remainder = messages[proc_block_size:]\n last_offset = msgs_to_proc[-1].offset\n self._processor_d = d = maybeDeferred(self.processor, self, msgs_to_proc)\n log.debug('self.processor return: %r, last_offset: %r', d, last_offset)\n d.addBoth(self._clear_processor_deferred)\n d.addCallback(self._update_processed_offset, last_offset)\n if self._stopping or self._start_d is None:\n d.cancel()\n else:\n d.addCallback(lambda _: self._process_messages(msgs_remainder))\n d.addErrback(self._handle_processor_error)\n",
"role": "user"
},
{
"content": "process_messages",
"role": "assistant"
}
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.