source_code stringlengths 45 7.24k | function_name stringlengths 1 111 | messages list |
|---|---|---|
def x(self, name):
address = _create_role_address(name)
role_list_bytes = None
try:
role_list_bytes = self._state_view.get(address=address)
except KeyError:
return None
if role_list_bytes is not None:
role_list = _create_from_bytes(role_list_bytes, identity_pb2.RoleList)
for role in role_list.roles:
if role.name == name:
return role
return None
| get_role | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, name):\n address = _create_role_address(name)\n role_list_bytes = None\n try:\n role_list_bytes = self._state_view.get(address=address)\n except KeyError:\n return None\n if role_list_bytes is not None:\n role_list = _create_from_bytes(role_list_bytes, identity_pb2.RoleList)\n for role in role_list.roles:\n if role.name == name:\n return role\n return None\n",
"role": "user"
},
{
"content": "get_role",
"role": "assistant"
}
] |
def x(self):
prefix = _IDENTITY_NS + _ROLE_NS
rolelist_list = [_create_from_bytes(d, identity_pb2.RoleList) for _, d in
self._state_view.leaves(prefix=prefix)]
roles = []
for role_list in rolelist_list:
for role in role_list.roles:
roles.append(role)
return sorted(roles, key=lambda r: r.name)
| get_roles | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n prefix = _IDENTITY_NS + _ROLE_NS\n rolelist_list = [_create_from_bytes(d, identity_pb2.RoleList) for _, d in\n self._state_view.leaves(prefix=prefix)]\n roles = []\n for role_list in rolelist_list:\n for role in role_list.roles:\n roles.append(role)\n return sorted(roles, key=lambda r: r.name)\n",
"role": "user"
},
{
"content": "get_roles",
"role": "assistant"
}
] |
def x(self, name):
address = _create_policy_address(name)
policy_list_bytes = None
try:
policy_list_bytes = self._state_view.get(address=address)
except KeyError:
return None
if policy_list_bytes is not None:
policy_list = _create_from_bytes(policy_list_bytes, identity_pb2.
PolicyList)
for policy in policy_list.policies:
if policy.name == name:
return policy
return None
| get_policy | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, name):\n address = _create_policy_address(name)\n policy_list_bytes = None\n try:\n policy_list_bytes = self._state_view.get(address=address)\n except KeyError:\n return None\n if policy_list_bytes is not None:\n policy_list = _create_from_bytes(policy_list_bytes, identity_pb2.\n PolicyList)\n for policy in policy_list.policies:\n if policy.name == name:\n return policy\n return None\n",
"role": "user"
},
{
"content": "get_policy",
"role": "assistant"
}
] |
def x(self):
prefix = _IDENTITY_NS + _POLICY_NS
policylist_list = [_create_from_bytes(d, identity_pb2.PolicyList) for _,
d in self._state_view.leaves(prefix=prefix)]
policies = []
for policy_list in policylist_list:
for policy in policy_list.policies:
policies.append(policy)
return sorted(policies, key=lambda p: p.name)
| get_policies | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n prefix = _IDENTITY_NS + _POLICY_NS\n policylist_list = [_create_from_bytes(d, identity_pb2.PolicyList) for _,\n d in self._state_view.leaves(prefix=prefix)]\n policies = []\n for policy_list in policylist_list:\n for policy in policy_list.policies:\n policies.append(policy)\n return sorted(policies, key=lambda p: p.name)\n",
"role": "user"
},
{
"content": "get_policies",
"role": "assistant"
}
] |
def x(self, previous_block):
previous_header_bytes = previous_block.header
previous_header = BlockHeader()
previous_header.ParseFromString(previous_header_bytes)
block_info = BlockInfo(block_num=previous_header.block_num,
previous_block_id=previous_header.previous_block_id,
signer_public_key=previous_header.signer_public_key,
header_signature=previous_block.header_signature, timestamp=int(
time.time()))
return [self.create_batch(block_info)]
| block_start | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, previous_block):\n previous_header_bytes = previous_block.header\n previous_header = BlockHeader()\n previous_header.ParseFromString(previous_header_bytes)\n block_info = BlockInfo(block_num=previous_header.block_num,\n previous_block_id=previous_header.previous_block_id,\n signer_public_key=previous_header.signer_public_key,\n header_signature=previous_block.header_signature, timestamp=int(\n time.time()))\n return [self.create_batch(block_info)]\n",
"role": "user"
},
{
"content": "block_start",
"role": "assistant"
}
] |
def x(settings_view, expected_signer, batches):
rules = settings_view.get_setting(
'sawtooth.validator.block_validation_rules')
if rules is None:
return True
transactions = []
for batch in batches:
transactions += batch.transactions
rules = rules.split(';')
valid = True
for rule in rules:
try:
rule_type, arguments = rule.split(':')
except ValueError:
LOGGER.warning(
'Validation Rule Ignored, not in the correct format: %s', rule)
continue
rule_type = rule_type.strip()
if rule_type == 'NofX':
valid = _do_nofx(transactions, arguments)
elif rule_type == 'XatY':
valid = _do_xaty(transactions, arguments)
elif rule_type == 'local':
valid = _do_local(transactions, expected_signer, arguments)
if not valid:
return False
return valid
| enforce_validation_rules | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(settings_view, expected_signer, batches):\n rules = settings_view.get_setting(\n 'sawtooth.validator.block_validation_rules')\n if rules is None:\n return True\n transactions = []\n for batch in batches:\n transactions += batch.transactions\n rules = rules.split(';')\n valid = True\n for rule in rules:\n try:\n rule_type, arguments = rule.split(':')\n except ValueError:\n LOGGER.warning(\n 'Validation Rule Ignored, not in the correct format: %s', rule)\n continue\n rule_type = rule_type.strip()\n if rule_type == 'NofX':\n valid = _do_nofx(transactions, arguments)\n elif rule_type == 'XatY':\n valid = _do_xaty(transactions, arguments)\n elif rule_type == 'local':\n valid = _do_local(transactions, expected_signer, arguments)\n if not valid:\n return False\n return valid\n",
"role": "user"
},
{
"content": "enforce_validation_rules",
"role": "assistant"
}
] |
def x(transactions, arguments):
try:
num, family = arguments.split(',')
limit = int(num.strip())
except ValueError:
LOGGER.warning(
'Ignore, NofX requires arguments in the format int,family not %s',
arguments)
return True
count = 0
family = family.strip()
for txn in transactions:
header = TransactionHeader()
header.ParseFromString(txn.header)
if header.family_name == family:
count += 1
if count > limit:
LOGGER.debug('Too many transactions of type %s', family)
return False
return True
| do_nofx | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(transactions, arguments):\n try:\n num, family = arguments.split(',')\n limit = int(num.strip())\n except ValueError:\n LOGGER.warning(\n 'Ignore, NofX requires arguments in the format int,family not %s',\n arguments)\n return True\n count = 0\n family = family.strip()\n for txn in transactions:\n header = TransactionHeader()\n header.ParseFromString(txn.header)\n if header.family_name == family:\n count += 1\n if count > limit:\n LOGGER.debug('Too many transactions of type %s', family)\n return False\n return True\n",
"role": "user"
},
{
"content": "do_nofx",
"role": "assistant"
}
] |
def x(transactions, arguments):
try:
family, num = arguments.split(',')
position = int(num.strip())
except ValueError:
LOGGER.warning(
'Ignore, XatY requires arguments in the format family,position not %s'
, arguments)
return True
family = family.strip()
if abs(position) >= len(transactions):
LOGGER.debug(
'Block does not have enough transactions to validate this rule XatY:%s'
, arguments)
return False
txn = transactions[position]
header = TransactionHeader()
header.ParseFromString(txn.header)
if header.family_name != family:
LOGGER.debug('Transaction at postion %s is not of type %s',
position, family)
return False
return True
| do_xaty | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(transactions, arguments):\n try:\n family, num = arguments.split(',')\n position = int(num.strip())\n except ValueError:\n LOGGER.warning(\n 'Ignore, XatY requires arguments in the format family,position not %s'\n , arguments)\n return True\n family = family.strip()\n if abs(position) >= len(transactions):\n LOGGER.debug(\n 'Block does not have enough transactions to validate this rule XatY:%s'\n , arguments)\n return False\n txn = transactions[position]\n header = TransactionHeader()\n header.ParseFromString(txn.header)\n if header.family_name != family:\n LOGGER.debug('Transaction at postion %s is not of type %s',\n position, family)\n return False\n return True\n",
"role": "user"
},
{
"content": "do_xaty",
"role": "assistant"
}
] |
def x(transactions, expected_signer, arguments):
indices = arguments.split(',')
txns_len = len(transactions)
for index in indices:
try:
index = int(index.strip())
except ValueError:
LOGGER.warning(
'Ignore, local requries one or more comma seperated integers that represent indices, not %s'
, arguments)
return True
if abs(index) >= txns_len:
LOGGER.debug(
'Ignore, Block does not have enough transactions to validate this rule local:%s'
, index)
continue
txn = transactions[index]
header = TransactionHeader()
header.ParseFromString(txn.header)
if header.signer_public_key != expected_signer:
LOGGER.debug(
'Transaction at postion %s was not signed by the same key as the block.'
, index)
return False
return True
| do_local | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(transactions, expected_signer, arguments):\n indices = arguments.split(',')\n txns_len = len(transactions)\n for index in indices:\n try:\n index = int(index.strip())\n except ValueError:\n LOGGER.warning(\n 'Ignore, local requries one or more comma seperated integers that represent indices, not %s'\n , arguments)\n return True\n if abs(index) >= txns_len:\n LOGGER.debug(\n 'Ignore, Block does not have enough transactions to validate this rule local:%s'\n , index)\n continue\n txn = transactions[index]\n header = TransactionHeader()\n header.ParseFromString(txn.header)\n if header.signer_public_key != expected_signer:\n LOGGER.debug(\n 'Transaction at postion %s was not signed by the same key as the block.'\n , index)\n return False\n return True\n",
"role": "user"
},
{
"content": "do_local",
"role": "assistant"
}
] |
def x(filename):
if not os.path.exists(filename):
LOGGER.info(
'Skipping validator config loading from non-existent config file: %s'
, filename)
return ValidatorConfig()
LOGGER.info('Loading validator information from config: %s', filename)
try:
with open(filename) as fd:
raw_config = fd.read()
except IOError as e:
raise LocalConfigurationError(
'Unable to load validator configuration file: {}'.format(str(e)))
toml_config = toml.loads(raw_config)
invalid_keys = set(toml_config.keys()).difference(['bind', 'endpoint',
'peering', 'seeds', 'peers', 'network_public_key',
'network_private_key', 'scheduler', 'permissions', 'roles',
'opentsdb_url', 'opentsdb_db', 'opentsdb_username',
'opentsdb_password', 'minimum_peer_connectivity',
'maximum_peer_connectivity', 'state_pruning_block_depth',
'fork_cache_keep_time', 'component_thread_pool_workers',
'network_thread_pool_workers', 'signature_thread_pool_workers'])
if invalid_keys:
raise LocalConfigurationError('Invalid keys in validator config: {}'
.format(', '.join(sorted(list(invalid_keys)))))
bind_network = None
bind_component = None
bind_consensus = None
for bind in toml_config.get('bind', []):
if 'network' in bind:
bind_network = bind[bind.find(':') + 1:]
if 'component' in bind:
bind_component = bind[bind.find(':') + 1:]
if 'consensus' in bind:
bind_consensus = bind[bind.find(':') + 1:]
network_public_key = None
network_private_key = None
if toml_config.get('network_public_key') is not None:
network_public_key = toml_config.get('network_public_key').encode()
if toml_config.get('network_private_key') is not None:
network_private_key = toml_config.get('network_private_key').encode()
config = ValidatorConfig(bind_network=bind_network, bind_component=
bind_component, bind_consensus=bind_consensus, endpoint=toml_config
.get('endpoint', None), peering=toml_config.get('peering', None),
seeds=toml_config.get('seeds', None), peers=toml_config.get('peers',
None), network_public_key=network_public_key, network_private_key=
network_private_key, scheduler=toml_config.get('scheduler', None),
permissions=parse_permissions(toml_config.get('permissions', None)),
roles=toml_config.get('roles', None), opentsdb_url=toml_config.get(
'opentsdb_url', None), opentsdb_db=toml_config.get('opentsdb_db',
None), opentsdb_username=toml_config.get('opentsdb_username', None),
opentsdb_password=toml_config.get('opentsdb_password', None),
minimum_peer_connectivity=toml_config.get(
'minimum_peer_connectivity', None), maximum_peer_connectivity=
toml_config.get('maximum_peer_connectivity', None),
state_pruning_block_depth=toml_config.get(
'state_pruning_block_depth', None), fork_cache_keep_time=
toml_config.get('fork_cache_keep_time', None),
component_thread_pool_workers=toml_config.get(
'component_thread_pool_workers', None), network_thread_pool_workers
=toml_config.get('network_thread_pool_workers', None),
signature_thread_pool_workers=toml_config.get(
'signature_thread_pool_workers', None))
return config
| load_toml_validator_config | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(filename):\n if not os.path.exists(filename):\n LOGGER.info(\n 'Skipping validator config loading from non-existent config file: %s'\n , filename)\n return ValidatorConfig()\n LOGGER.info('Loading validator information from config: %s', filename)\n try:\n with open(filename) as fd:\n raw_config = fd.read()\n except IOError as e:\n raise LocalConfigurationError(\n 'Unable to load validator configuration file: {}'.format(str(e)))\n toml_config = toml.loads(raw_config)\n invalid_keys = set(toml_config.keys()).difference(['bind', 'endpoint',\n 'peering', 'seeds', 'peers', 'network_public_key',\n 'network_private_key', 'scheduler', 'permissions', 'roles',\n 'opentsdb_url', 'opentsdb_db', 'opentsdb_username',\n 'opentsdb_password', 'minimum_peer_connectivity',\n 'maximum_peer_connectivity', 'state_pruning_block_depth',\n 'fork_cache_keep_time', 'component_thread_pool_workers',\n 'network_thread_pool_workers', 'signature_thread_pool_workers'])\n if invalid_keys:\n raise LocalConfigurationError('Invalid keys in validator config: {}'\n .format(', '.join(sorted(list(invalid_keys)))))\n bind_network = None\n bind_component = None\n bind_consensus = None\n for bind in toml_config.get('bind', []):\n if 'network' in bind:\n bind_network = bind[bind.find(':') + 1:]\n if 'component' in bind:\n bind_component = bind[bind.find(':') + 1:]\n if 'consensus' in bind:\n bind_consensus = bind[bind.find(':') + 1:]\n network_public_key = None\n network_private_key = None\n if toml_config.get('network_public_key') is not None:\n network_public_key = toml_config.get('network_public_key').encode()\n if toml_config.get('network_private_key') is not None:\n network_private_key = toml_config.get('network_private_key').encode()\n config = ValidatorConfig(bind_network=bind_network, bind_component=\n bind_component, bind_consensus=bind_consensus, endpoint=toml_config\n .get('endpoint', None), peering=toml_config.get('peering', None),\n seeds=toml_config.get('seeds', None), peers=toml_config.get('peers',\n None), network_public_key=network_public_key, network_private_key=\n network_private_key, scheduler=toml_config.get('scheduler', None),\n permissions=parse_permissions(toml_config.get('permissions', None)),\n roles=toml_config.get('roles', None), opentsdb_url=toml_config.get(\n 'opentsdb_url', None), opentsdb_db=toml_config.get('opentsdb_db',\n None), opentsdb_username=toml_config.get('opentsdb_username', None),\n opentsdb_password=toml_config.get('opentsdb_password', None),\n minimum_peer_connectivity=toml_config.get(\n 'minimum_peer_connectivity', None), maximum_peer_connectivity=\n toml_config.get('maximum_peer_connectivity', None),\n state_pruning_block_depth=toml_config.get(\n 'state_pruning_block_depth', None), fork_cache_keep_time=\n toml_config.get('fork_cache_keep_time', None),\n component_thread_pool_workers=toml_config.get(\n 'component_thread_pool_workers', None), network_thread_pool_workers\n =toml_config.get('network_thread_pool_workers', None),\n signature_thread_pool_workers=toml_config.get(\n 'signature_thread_pool_workers', None))\n return config\n",
"role": "user"
},
{
"content": "load_toml_validator_config",
"role": "assistant"
}
] |
def x(configs):
bind_network = None
bind_component = None
bind_consensus = None
endpoint = None
peering = None
seeds = None
peers = None
network_public_key = None
network_private_key = None
scheduler = None
permissions = None
roles = None
opentsdb_url = None
opentsdb_db = None
opentsdb_username = None
opentsdb_password = None
minimum_peer_connectivity = None
maximum_peer_connectivity = None
state_pruning_block_depth = None
fork_cache_keep_time = None
component_thread_pool_workers = None
network_thread_pool_workers = None
signature_thread_pool_workers = None
for config in reversed(configs):
if config.bind_network is not None:
bind_network = config.bind_network
if config.bind_component is not None:
bind_component = config.bind_component
if config.bind_consensus is not None:
bind_consensus = config.bind_consensus
if config.endpoint is not None:
endpoint = config.endpoint
if config.peering is not None:
peering = config.peering
if config.seeds is not None:
seeds = config.seeds
if config.peers is not None:
peers = config.peers
if config.network_public_key is not None:
network_public_key = config.network_public_key
if config.network_private_key is not None:
network_private_key = config.network_private_key
if config.scheduler is not None:
scheduler = config.scheduler
if config.permissions is not None or config.permissions == {}:
permissions = config.permissions
if config.roles is not None:
roles = config.roles
if config.opentsdb_url is not None:
opentsdb_url = config.opentsdb_url
if config.opentsdb_db is not None:
opentsdb_db = config.opentsdb_db
if config.opentsdb_username is not None:
opentsdb_username = config.opentsdb_username
if config.opentsdb_password is not None:
opentsdb_password = config.opentsdb_password
if config.minimum_peer_connectivity is not None:
minimum_peer_connectivity = config.minimum_peer_connectivity
if config.maximum_peer_connectivity is not None:
maximum_peer_connectivity = config.maximum_peer_connectivity
if config.state_pruning_block_depth is not None:
state_pruning_block_depth = config.state_pruning_block_depth
if config.fork_cache_keep_time is not None:
fork_cache_keep_time = config.fork_cache_keep_time
if config.component_thread_pool_workers is not None:
component_thread_pool_workers = (config.
component_thread_pool_workers)
if config.network_thread_pool_workers is not None:
network_thread_pool_workers = config.network_thread_pool_workers
if config.signature_thread_pool_workers is not None:
signature_thread_pool_workers = (config.
signature_thread_pool_workers)
return ValidatorConfig(bind_network=bind_network, bind_component=
bind_component, bind_consensus=bind_consensus, endpoint=endpoint,
peering=peering, seeds=seeds, peers=peers, network_public_key=
network_public_key, network_private_key=network_private_key,
scheduler=scheduler, permissions=permissions, roles=roles,
opentsdb_url=opentsdb_url, opentsdb_db=opentsdb_db,
opentsdb_username=opentsdb_username, opentsdb_password=
opentsdb_password, minimum_peer_connectivity=
minimum_peer_connectivity, maximum_peer_connectivity=
maximum_peer_connectivity, state_pruning_block_depth=
state_pruning_block_depth, fork_cache_keep_time=
fork_cache_keep_time, component_thread_pool_workers=
component_thread_pool_workers, network_thread_pool_workers=
network_thread_pool_workers, signature_thread_pool_workers=
signature_thread_pool_workers)
| merge_validator_config | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(configs):\n bind_network = None\n bind_component = None\n bind_consensus = None\n endpoint = None\n peering = None\n seeds = None\n peers = None\n network_public_key = None\n network_private_key = None\n scheduler = None\n permissions = None\n roles = None\n opentsdb_url = None\n opentsdb_db = None\n opentsdb_username = None\n opentsdb_password = None\n minimum_peer_connectivity = None\n maximum_peer_connectivity = None\n state_pruning_block_depth = None\n fork_cache_keep_time = None\n component_thread_pool_workers = None\n network_thread_pool_workers = None\n signature_thread_pool_workers = None\n for config in reversed(configs):\n if config.bind_network is not None:\n bind_network = config.bind_network\n if config.bind_component is not None:\n bind_component = config.bind_component\n if config.bind_consensus is not None:\n bind_consensus = config.bind_consensus\n if config.endpoint is not None:\n endpoint = config.endpoint\n if config.peering is not None:\n peering = config.peering\n if config.seeds is not None:\n seeds = config.seeds\n if config.peers is not None:\n peers = config.peers\n if config.network_public_key is not None:\n network_public_key = config.network_public_key\n if config.network_private_key is not None:\n network_private_key = config.network_private_key\n if config.scheduler is not None:\n scheduler = config.scheduler\n if config.permissions is not None or config.permissions == {}:\n permissions = config.permissions\n if config.roles is not None:\n roles = config.roles\n if config.opentsdb_url is not None:\n opentsdb_url = config.opentsdb_url\n if config.opentsdb_db is not None:\n opentsdb_db = config.opentsdb_db\n if config.opentsdb_username is not None:\n opentsdb_username = config.opentsdb_username\n if config.opentsdb_password is not None:\n opentsdb_password = config.opentsdb_password\n if config.minimum_peer_connectivity is not None:\n minimum_peer_connectivity = config.minimum_peer_connectivity\n if config.maximum_peer_connectivity is not None:\n maximum_peer_connectivity = config.maximum_peer_connectivity\n if config.state_pruning_block_depth is not None:\n state_pruning_block_depth = config.state_pruning_block_depth\n if config.fork_cache_keep_time is not None:\n fork_cache_keep_time = config.fork_cache_keep_time\n if config.component_thread_pool_workers is not None:\n component_thread_pool_workers = (config.\n component_thread_pool_workers)\n if config.network_thread_pool_workers is not None:\n network_thread_pool_workers = config.network_thread_pool_workers\n if config.signature_thread_pool_workers is not None:\n signature_thread_pool_workers = (config.\n signature_thread_pool_workers)\n return ValidatorConfig(bind_network=bind_network, bind_component=\n bind_component, bind_consensus=bind_consensus, endpoint=endpoint,\n peering=peering, seeds=seeds, peers=peers, network_public_key=\n network_public_key, network_private_key=network_private_key,\n scheduler=scheduler, permissions=permissions, roles=roles,\n opentsdb_url=opentsdb_url, opentsdb_db=opentsdb_db,\n opentsdb_username=opentsdb_username, opentsdb_password=\n opentsdb_password, minimum_peer_connectivity=\n minimum_peer_connectivity, maximum_peer_connectivity=\n maximum_peer_connectivity, state_pruning_block_depth=\n state_pruning_block_depth, fork_cache_keep_time=\n fork_cache_keep_time, component_thread_pool_workers=\n component_thread_pool_workers, network_thread_pool_workers=\n network_thread_pool_workers, signature_thread_pool_workers=\n signature_thread_pool_workers)\n",
"role": "user"
},
{
"content": "merge_validator_config",
"role": "assistant"
}
] |
def x(key_dir, key_name):
key_path = os.path.join(key_dir, '{}.priv'.format(key_name))
if not os.path.exists(key_path):
raise LocalConfigurationError('No such signing key file: {}'.format
(key_path))
if not os.access(key_path, os.R_OK):
raise LocalConfigurationError('Key file is not readable: {}'.format
(key_path))
LOGGER.info('Loading signing key: %s', key_path)
try:
with open(key_path, 'r') as key_file:
private_key_str = key_file.read().strip()
except IOError as e:
raise LocalConfigurationError('Could not load key file: {}'.format(
str(e)))
try:
private_key = Secp256k1PrivateKey.from_hex(private_key_str)
except signing.ParseError as e:
raise LocalConfigurationError('Invalid key in file {}: {}'.format(
key_path, str(e)))
context = signing.create_context('secp256k1')
crypto_factory = CryptoFactory(context)
return crypto_factory.new_signer(private_key)
| load_identity_signer | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(key_dir, key_name):\n key_path = os.path.join(key_dir, '{}.priv'.format(key_name))\n if not os.path.exists(key_path):\n raise LocalConfigurationError('No such signing key file: {}'.format\n (key_path))\n if not os.access(key_path, os.R_OK):\n raise LocalConfigurationError('Key file is not readable: {}'.format\n (key_path))\n LOGGER.info('Loading signing key: %s', key_path)\n try:\n with open(key_path, 'r') as key_file:\n private_key_str = key_file.read().strip()\n except IOError as e:\n raise LocalConfigurationError('Could not load key file: {}'.format(\n str(e)))\n try:\n private_key = Secp256k1PrivateKey.from_hex(private_key_str)\n except signing.ParseError as e:\n raise LocalConfigurationError('Invalid key in file {}: {}'.format(\n key_path, str(e)))\n context = signing.create_context('secp256k1')\n crypto_factory = CryptoFactory(context)\n return crypto_factory.new_signer(private_key)\n",
"role": "user"
},
{
"content": "load_identity_signer",
"role": "assistant"
}
] |
def x(subparsers, parent_parser):
parser = subparsers.add_parser('status', help=
'Displays information about validator status', description=
"Provides a subcommand to show a validator's status")
grand_parsers = parser.add_subparsers(title='subcommands', dest=
'subcommand')
grand_parsers.required = True
add_status_show_parser(grand_parsers, parent_parser)
| add_status_parser | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(subparsers, parent_parser):\n parser = subparsers.add_parser('status', help=\n 'Displays information about validator status', description=\n \"Provides a subcommand to show a validator's status\")\n grand_parsers = parser.add_subparsers(title='subcommands', dest=\n 'subcommand')\n grand_parsers.required = True\n add_status_show_parser(grand_parsers, parent_parser)\n",
"role": "user"
},
{
"content": "add_status_parser",
"role": "assistant"
}
] |
def x(self, processor_type):
with self._condition:
if processor_type not in self:
self.wait_for_registration(processor_type)
try:
processor = self[processor_type].next_processor()
except NoProcessorVacancyError:
processor = self.wait_for_vacancy(processor_type)
processor.inc_occupancy()
return processor
| get_next_of_type | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, processor_type):\n with self._condition:\n if processor_type not in self:\n self.wait_for_registration(processor_type)\n try:\n processor = self[processor_type].next_processor()\n except NoProcessorVacancyError:\n processor = self.wait_for_vacancy(processor_type)\n processor.inc_occupancy()\n return processor\n",
"role": "user"
},
{
"content": "get_next_of_type",
"role": "assistant"
}
] |
def x(self, processor_identity):
with self._condition:
processor_types = self._identities.get(processor_identity)
if processor_types is None:
LOGGER.warning(
'transaction processor with identity %s tried to unregister but was not registered'
, processor_identity)
return
for processor_type in processor_types:
if processor_type not in self._processors:
LOGGER.warning(
'processor type %s not a known processor type but is associated with identity %s'
, processor_type, processor_identity)
continue
self._processors[processor_type].remove_processor(
processor_identity=processor_identity)
if not self._processors[processor_type]:
del self._processors[processor_type]
| remove | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, processor_identity):\n with self._condition:\n processor_types = self._identities.get(processor_identity)\n if processor_types is None:\n LOGGER.warning(\n 'transaction processor with identity %s tried to unregister but was not registered'\n , processor_identity)\n return\n for processor_type in processor_types:\n if processor_type not in self._processors:\n LOGGER.warning(\n 'processor type %s not a known processor type but is associated with identity %s'\n , processor_type, processor_identity)\n continue\n self._processors[processor_type].remove_processor(\n processor_identity=processor_identity)\n if not self._processors[processor_type]:\n del self._processors[processor_type]\n",
"role": "user"
},
{
"content": "remove",
"role": "assistant"
}
] |
def x(self, processor_type):
with self._condition:
self._condition.wait_for(lambda : processor_type in self or self.
_cancelled_event.is_set())
if self._cancelled_event.is_set():
raise WaitCancelledException()
| wait_for_registration | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, processor_type):\n with self._condition:\n self._condition.wait_for(lambda : processor_type in self or self.\n _cancelled_event.is_set())\n if self._cancelled_event.is_set():\n raise WaitCancelledException()\n",
"role": "user"
},
{
"content": "wait_for_registration",
"role": "assistant"
}
] |
def x(self, processor_type):
with self._condition:
self._condition.wait_for(lambda : self._processor_available(
processor_type) or self._cancelled_event.is_set())
if self._cancelled_event.is_set():
raise WaitCancelledException()
processor = self[processor_type].next_processor()
return processor
| wait_for_vacancy | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, processor_type):\n with self._condition:\n self._condition.wait_for(lambda : self._processor_available(\n processor_type) or self._cancelled_event.is_set())\n if self._cancelled_event.is_set():\n raise WaitCancelledException()\n processor = self[processor_type].next_processor()\n return processor\n",
"role": "user"
},
{
"content": "wait_for_vacancy",
"role": "assistant"
}
] |
def x(self, key, default_value=None, value_type=str):
try:
state_entry = self._state_view.get(SettingsView.setting_address(key))
except KeyError:
return default_value
if state_entry is not None:
setting = Setting()
setting.ParseFromString(state_entry)
for setting_entry in setting.entries:
if setting_entry.key == key:
return value_type(setting_entry.value)
return default_value
| get_setting | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, key, default_value=None, value_type=str):\n try:\n state_entry = self._state_view.get(SettingsView.setting_address(key))\n except KeyError:\n return default_value\n if state_entry is not None:\n setting = Setting()\n setting.ParseFromString(state_entry)\n for setting_entry in setting.entries:\n if setting_entry.key == key:\n return value_type(setting_entry.value)\n return default_value\n",
"role": "user"
},
{
"content": "get_setting",
"role": "assistant"
}
] |
def x(self, key, default_value=None, delimiter=',', value_type=str):
value = self.get_setting(key)
if value is not None:
setting_list = [value_type(v) for v in value.split(delimiter)]
else:
setting_list = default_value
return setting_list
| get_setting_list | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, key, default_value=None, delimiter=',', value_type=str):\n value = self.get_setting(key)\n if value is not None:\n setting_list = [value_type(v) for v in value.split(delimiter)]\n else:\n setting_list = default_value\n return setting_list\n",
"role": "user"
},
{
"content": "get_setting_list",
"role": "assistant"
}
] |
def x(key):
key_parts = key.split('.', maxsplit=_MAX_KEY_PARTS - 1)
addr_parts = [_short_hash(x.encode()) for x in key_parts]
addr_parts.extend([_EMPTY_PART] * (_MAX_KEY_PARTS - len(addr_parts)))
return CONFIG_STATE_NAMESPACE + ''.join(addr_parts)
| setting_address | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(key):\n key_parts = key.split('.', maxsplit=_MAX_KEY_PARTS - 1)\n addr_parts = [_short_hash(x.encode()) for x in key_parts]\n addr_parts.extend([_EMPTY_PART] * (_MAX_KEY_PARTS - len(addr_parts)))\n return CONFIG_STATE_NAMESPACE + ''.join(addr_parts)\n",
"role": "user"
},
{
"content": "setting_address",
"role": "assistant"
}
] |
async def submit_batches(self, request):
"""Accepts a binary encoded BatchList and submits it to the validator.
Request:
body: octet-stream BatchList of one or more Batches
Response:
status:
- 202: Batches submitted and pending
link: /batches or /batch_statuses link for submitted batches
"""
timer_ctx = self._post_batches_total_time.time()
self._post_batches_count.inc()
if request.headers['Content-Type'] != 'application/octet-stream':
LOGGER.debug('Submission headers had wrong Content-Type: %s',
request.headers['Content-Type'])
self._post_batches_error.inc()
raise errors.SubmissionWrongContentType()
body = await request.read()
if not body:
LOGGER.debug('Submission contained an empty body')
self._post_batches_error.inc()
raise errors.NoBatchesSubmitted()
try:
batch_list = BatchList()
batch_list.ParseFromString(body)
except DecodeError:
LOGGER.debug('Submission body could not be decoded: %s', body)
self._post_batches_error.inc()
raise errors.BadProtobufSubmitted()
error_traps = [error_handlers.BatchInvalidTrap, error_handlers.
BatchQueueFullTrap]
validator_query = client_batch_submit_pb2.ClientBatchSubmitRequest(batches
=batch_list.batches)
with self._post_batches_validator_time.time():
await self._query_validator(Message.CLIENT_BATCH_SUBMIT_REQUEST,
client_batch_submit_pb2.ClientBatchSubmitResponse,
validator_query, error_traps)
id_string = ','.join(b.header_signature for b in batch_list.batches)
status = 202
link = self._build_url(request, path='/batch_statuses', id=id_string)
retval = self._wrap_response(request, metadata={'link': link}, status=
status)
timer_ctx.stop()
return retval
| submit_batches | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def submit_batches(self, request):\n \"\"\"Accepts a binary encoded BatchList and submits it to the validator.\n\n Request:\n body: octet-stream BatchList of one or more Batches\n Response:\n status:\n - 202: Batches submitted and pending\n link: /batches or /batch_statuses link for submitted batches\n\n \"\"\"\n timer_ctx = self._post_batches_total_time.time()\n self._post_batches_count.inc()\n if request.headers['Content-Type'] != 'application/octet-stream':\n LOGGER.debug('Submission headers had wrong Content-Type: %s',\n request.headers['Content-Type'])\n self._post_batches_error.inc()\n raise errors.SubmissionWrongContentType()\n body = await request.read()\n if not body:\n LOGGER.debug('Submission contained an empty body')\n self._post_batches_error.inc()\n raise errors.NoBatchesSubmitted()\n try:\n batch_list = BatchList()\n batch_list.ParseFromString(body)\n except DecodeError:\n LOGGER.debug('Submission body could not be decoded: %s', body)\n self._post_batches_error.inc()\n raise errors.BadProtobufSubmitted()\n error_traps = [error_handlers.BatchInvalidTrap, error_handlers.\n BatchQueueFullTrap]\n validator_query = client_batch_submit_pb2.ClientBatchSubmitRequest(batches\n =batch_list.batches)\n with self._post_batches_validator_time.time():\n await self._query_validator(Message.CLIENT_BATCH_SUBMIT_REQUEST,\n client_batch_submit_pb2.ClientBatchSubmitResponse,\n validator_query, error_traps)\n id_string = ','.join(b.header_signature for b in batch_list.batches)\n status = 202\n link = self._build_url(request, path='/batch_statuses', id=id_string)\n retval = self._wrap_response(request, metadata={'link': link}, status=\n status)\n timer_ctx.stop()\n return retval\n",
"role": "user"
},
{
"content": "submit_batches",
"role": "assistant"
}
] |
async def list_statuses(self, request):
"""Fetches the committed status of batches by either a POST or GET.
Request:
body: A JSON array of one or more id strings (if POST)
query:
- id: A comma separated list of up to 15 ids (if GET)
- wait: Request should not return until all batches committed
Response:
data: A JSON object, with batch ids as keys, and statuses as values
link: The /batch_statuses link queried (if GET)
"""
error_traps = [error_handlers.StatusResponseMissing]
if request.method == 'POST':
if request.headers['Content-Type'] != 'application/json':
LOGGER.debug('Request headers had wrong Content-Type: %s',
request.headers['Content-Type'])
raise errors.StatusWrongContentType()
ids = await request.json()
if not ids or not isinstance(ids, list) or not all(isinstance(i,
str) for i in ids):
LOGGER.debug('Request body was invalid: %s', ids)
raise errors.StatusBodyInvalid()
for i in ids:
self._validate_id(i)
else:
ids = self._get_filter_ids(request)
if not ids:
LOGGER.debug('Request for statuses missing id query')
raise errors.StatusIdQueryInvalid()
validator_query = client_batch_submit_pb2.ClientBatchStatusRequest(
batch_ids=ids)
self._set_wait(request, validator_query)
response = await self._query_validator(Message.
CLIENT_BATCH_STATUS_REQUEST, client_batch_submit_pb2.
ClientBatchStatusResponse, validator_query, error_traps)
if request.method != 'POST':
metadata = self._get_metadata(request, response)
else:
metadata = None
data = self._drop_id_prefixes(self._drop_empty_props(response[
'batch_statuses']))
return self._wrap_response(request, data=data, metadata=metadata)
| list_statuses | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def list_statuses(self, request):\n \"\"\"Fetches the committed status of batches by either a POST or GET.\n\n Request:\n body: A JSON array of one or more id strings (if POST)\n query:\n - id: A comma separated list of up to 15 ids (if GET)\n - wait: Request should not return until all batches committed\n\n Response:\n data: A JSON object, with batch ids as keys, and statuses as values\n link: The /batch_statuses link queried (if GET)\n \"\"\"\n error_traps = [error_handlers.StatusResponseMissing]\n if request.method == 'POST':\n if request.headers['Content-Type'] != 'application/json':\n LOGGER.debug('Request headers had wrong Content-Type: %s',\n request.headers['Content-Type'])\n raise errors.StatusWrongContentType()\n ids = await request.json()\n if not ids or not isinstance(ids, list) or not all(isinstance(i,\n str) for i in ids):\n LOGGER.debug('Request body was invalid: %s', ids)\n raise errors.StatusBodyInvalid()\n for i in ids:\n self._validate_id(i)\n else:\n ids = self._get_filter_ids(request)\n if not ids:\n LOGGER.debug('Request for statuses missing id query')\n raise errors.StatusIdQueryInvalid()\n validator_query = client_batch_submit_pb2.ClientBatchStatusRequest(\n batch_ids=ids)\n self._set_wait(request, validator_query)\n response = await self._query_validator(Message.\n CLIENT_BATCH_STATUS_REQUEST, client_batch_submit_pb2.\n ClientBatchStatusResponse, validator_query, error_traps)\n if request.method != 'POST':\n metadata = self._get_metadata(request, response)\n else:\n metadata = None\n data = self._drop_id_prefixes(self._drop_empty_props(response[\n 'batch_statuses']))\n return self._wrap_response(request, data=data, metadata=metadata)\n",
"role": "user"
},
{
"content": "list_statuses",
"role": "assistant"
}
] |
async def list_state(self, request):
"""Fetches list of data entries, optionally filtered by address prefix.
Request:
query:
- head: The id of the block to use as the head of the chain
- address: Return entries whose addresses begin with this
prefix
Response:
data: An array of leaf objects with address and data keys
head: The head used for this query (most recent if unspecified)
link: The link to this exact query, including head block
paging: Paging info and nav, like total resources and a next link
"""
paging_controls = self._get_paging_controls(request)
head, root = await self._head_to_root(request.url.query.get('head', None))
validator_query = client_state_pb2.ClientStateListRequest(state_root=
root, address=request.url.query.get('address', None), sorting=self.
_get_sorting_message(request, 'default'), paging=self.
_make_paging_message(paging_controls))
response = await self._query_validator(Message.
CLIENT_STATE_LIST_REQUEST, client_state_pb2.ClientStateListResponse,
validator_query)
return self._wrap_paginated_response(request=request, response=response,
controls=paging_controls, data=response.get('entries', []), head=head)
| list_state | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def list_state(self, request):\n \"\"\"Fetches list of data entries, optionally filtered by address prefix.\n\n Request:\n query:\n - head: The id of the block to use as the head of the chain\n - address: Return entries whose addresses begin with this\n prefix\n\n Response:\n data: An array of leaf objects with address and data keys\n head: The head used for this query (most recent if unspecified)\n link: The link to this exact query, including head block\n paging: Paging info and nav, like total resources and a next link\n \"\"\"\n paging_controls = self._get_paging_controls(request)\n head, root = await self._head_to_root(request.url.query.get('head', None))\n validator_query = client_state_pb2.ClientStateListRequest(state_root=\n root, address=request.url.query.get('address', None), sorting=self.\n _get_sorting_message(request, 'default'), paging=self.\n _make_paging_message(paging_controls))\n response = await self._query_validator(Message.\n CLIENT_STATE_LIST_REQUEST, client_state_pb2.ClientStateListResponse,\n validator_query)\n return self._wrap_paginated_response(request=request, response=response,\n controls=paging_controls, data=response.get('entries', []), head=head)\n",
"role": "user"
},
{
"content": "list_state",
"role": "assistant"
}
] |
async def fetch_state(self, request):
"""Fetches data from a specific address in the validator's state tree.
Request:
query:
- head: The id of the block to use as the head of the chain
- address: The 70 character address of the data to be fetched
Response:
data: The base64 encoded binary data stored at that address
head: The head used for this query (most recent if unspecified)
link: The link to this exact query, including head block
"""
error_traps = [error_handlers.InvalidAddressTrap, error_handlers.
StateNotFoundTrap]
address = request.match_info.get('address', '')
head = request.url.query.get('head', None)
head, root = await self._head_to_root(head)
response = await self._query_validator(Message.CLIENT_STATE_GET_REQUEST,
client_state_pb2.ClientStateGetResponse, client_state_pb2.
ClientStateGetRequest(state_root=root, address=address), error_traps)
return self._wrap_response(request, data=response['value'], metadata=
self._get_metadata(request, response, head=head))
| fetch_state | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def fetch_state(self, request):\n \"\"\"Fetches data from a specific address in the validator's state tree.\n\n Request:\n query:\n - head: The id of the block to use as the head of the chain\n - address: The 70 character address of the data to be fetched\n\n Response:\n data: The base64 encoded binary data stored at that address\n head: The head used for this query (most recent if unspecified)\n link: The link to this exact query, including head block\n \"\"\"\n error_traps = [error_handlers.InvalidAddressTrap, error_handlers.\n StateNotFoundTrap]\n address = request.match_info.get('address', '')\n head = request.url.query.get('head', None)\n head, root = await self._head_to_root(head)\n response = await self._query_validator(Message.CLIENT_STATE_GET_REQUEST,\n client_state_pb2.ClientStateGetResponse, client_state_pb2.\n ClientStateGetRequest(state_root=root, address=address), error_traps)\n return self._wrap_response(request, data=response['value'], metadata=\n self._get_metadata(request, response, head=head))\n",
"role": "user"
},
{
"content": "fetch_state",
"role": "assistant"
}
] |
async def list_blocks(self, request):
"""Fetches list of blocks from validator, optionally filtered by id.
Request:
query:
- head: The id of the block to use as the head of the chain
- id: Comma separated list of block ids to include in results
Response:
data: JSON array of fully expanded Block objects
head: The head used for this query (most recent if unspecified)
link: The link to this exact query, including head block
paging: Paging info and nav, like total resources and a next link
"""
paging_controls = self._get_paging_controls(request)
validator_query = client_block_pb2.ClientBlockListRequest(head_id=self.
_get_head_id(request), block_ids=self._get_filter_ids(request),
sorting=self._get_sorting_message(request, 'block_num'), paging=
self._make_paging_message(paging_controls))
response = await self._query_validator(Message.
CLIENT_BLOCK_LIST_REQUEST, client_block_pb2.ClientBlockListResponse,
validator_query)
return self._wrap_paginated_response(request=request, response=response,
controls=paging_controls, data=[self._expand_block(b) for b in
response['blocks']])
| list_blocks | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def list_blocks(self, request):\n \"\"\"Fetches list of blocks from validator, optionally filtered by id.\n\n Request:\n query:\n - head: The id of the block to use as the head of the chain\n - id: Comma separated list of block ids to include in results\n\n Response:\n data: JSON array of fully expanded Block objects\n head: The head used for this query (most recent if unspecified)\n link: The link to this exact query, including head block\n paging: Paging info and nav, like total resources and a next link\n \"\"\"\n paging_controls = self._get_paging_controls(request)\n validator_query = client_block_pb2.ClientBlockListRequest(head_id=self.\n _get_head_id(request), block_ids=self._get_filter_ids(request),\n sorting=self._get_sorting_message(request, 'block_num'), paging=\n self._make_paging_message(paging_controls))\n response = await self._query_validator(Message.\n CLIENT_BLOCK_LIST_REQUEST, client_block_pb2.ClientBlockListResponse,\n validator_query)\n return self._wrap_paginated_response(request=request, response=response,\n controls=paging_controls, data=[self._expand_block(b) for b in\n response['blocks']])\n",
"role": "user"
},
{
"content": "list_blocks",
"role": "assistant"
}
] |
async def fetch_block(self, request):
"""Fetches a specific block from the validator, specified by id.
Request:
path:
- block_id: The 128-character id of the block to be fetched
Response:
data: A JSON object with the data from the fully expanded Block
link: The link to this exact query
"""
error_traps = [error_handlers.BlockNotFoundTrap]
block_id = request.match_info.get('block_id', '')
self._validate_id(block_id)
response = await self._query_validator(Message.
CLIENT_BLOCK_GET_BY_ID_REQUEST, client_block_pb2.
ClientBlockGetResponse, client_block_pb2.ClientBlockGetByIdRequest(
block_id=block_id), error_traps)
return self._wrap_response(request, data=self._expand_block(response[
'block']), metadata=self._get_metadata(request, response))
| fetch_block | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def fetch_block(self, request):\n \"\"\"Fetches a specific block from the validator, specified by id.\n Request:\n path:\n - block_id: The 128-character id of the block to be fetched\n\n Response:\n data: A JSON object with the data from the fully expanded Block\n link: The link to this exact query\n \"\"\"\n error_traps = [error_handlers.BlockNotFoundTrap]\n block_id = request.match_info.get('block_id', '')\n self._validate_id(block_id)\n response = await self._query_validator(Message.\n CLIENT_BLOCK_GET_BY_ID_REQUEST, client_block_pb2.\n ClientBlockGetResponse, client_block_pb2.ClientBlockGetByIdRequest(\n block_id=block_id), error_traps)\n return self._wrap_response(request, data=self._expand_block(response[\n 'block']), metadata=self._get_metadata(request, response))\n",
"role": "user"
},
{
"content": "fetch_block",
"role": "assistant"
}
] |
async def list_batches(self, request):
"""Fetches list of batches from validator, optionally filtered by id.
Request:
query:
- head: The id of the block to use as the head of the chain
- id: Comma separated list of batch ids to include in results
Response:
data: JSON array of fully expanded Batch objects
head: The head used for this query (most recent if unspecified)
link: The link to this exact query, including head block
paging: Paging info and nav, like total resources and a next link
"""
paging_controls = self._get_paging_controls(request)
validator_query = client_batch_pb2.ClientBatchListRequest(head_id=self.
_get_head_id(request), batch_ids=self._get_filter_ids(request),
sorting=self._get_sorting_message(request, 'default'), paging=self.
_make_paging_message(paging_controls))
response = await self._query_validator(Message.
CLIENT_BATCH_LIST_REQUEST, client_batch_pb2.ClientBatchListResponse,
validator_query)
return self._wrap_paginated_response(request=request, response=response,
controls=paging_controls, data=[self._expand_batch(b) for b in
response['batches']])
| list_batches | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def list_batches(self, request):\n \"\"\"Fetches list of batches from validator, optionally filtered by id.\n\n Request:\n query:\n - head: The id of the block to use as the head of the chain\n - id: Comma separated list of batch ids to include in results\n\n Response:\n data: JSON array of fully expanded Batch objects\n head: The head used for this query (most recent if unspecified)\n link: The link to this exact query, including head block\n paging: Paging info and nav, like total resources and a next link\n \"\"\"\n paging_controls = self._get_paging_controls(request)\n validator_query = client_batch_pb2.ClientBatchListRequest(head_id=self.\n _get_head_id(request), batch_ids=self._get_filter_ids(request),\n sorting=self._get_sorting_message(request, 'default'), paging=self.\n _make_paging_message(paging_controls))\n response = await self._query_validator(Message.\n CLIENT_BATCH_LIST_REQUEST, client_batch_pb2.ClientBatchListResponse,\n validator_query)\n return self._wrap_paginated_response(request=request, response=response,\n controls=paging_controls, data=[self._expand_batch(b) for b in\n response['batches']])\n",
"role": "user"
},
{
"content": "list_batches",
"role": "assistant"
}
] |
async def fetch_batch(self, request):
"""Fetches a specific batch from the validator, specified by id.
Request:
path:
- batch_id: The 128-character id of the batch to be fetched
Response:
data: A JSON object with the data from the fully expanded Batch
link: The link to this exact query
"""
error_traps = [error_handlers.BatchNotFoundTrap]
batch_id = request.match_info.get('batch_id', '')
self._validate_id(batch_id)
response = await self._query_validator(Message.CLIENT_BATCH_GET_REQUEST,
client_batch_pb2.ClientBatchGetResponse, client_batch_pb2.
ClientBatchGetRequest(batch_id=batch_id), error_traps)
return self._wrap_response(request, data=self._expand_batch(response[
'batch']), metadata=self._get_metadata(request, response))
| fetch_batch | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def fetch_batch(self, request):\n \"\"\"Fetches a specific batch from the validator, specified by id.\n\n Request:\n path:\n - batch_id: The 128-character id of the batch to be fetched\n\n Response:\n data: A JSON object with the data from the fully expanded Batch\n link: The link to this exact query\n \"\"\"\n error_traps = [error_handlers.BatchNotFoundTrap]\n batch_id = request.match_info.get('batch_id', '')\n self._validate_id(batch_id)\n response = await self._query_validator(Message.CLIENT_BATCH_GET_REQUEST,\n client_batch_pb2.ClientBatchGetResponse, client_batch_pb2.\n ClientBatchGetRequest(batch_id=batch_id), error_traps)\n return self._wrap_response(request, data=self._expand_batch(response[\n 'batch']), metadata=self._get_metadata(request, response))\n",
"role": "user"
},
{
"content": "fetch_batch",
"role": "assistant"
}
] |
async def list_transactions(self, request):
"""Fetches list of txns from validator, optionally filtered by id.
Request:
query:
- head: The id of the block to use as the head of the chain
- id: Comma separated list of txn ids to include in results
Response:
data: JSON array of Transaction objects with expanded headers
head: The head used for this query (most recent if unspecified)
link: The link to this exact query, including head block
paging: Paging info and nav, like total resources and a next link
"""
paging_controls = self._get_paging_controls(request)
validator_query = client_transaction_pb2.ClientTransactionListRequest(
head_id=self._get_head_id(request), transaction_ids=self.
_get_filter_ids(request), sorting=self._get_sorting_message(request,
'default'), paging=self._make_paging_message(paging_controls))
response = await self._query_validator(Message.
CLIENT_TRANSACTION_LIST_REQUEST, client_transaction_pb2.
ClientTransactionListResponse, validator_query)
data = [self._expand_transaction(t) for t in response['transactions']]
return self._wrap_paginated_response(request=request, response=response,
controls=paging_controls, data=data)
| list_transactions | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def list_transactions(self, request):\n \"\"\"Fetches list of txns from validator, optionally filtered by id.\n\n Request:\n query:\n - head: The id of the block to use as the head of the chain\n - id: Comma separated list of txn ids to include in results\n\n Response:\n data: JSON array of Transaction objects with expanded headers\n head: The head used for this query (most recent if unspecified)\n link: The link to this exact query, including head block\n paging: Paging info and nav, like total resources and a next link\n \"\"\"\n paging_controls = self._get_paging_controls(request)\n validator_query = client_transaction_pb2.ClientTransactionListRequest(\n head_id=self._get_head_id(request), transaction_ids=self.\n _get_filter_ids(request), sorting=self._get_sorting_message(request,\n 'default'), paging=self._make_paging_message(paging_controls))\n response = await self._query_validator(Message.\n CLIENT_TRANSACTION_LIST_REQUEST, client_transaction_pb2.\n ClientTransactionListResponse, validator_query)\n data = [self._expand_transaction(t) for t in response['transactions']]\n return self._wrap_paginated_response(request=request, response=response,\n controls=paging_controls, data=data)\n",
"role": "user"
},
{
"content": "list_transactions",
"role": "assistant"
}
] |
async def fetch_transaction(self, request):
"""Fetches a specific transaction from the validator, specified by id.
Request:
path:
- transaction_id: The 128-character id of the txn to be fetched
Response:
data: A JSON object with the data from the expanded Transaction
link: The link to this exact query
"""
error_traps = [error_handlers.TransactionNotFoundTrap]
txn_id = request.match_info.get('transaction_id', '')
self._validate_id(txn_id)
response = await self._query_validator(Message.
CLIENT_TRANSACTION_GET_REQUEST, client_transaction_pb2.
ClientTransactionGetResponse, client_transaction_pb2.
ClientTransactionGetRequest(transaction_id=txn_id), error_traps)
return self._wrap_response(request, data=self._expand_transaction(
response['transaction']), metadata=self._get_metadata(request,
response))
| fetch_transaction | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def fetch_transaction(self, request):\n \"\"\"Fetches a specific transaction from the validator, specified by id.\n\n Request:\n path:\n - transaction_id: The 128-character id of the txn to be fetched\n\n Response:\n data: A JSON object with the data from the expanded Transaction\n link: The link to this exact query\n \"\"\"\n error_traps = [error_handlers.TransactionNotFoundTrap]\n txn_id = request.match_info.get('transaction_id', '')\n self._validate_id(txn_id)\n response = await self._query_validator(Message.\n CLIENT_TRANSACTION_GET_REQUEST, client_transaction_pb2.\n ClientTransactionGetResponse, client_transaction_pb2.\n ClientTransactionGetRequest(transaction_id=txn_id), error_traps)\n return self._wrap_response(request, data=self._expand_transaction(\n response['transaction']), metadata=self._get_metadata(request,\n response))\n",
"role": "user"
},
{
"content": "fetch_transaction",
"role": "assistant"
}
] |
async def list_receipts(self, request):
"""Fetches the receipts for transaction by either a POST or GET.
Request:
body: A JSON array of one or more transaction id strings (if POST)
query:
- id: A comma separated list of up to 15 ids (if GET)
- wait: Request should return as soon as some receipts are
available
Response:
data: A JSON object, with transaction ids as keys, and receipts as
values
link: The /receipts link queried (if GET)
"""
error_traps = [error_handlers.ReceiptNotFoundTrap]
if request.method == 'POST':
if request.headers['Content-Type'] != 'application/json':
LOGGER.debug('Request headers had wrong Content-Type: %s',
request.headers['Content-Type'])
raise errors.ReceiptWrongContentType()
ids = await request.json()
if not ids or not isinstance(ids, list) or not all(isinstance(i,
str) for i in ids):
LOGGER.debug('Request body was invalid: %s', ids)
raise errors.ReceiptBodyInvalid()
for i in ids:
self._validate_id(i)
else:
ids = self._get_filter_ids(request)
if not ids:
LOGGER.debug('Request for receipts missing id query')
raise errors.ReceiptIdQueryInvalid()
validator_query = client_receipt_pb2.ClientReceiptGetRequest(
transaction_ids=ids)
self._set_wait(request, validator_query)
response = await self._query_validator(Message.
CLIENT_RECEIPT_GET_REQUEST, client_receipt_pb2.
ClientReceiptGetResponse, validator_query, error_traps)
if request.method != 'POST':
metadata = self._get_metadata(request, response)
else:
metadata = None
data = self._drop_id_prefixes(self._drop_empty_props(response['receipts']))
return self._wrap_response(request, data=data, metadata=metadata)
| list_receipts | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def list_receipts(self, request):\n \"\"\"Fetches the receipts for transaction by either a POST or GET.\n\n Request:\n body: A JSON array of one or more transaction id strings (if POST)\n query:\n - id: A comma separated list of up to 15 ids (if GET)\n - wait: Request should return as soon as some receipts are\n available\n\n Response:\n data: A JSON object, with transaction ids as keys, and receipts as\n values\n link: The /receipts link queried (if GET)\n \"\"\"\n error_traps = [error_handlers.ReceiptNotFoundTrap]\n if request.method == 'POST':\n if request.headers['Content-Type'] != 'application/json':\n LOGGER.debug('Request headers had wrong Content-Type: %s',\n request.headers['Content-Type'])\n raise errors.ReceiptWrongContentType()\n ids = await request.json()\n if not ids or not isinstance(ids, list) or not all(isinstance(i,\n str) for i in ids):\n LOGGER.debug('Request body was invalid: %s', ids)\n raise errors.ReceiptBodyInvalid()\n for i in ids:\n self._validate_id(i)\n else:\n ids = self._get_filter_ids(request)\n if not ids:\n LOGGER.debug('Request for receipts missing id query')\n raise errors.ReceiptIdQueryInvalid()\n validator_query = client_receipt_pb2.ClientReceiptGetRequest(\n transaction_ids=ids)\n self._set_wait(request, validator_query)\n response = await self._query_validator(Message.\n CLIENT_RECEIPT_GET_REQUEST, client_receipt_pb2.\n ClientReceiptGetResponse, validator_query, error_traps)\n if request.method != 'POST':\n metadata = self._get_metadata(request, response)\n else:\n metadata = None\n data = self._drop_id_prefixes(self._drop_empty_props(response['receipts']))\n return self._wrap_response(request, data=data, metadata=metadata)\n",
"role": "user"
},
{
"content": "list_receipts",
"role": "assistant"
}
] |
async def fetch_peers(self, request):
"""Fetches the peers from the validator.
Request:
Response:
data: JSON array of peer endpoints
link: The link to this exact query
"""
response = await self._query_validator(Message.CLIENT_PEERS_GET_REQUEST,
client_peers_pb2.ClientPeersGetResponse, client_peers_pb2.
ClientPeersGetRequest())
return self._wrap_response(request, data=response['peers'], metadata=
self._get_metadata(request, response))
| fetch_peers | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def fetch_peers(self, request):\n \"\"\"Fetches the peers from the validator.\n Request:\n\n Response:\n data: JSON array of peer endpoints\n link: The link to this exact query\n \"\"\"\n response = await self._query_validator(Message.CLIENT_PEERS_GET_REQUEST,\n client_peers_pb2.ClientPeersGetResponse, client_peers_pb2.\n ClientPeersGetRequest())\n return self._wrap_response(request, data=response['peers'], metadata=\n self._get_metadata(request, response))\n",
"role": "user"
},
{
"content": "fetch_peers",
"role": "assistant"
}
] |
async def fetch_status(self, request):
"""Fetches information pertaining to the valiator's status."""
response = await self._query_validator(Message.
CLIENT_STATUS_GET_REQUEST, client_status_pb2.
ClientStatusGetResponse, client_status_pb2.ClientStatusGetRequest())
return self._wrap_response(request, data={'peers': response['peers'],
'endpoint': response['endpoint']}, metadata=self._get_metadata(
request, response))
| fetch_status | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def fetch_status(self, request):\n \"\"\"Fetches information pertaining to the valiator's status.\"\"\"\n response = await self._query_validator(Message.\n CLIENT_STATUS_GET_REQUEST, client_status_pb2.\n ClientStatusGetResponse, client_status_pb2.ClientStatusGetRequest())\n return self._wrap_response(request, data={'peers': response['peers'],\n 'endpoint': response['endpoint']}, metadata=self._get_metadata(\n request, response))\n",
"role": "user"
},
{
"content": "fetch_status",
"role": "assistant"
}
] |
async def _query_validator(self, request_type, response_proto, payload,
error_traps=None):
"""Sends a request to the validator and parses the response.
"""
LOGGER.debug('Sending %s request to validator', self._get_type_name(
request_type))
payload_bytes = payload.SerializeToString()
response = await self._send_request(request_type, payload_bytes)
content = self._parse_response(response_proto, response)
LOGGER.debug('Received %s response from validator with status %s', self
._get_type_name(response.message_type), self._get_status_name(
response_proto, content.status))
self._check_status_errors(response_proto, content, error_traps)
return self._message_to_dict(content)
| query_validator | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def _query_validator(self, request_type, response_proto, payload,\n error_traps=None):\n \"\"\"Sends a request to the validator and parses the response.\n \"\"\"\n LOGGER.debug('Sending %s request to validator', self._get_type_name(\n request_type))\n payload_bytes = payload.SerializeToString()\n response = await self._send_request(request_type, payload_bytes)\n content = self._parse_response(response_proto, response)\n LOGGER.debug('Received %s response from validator with status %s', self\n ._get_type_name(response.message_type), self._get_status_name(\n response_proto, content.status))\n self._check_status_errors(response_proto, content, error_traps)\n return self._message_to_dict(content)\n",
"role": "user"
},
{
"content": "query_validator",
"role": "assistant"
}
] |
async def _send_request(self, request_type, payload):
"""Uses an executor to send an asynchronous ZMQ request to the
validator with the handler's Connection
"""
try:
return await self._connection.send(message_type=request_type,
message_content=payload, timeout=self._timeout)
except DisconnectError:
LOGGER.warning('Validator disconnected while waiting for response')
raise errors.ValidatorDisconnected()
except asyncio.TimeoutError:
LOGGER.warning('Timed out while waiting for validator response')
raise errors.ValidatorTimedOut()
except SendBackoffTimeoutError:
LOGGER.warning('Failed sending message - Backoff timed out')
raise errors.SendBackoffTimeout()
| send_request | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def _send_request(self, request_type, payload):\n \"\"\"Uses an executor to send an asynchronous ZMQ request to the\n validator with the handler's Connection\n \"\"\"\n try:\n return await self._connection.send(message_type=request_type,\n message_content=payload, timeout=self._timeout)\n except DisconnectError:\n LOGGER.warning('Validator disconnected while waiting for response')\n raise errors.ValidatorDisconnected()\n except asyncio.TimeoutError:\n LOGGER.warning('Timed out while waiting for validator response')\n raise errors.ValidatorTimedOut()\n except SendBackoffTimeoutError:\n LOGGER.warning('Failed sending message - Backoff timed out')\n raise errors.SendBackoffTimeout()\n",
"role": "user"
},
{
"content": "send_request",
"role": "assistant"
}
] |
def x(proto, response):
try:
content = proto()
content.ParseFromString(response.content)
return content
except (DecodeError, AttributeError):
LOGGER.error('Validator response was not parsable: %s', response)
raise errors.ValidatorResponseInvalid()
| parse_response | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(proto, response):\n try:\n content = proto()\n content.ParseFromString(response.content)\n return content\n except (DecodeError, AttributeError):\n LOGGER.error('Validator response was not parsable: %s', response)\n raise errors.ValidatorResponseInvalid()\n",
"role": "user"
},
{
"content": "parse_response",
"role": "assistant"
}
] |
def x(proto, content, error_traps=None):
if content.status == proto.OK:
return
try:
if content.status == proto.INTERNAL_ERROR:
raise errors.UnknownValidatorError()
except AttributeError:
pass
try:
if content.status == proto.NOT_READY:
raise errors.ValidatorNotReady()
except AttributeError:
pass
try:
if content.status == proto.NO_ROOT:
raise errors.HeadNotFound()
except AttributeError:
pass
try:
if content.status == proto.INVALID_PAGING:
raise errors.PagingInvalid()
except AttributeError:
pass
try:
if content.status == proto.INVALID_SORT:
raise errors.SortInvalid()
except AttributeError:
pass
if error_traps is not None:
for trap in error_traps:
trap.check(content.status)
| check_status_errors | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(proto, content, error_traps=None):\n if content.status == proto.OK:\n return\n try:\n if content.status == proto.INTERNAL_ERROR:\n raise errors.UnknownValidatorError()\n except AttributeError:\n pass\n try:\n if content.status == proto.NOT_READY:\n raise errors.ValidatorNotReady()\n except AttributeError:\n pass\n try:\n if content.status == proto.NO_ROOT:\n raise errors.HeadNotFound()\n except AttributeError:\n pass\n try:\n if content.status == proto.INVALID_PAGING:\n raise errors.PagingInvalid()\n except AttributeError:\n pass\n try:\n if content.status == proto.INVALID_SORT:\n raise errors.SortInvalid()\n except AttributeError:\n pass\n if error_traps is not None:\n for trap in error_traps:\n trap.check(content.status)\n",
"role": "user"
},
{
"content": "check_status_errors",
"role": "assistant"
}
] |
def x(request, data=None, metadata=None, status=200):
envelope = metadata or {}
if data is not None:
envelope['data'] = data
return web.Response(status=status, content_type='application/json',
text=json.dumps(envelope, indent=2, separators=(',', ': '),
sort_keys=True))
| wrap_response | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(request, data=None, metadata=None, status=200):\n envelope = metadata or {}\n if data is not None:\n envelope['data'] = data\n return web.Response(status=status, content_type='application/json',\n text=json.dumps(envelope, indent=2, separators=(',', ': '),\n sort_keys=True))\n",
"role": "user"
},
{
"content": "wrap_response",
"role": "assistant"
}
] |
def x(cls, request, response, controls, data, head=None):
paging_response = response['paging']
if head is None:
head = response['head_id']
link = cls._build_url(request, head=head, start=paging_response['start'
], limit=paging_response['limit'])
paging = {}
limit = controls.get('limit')
start = controls.get('start')
paging['limit'] = limit
paging['start'] = start
if paging_response.get('next') == '':
return cls._wrap_response(request, data=data, metadata={'head':
head, 'link': link, 'paging': paging})
next_id = paging_response['next']
paging['next_position'] = next_id
def x(start=None):
return cls._build_url(request, head=head, limit=limit, start=start)
paging['next'] = build_pg_url(paging_response['next'])
return cls._wrap_response(request, data=data, metadata={'head': head,
'link': link, 'paging': paging})
| wrap_paginated_response | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, request, response, controls, data, head=None):\n paging_response = response['paging']\n if head is None:\n head = response['head_id']\n link = cls._build_url(request, head=head, start=paging_response['start'\n ], limit=paging_response['limit'])\n paging = {}\n limit = controls.get('limit')\n start = controls.get('start')\n paging['limit'] = limit\n paging['start'] = start\n if paging_response.get('next') == '':\n return cls._wrap_response(request, data=data, metadata={'head':\n head, 'link': link, 'paging': paging})\n next_id = paging_response['next']\n paging['next_position'] = next_id\n\n def x(start=None):\n return cls._build_url(request, head=head, limit=limit, start=start)\n paging['next'] = build_pg_url(paging_response['next'])\n return cls._wrap_response(request, data=data, metadata={'head': head,\n 'link': link, 'paging': paging})\n",
"role": "user"
},
{
"content": "wrap_paginated_response",
"role": "assistant"
}
] |
def x(cls, request, response, head=None):
head = response.get('head_id', head)
metadata = {'link': cls._build_url(request, head=head)}
if head is not None:
metadata['head'] = head
return metadata
| get_metadata | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, request, response, head=None):\n head = response.get('head_id', head)\n metadata = {'link': cls._build_url(request, head=head)}\n if head is not None:\n metadata['head'] = head\n return metadata\n",
"role": "user"
},
{
"content": "get_metadata",
"role": "assistant"
}
] |
def x(cls, request, path=None, **changes):
changes = {k: v for k, v in changes.items() if v is not None}
queries = {**request.url.query, **changes}
queries = {k: v for k, v in queries.items() if v is not False}
query_strings = []
def x(key):
query_strings.append('{}={}'.format(key, queries[key]) if queries[
key] != '' else key)
def x(key):
queries.pop(key, None)
if 'head' in queries:
add_query('head')
del_query('head')
if 'start' in changes:
add_query('start')
elif 'start' in queries:
add_query('start')
del_query('start')
if 'limit' in queries:
add_query('limit')
del_query('limit')
for key in sorted(queries):
add_query(key)
scheme = cls._get_forwarded(request, 'proto') or request.url.scheme
host = cls._get_forwarded(request, 'host') or request.host
forwarded_path = cls._get_forwarded(request, 'path')
path = path if path is not None else request.path
query = '?' + '&'.join(query_strings) if query_strings else ''
url = '{}://{}{}{}{}'.format(scheme, host, forwarded_path, path, query)
return url
| build_url | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, request, path=None, **changes):\n changes = {k: v for k, v in changes.items() if v is not None}\n queries = {**request.url.query, **changes}\n queries = {k: v for k, v in queries.items() if v is not False}\n query_strings = []\n\n def x(key):\n query_strings.append('{}={}'.format(key, queries[key]) if queries[\n key] != '' else key)\n\n def x(key):\n queries.pop(key, None)\n if 'head' in queries:\n add_query('head')\n del_query('head')\n if 'start' in changes:\n add_query('start')\n elif 'start' in queries:\n add_query('start')\n del_query('start')\n if 'limit' in queries:\n add_query('limit')\n del_query('limit')\n for key in sorted(queries):\n add_query(key)\n scheme = cls._get_forwarded(request, 'proto') or request.url.scheme\n host = cls._get_forwarded(request, 'host') or request.host\n forwarded_path = cls._get_forwarded(request, 'path')\n path = path if path is not None else request.path\n query = '?' + '&'.join(query_strings) if query_strings else ''\n url = '{}://{}{}{}{}'.format(scheme, host, forwarded_path, path, query)\n return url\n",
"role": "user"
},
{
"content": "build_url",
"role": "assistant"
}
] |
def x(request, key):
forwarded = request.headers.get('Forwarded', '')
match = re.search('(?<={}=).+?(?=[\\s,;]|$)'.format(key), forwarded, re
.IGNORECASE)
if match is not None:
header = match.group(0)
if header[0] == '"' and header[-1] == '"':
return header[1:-1]
return header
return request.headers.get('X-Forwarded-{}'.format(key.title()), '')
| get_forwarded | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(request, key):\n forwarded = request.headers.get('Forwarded', '')\n match = re.search('(?<={}=).+?(?=[\\\\s,;]|$)'.format(key), forwarded, re\n .IGNORECASE)\n if match is not None:\n header = match.group(0)\n if header[0] == '\"' and header[-1] == '\"':\n return header[1:-1]\n return header\n return request.headers.get('X-Forwarded-{}'.format(key.title()), '')\n",
"role": "user"
},
{
"content": "get_forwarded",
"role": "assistant"
}
] |
def x(cls, block):
cls._parse_header(BlockHeader, block)
if 'batches' in block:
block['batches'] = [cls._expand_batch(b) for b in block['batches']]
return block
| expand_block | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, block):\n cls._parse_header(BlockHeader, block)\n if 'batches' in block:\n block['batches'] = [cls._expand_batch(b) for b in block['batches']]\n return block\n",
"role": "user"
},
{
"content": "expand_block",
"role": "assistant"
}
] |
def x(cls, batch):
cls._parse_header(BatchHeader, batch)
if 'transactions' in batch:
batch['transactions'] = [cls._expand_transaction(t) for t in batch[
'transactions']]
return batch
| expand_batch | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, batch):\n cls._parse_header(BatchHeader, batch)\n if 'transactions' in batch:\n batch['transactions'] = [cls._expand_transaction(t) for t in batch[\n 'transactions']]\n return batch\n",
"role": "user"
},
{
"content": "expand_batch",
"role": "assistant"
}
] |
def x(cls, header_proto, resource):
header = header_proto()
try:
header_bytes = base64.b64decode(resource['header'])
header.ParseFromString(header_bytes)
except (KeyError, TypeError, ValueError, DecodeError):
header = resource.get('header', None)
LOGGER.error('The validator sent a resource with %s %s',
'a missing header' if header is None else 'an invalid header:',
header or '')
raise errors.ResourceHeaderInvalid()
resource['header'] = cls._message_to_dict(header)
return resource
| parse_header | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, header_proto, resource):\n header = header_proto()\n try:\n header_bytes = base64.b64decode(resource['header'])\n header.ParseFromString(header_bytes)\n except (KeyError, TypeError, ValueError, DecodeError):\n header = resource.get('header', None)\n LOGGER.error('The validator sent a resource with %s %s', \n 'a missing header' if header is None else 'an invalid header:',\n header or '')\n raise errors.ResourceHeaderInvalid()\n resource['header'] = cls._message_to_dict(header)\n return resource\n",
"role": "user"
},
{
"content": "parse_header",
"role": "assistant"
}
] |
def x(request):
start = request.url.query.get('start', None)
limit = request.url.query.get('limit', None)
controls = {}
if limit is not None:
try:
controls['limit'] = int(limit)
except ValueError:
LOGGER.debug('Request query had an invalid limit: %s', limit)
raise errors.CountInvalid()
if controls['limit'] <= 0:
LOGGER.debug('Request query had an invalid limit: %s', limit)
raise errors.CountInvalid()
if start is not None:
controls['start'] = start
return controls
| get_paging_controls | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(request):\n start = request.url.query.get('start', None)\n limit = request.url.query.get('limit', None)\n controls = {}\n if limit is not None:\n try:\n controls['limit'] = int(limit)\n except ValueError:\n LOGGER.debug('Request query had an invalid limit: %s', limit)\n raise errors.CountInvalid()\n if controls['limit'] <= 0:\n LOGGER.debug('Request query had an invalid limit: %s', limit)\n raise errors.CountInvalid()\n if start is not None:\n controls['start'] = start\n return controls\n",
"role": "user"
},
{
"content": "get_paging_controls",
"role": "assistant"
}
] |
def x(controls):
return client_list_control_pb2.ClientPagingControls(start=controls.get(
'start', None), limit=controls.get('limit', None))
| make_paging_message | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(controls):\n return client_list_control_pb2.ClientPagingControls(start=controls.get(\n 'start', None), limit=controls.get('limit', None))\n",
"role": "user"
},
{
"content": "make_paging_message",
"role": "assistant"
}
] |
def x(request, key):
control_list = []
reverse = request.url.query.get('reverse', None)
if reverse is None:
return control_list
if reverse.lower() == '':
control_list.append(client_list_control_pb2.ClientSortControls(
reverse=True, keys=key.split(',')))
elif reverse.lower() != 'false':
control_list.append(client_list_control_pb2.ClientSortControls(
reverse=True, keys=reverse.split(',')))
return control_list
| get_sorting_message | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(request, key):\n control_list = []\n reverse = request.url.query.get('reverse', None)\n if reverse is None:\n return control_list\n if reverse.lower() == '':\n control_list.append(client_list_control_pb2.ClientSortControls(\n reverse=True, keys=key.split(',')))\n elif reverse.lower() != 'false':\n control_list.append(client_list_control_pb2.ClientSortControls(\n reverse=True, keys=reverse.split(',')))\n return control_list\n",
"role": "user"
},
{
"content": "get_sorting_message",
"role": "assistant"
}
] |
def x(self, request, validator_query):
wait = request.url.query.get('wait', 'false')
if wait.lower() != 'false':
validator_query.wait = True
try:
validator_query.timeout = int(wait)
except ValueError:
validator_query.timeout = int(self._timeout * 0.95)
| set_wait | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, request, validator_query):\n wait = request.url.query.get('wait', 'false')\n if wait.lower() != 'false':\n validator_query.wait = True\n try:\n validator_query.timeout = int(wait)\n except ValueError:\n validator_query.timeout = int(self._timeout * 0.95)\n",
"role": "user"
},
{
"content": "set_wait",
"role": "assistant"
}
] |
def x(self, item):
if isinstance(item, list):
return [self._drop_empty_props(i) for i in item]
if isinstance(item, dict):
return {k: self._drop_empty_props(v) for k, v in item.items() if v !=
''}
return item
| drop_empty_props | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, item):\n if isinstance(item, list):\n return [self._drop_empty_props(i) for i in item]\n if isinstance(item, dict):\n return {k: self._drop_empty_props(v) for k, v in item.items() if v !=\n ''}\n return item\n",
"role": "user"
},
{
"content": "drop_empty_props",
"role": "assistant"
}
] |
def x(self, item):
if isinstance(item, list):
return [self._drop_id_prefixes(i) for i in item]
if isinstance(item, dict):
return {('id' if k.endswith('id') else k): self._drop_id_prefixes(v
) for k, v in item.items()}
return item
| drop_id_prefixes | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, item):\n if isinstance(item, list):\n return [self._drop_id_prefixes(i) for i in item]\n if isinstance(item, dict):\n return {('id' if k.endswith('id') else k): self._drop_id_prefixes(v\n ) for k, v in item.items()}\n return item\n",
"role": "user"
},
{
"content": "drop_id_prefixes",
"role": "assistant"
}
] |
def x(cls, request):
head_id = request.url.query.get('head', None)
if head_id is not None:
cls._validate_id(head_id)
return head_id
| get_head_id | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, request):\n head_id = request.url.query.get('head', None)\n if head_id is not None:\n cls._validate_id(head_id)\n return head_id\n",
"role": "user"
},
{
"content": "get_head_id",
"role": "assistant"
}
] |
def x(cls, request):
id_query = request.url.query.get('id', None)
if id_query is None:
return None
filter_ids = id_query.split(',')
for filter_id in filter_ids:
cls._validate_id(filter_id)
return filter_ids
| get_filter_ids | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, request):\n id_query = request.url.query.get('id', None)\n if id_query is None:\n return None\n filter_ids = id_query.split(',')\n for filter_id in filter_ids:\n cls._validate_id(filter_id)\n return filter_ids\n",
"role": "user"
},
{
"content": "get_filter_ids",
"role": "assistant"
}
] |
def x(self, state_root_hash=None):
if state_root_hash is None:
state_root_hash = INIT_ROOT_KEY
merkle_db = MerkleDatabase(self._database, merkle_root=state_root_hash)
return StateView(merkle_db)
| create_view | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, state_root_hash=None):\n if state_root_hash is None:\n state_root_hash = INIT_ROOT_KEY\n merkle_db = MerkleDatabase(self._database, merkle_root=state_root_hash)\n return StateView(merkle_db)\n",
"role": "user"
},
{
"content": "create_view",
"role": "assistant"
}
] |
def x(self, connection_id, subscriptions, last_known_block_id):
with self._subscribers_cv:
self._subscribers[connection_id] = EventSubscriber(connection_id,
subscriptions, last_known_block_id)
LOGGER.debug('Added Subscriber %s for %s', connection_id, subscriptions)
| add_subscriber | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, connection_id, subscriptions, last_known_block_id):\n with self._subscribers_cv:\n self._subscribers[connection_id] = EventSubscriber(connection_id,\n subscriptions, last_known_block_id)\n LOGGER.debug('Added Subscriber %s for %s', connection_id, subscriptions)\n",
"role": "user"
},
{
"content": "add_subscriber",
"role": "assistant"
}
] |
def x(self, connection_id):
with self._subscribers_cv:
subscriber = self._subscribers[connection_id]
last_known_block_id = subscriber.get_last_known_block_id()
subscriptions = subscriber.subscriptions
if last_known_block_id is not None:
LOGGER.debug('Catching up Subscriber %s from %s', connection_id,
last_known_block_id)
for block_id in self.get_catchup_block_ids(last_known_block_id):
events = self.get_events_for_block_id(block_id, subscriptions)
event_list = EventList(events=events)
self._send(connection_id, event_list.SerializeToString())
| catchup_subscriber | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, connection_id):\n with self._subscribers_cv:\n subscriber = self._subscribers[connection_id]\n last_known_block_id = subscriber.get_last_known_block_id()\n subscriptions = subscriber.subscriptions\n if last_known_block_id is not None:\n LOGGER.debug('Catching up Subscriber %s from %s', connection_id,\n last_known_block_id)\n for block_id in self.get_catchup_block_ids(last_known_block_id):\n events = self.get_events_for_block_id(block_id, subscriptions)\n event_list = EventList(events=events)\n self._send(connection_id, event_list.SerializeToString())\n",
"role": "user"
},
{
"content": "catchup_subscriber",
"role": "assistant"
}
] |
def x(self, last_known_block_id):
catchup_up_blocks = []
chain_head = self._block_store.chain_head
if chain_head and last_known_block_id != chain_head.identifier:
for block in self._block_store.get_predecessor_iter():
if last_known_block_id != NULL_BLOCK_IDENTIFIER:
if block.identifier == last_known_block_id:
break
catchup_up_blocks.append(block.identifier)
return list(reversed(catchup_up_blocks))
| get_catchup_block_ids | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, last_known_block_id):\n catchup_up_blocks = []\n chain_head = self._block_store.chain_head\n if chain_head and last_known_block_id != chain_head.identifier:\n for block in self._block_store.get_predecessor_iter():\n if last_known_block_id != NULL_BLOCK_IDENTIFIER:\n if block.identifier == last_known_block_id:\n break\n catchup_up_blocks.append(block.identifier)\n return list(reversed(catchup_up_blocks))\n",
"role": "user"
},
{
"content": "get_catchup_block_ids",
"role": "assistant"
}
] |
def x(self, block_ids, subscriptions):
blocks = [self._block_store[block_id] for block_id in block_ids]
return self.get_events_for_blocks(blocks, subscriptions)
| get_events_for_block_ids | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, block_ids, subscriptions):\n blocks = [self._block_store[block_id] for block_id in block_ids]\n return self.get_events_for_blocks(blocks, subscriptions)\n",
"role": "user"
},
{
"content": "get_events_for_block_ids",
"role": "assistant"
}
] |
def x(self, blocks, subscriptions):
events = []
for blkw in blocks:
events.extend(self.get_events_for_block(blkw, subscriptions))
return events
| get_events_for_blocks | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, blocks, subscriptions):\n events = []\n for blkw in blocks:\n events.extend(self.get_events_for_block(blkw, subscriptions))\n return events\n",
"role": "user"
},
{
"content": "get_events_for_blocks",
"role": "assistant"
}
] |
def x(module_name):
module_package = module_name
if module_name == 'genesis':
module_package = (
'sawtooth_validator.journal.consensus.genesis.genesis_consensus')
elif module_name == 'devmode':
module_package = (
'sawtooth_validator.journal.consensus.dev_mode.dev_mode_consensus')
try:
return importlib.import_module(module_package)
except ImportError:
raise UnknownConsensusModuleError(
'Consensus module "{}" does not exist.'.format(module_name))
| get_consensus_module | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(module_name):\n module_package = module_name\n if module_name == 'genesis':\n module_package = (\n 'sawtooth_validator.journal.consensus.genesis.genesis_consensus')\n elif module_name == 'devmode':\n module_package = (\n 'sawtooth_validator.journal.consensus.dev_mode.dev_mode_consensus')\n try:\n return importlib.import_module(module_package)\n except ImportError:\n raise UnknownConsensusModuleError(\n 'Consensus module \"{}\" does not exist.'.format(module_name))\n",
"role": "user"
},
{
"content": "get_consensus_module",
"role": "assistant"
}
] |
def x(block_id, state_view):
settings_view = SettingsView(state_view)
default_consensus = ('genesis' if block_id == NULL_BLOCK_IDENTIFIER else
'devmode')
consensus_module_name = settings_view.get_setting(
'sawtooth.consensus.algorithm', default_value=default_consensus)
return ConsensusFactory.get_consensus_module(consensus_module_name)
| get_configured_consensus_module | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(block_id, state_view):\n settings_view = SettingsView(state_view)\n default_consensus = ('genesis' if block_id == NULL_BLOCK_IDENTIFIER else\n 'devmode')\n consensus_module_name = settings_view.get_setting(\n 'sawtooth.consensus.algorithm', default_value=default_consensus)\n return ConsensusFactory.get_consensus_module(consensus_module_name)\n",
"role": "user"
},
{
"content": "get_configured_consensus_module",
"role": "assistant"
}
] |
def x(args):
parser = argparse.ArgumentParser(description=
'Starts the REST API application and connects to a specified validator.'
)
parser.add_argument('-B', '--bind', help=
'identify host and port for API to run on default: http://localhost:8008)'
, action='append')
parser.add_argument('-C', '--connect', help=
'specify URL to connect to a running validator')
parser.add_argument('-t', '--timeout', help=
'set time (in seconds) to wait for validator response'
)
parser.add_argument('--client-max-size', type=int, help=
'the max size (in bytes) of a request body')
parser.add_argument('-v', '--verbose', action='count', default=0, help=
'enable more verbose output to stderr')
parser.add_argument('--opentsdb-url', help=
'specify host and port for Open TSDB database used for metrics'
)
parser.add_argument('--opentsdb-db', help=
'specify name of database for storing metrics')
try:
version = pkg_resources.get_distribution(DISTRIBUTION_NAME).version
except pkg_resources.DistributionNotFound:
version = 'UNKNOWN'
parser.add_argument('-V', '--version', action='version', version=(
DISTRIBUTION_NAME + ' (Hyperledger Sawtooth) version {}').format(
version), help='display version information')
return parser.parse_args(args)
| parse_args | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(args):\n parser = argparse.ArgumentParser(description=\n 'Starts the REST API application and connects to a specified validator.'\n )\n parser.add_argument('-B', '--bind', help=\n 'identify host and port for API to run on default: http://localhost:8008)'\n , action='append')\n parser.add_argument('-C', '--connect', help=\n 'specify URL to connect to a running validator')\n parser.add_argument('-t', '--timeout', help=\n 'set time (in seconds) to wait for validator response'\n )\n parser.add_argument('--client-max-size', type=int, help=\n 'the max size (in bytes) of a request body')\n parser.add_argument('-v', '--verbose', action='count', default=0, help=\n 'enable more verbose output to stderr')\n parser.add_argument('--opentsdb-url', help=\n 'specify host and port for Open TSDB database used for metrics'\n )\n parser.add_argument('--opentsdb-db', help=\n 'specify name of database for storing metrics')\n try:\n version = pkg_resources.get_distribution(DISTRIBUTION_NAME).version\n except pkg_resources.DistributionNotFound:\n version = 'UNKNOWN'\n parser.add_argument('-V', '--version', action='version', version=(\n DISTRIBUTION_NAME + ' (Hyperledger Sawtooth) version {}').format(\n version), help='display version information')\n return parser.parse_args(args)\n",
"role": "user"
},
{
"content": "parse_args",
"role": "assistant"
}
] |
def x(host, port, connection, timeout, registry, client_max_size=None):
loop = asyncio.get_event_loop()
connection.open()
app = web.Application(loop=loop, client_max_size=client_max_size)
app.on_cleanup.append(lambda app: connection.close())
LOGGER.info('Creating handlers for validator at %s', connection.url)
handler = RouteHandler(loop, connection, timeout, registry)
app.router.add_post('/batches', handler.submit_batches)
app.router.add_get('/batch_statuses', handler.list_statuses)
app.router.add_post('/batch_statuses', handler.list_statuses)
app.router.add_get('/state', handler.list_state)
app.router.add_get('/state/{address}', handler.fetch_state)
app.router.add_get('/blocks', handler.list_blocks)
app.router.add_get('/blocks/{block_id}', handler.fetch_block)
app.router.add_get('/batches', handler.list_batches)
app.router.add_get('/batches/{batch_id}', handler.fetch_batch)
app.router.add_get('/transactions', handler.list_transactions)
app.router.add_get('/transactions/{transaction_id}', handler.
fetch_transaction)
app.router.add_get('/receipts', handler.list_receipts)
app.router.add_post('/receipts', handler.list_receipts)
app.router.add_get('/peers', handler.fetch_peers)
app.router.add_get('/status', handler.fetch_status)
subscriber_handler = StateDeltaSubscriberHandler(connection)
app.router.add_get('/subscriptions', subscriber_handler.subscriptions)
app.on_shutdown.append(lambda app: subscriber_handler.on_shutdown())
LOGGER.info('Starting REST API on %s:%s', host, port)
web.run_app(app, host=host, port=port, access_log=LOGGER,
access_log_format='%r: %s status, %b size, in %Tf s')
| start_rest_api | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(host, port, connection, timeout, registry, client_max_size=None):\n loop = asyncio.get_event_loop()\n connection.open()\n app = web.Application(loop=loop, client_max_size=client_max_size)\n app.on_cleanup.append(lambda app: connection.close())\n LOGGER.info('Creating handlers for validator at %s', connection.url)\n handler = RouteHandler(loop, connection, timeout, registry)\n app.router.add_post('/batches', handler.submit_batches)\n app.router.add_get('/batch_statuses', handler.list_statuses)\n app.router.add_post('/batch_statuses', handler.list_statuses)\n app.router.add_get('/state', handler.list_state)\n app.router.add_get('/state/{address}', handler.fetch_state)\n app.router.add_get('/blocks', handler.list_blocks)\n app.router.add_get('/blocks/{block_id}', handler.fetch_block)\n app.router.add_get('/batches', handler.list_batches)\n app.router.add_get('/batches/{batch_id}', handler.fetch_batch)\n app.router.add_get('/transactions', handler.list_transactions)\n app.router.add_get('/transactions/{transaction_id}', handler.\n fetch_transaction)\n app.router.add_get('/receipts', handler.list_receipts)\n app.router.add_post('/receipts', handler.list_receipts)\n app.router.add_get('/peers', handler.fetch_peers)\n app.router.add_get('/status', handler.fetch_status)\n subscriber_handler = StateDeltaSubscriberHandler(connection)\n app.router.add_get('/subscriptions', subscriber_handler.subscriptions)\n app.on_shutdown.append(lambda app: subscriber_handler.on_shutdown())\n LOGGER.info('Starting REST API on %s:%s', host, port)\n web.run_app(app, host=host, port=port, access_log=LOGGER,\n access_log_format='%r: %s status, %b size, in %Tf s')\n",
"role": "user"
},
{
"content": "start_rest_api",
"role": "assistant"
}
] |
def x(subparsers, parent_parser):
parser = subparsers.add_parser('genesis', help=
'Creates the genesis.batch file for initializing the validator',
description=
'Generates the genesis.batch file for initializing the validator.',
epilog=
'This command generates a serialized GenesisData protobuf message and stores it in the genesis.batch file. One or more input files (optional) can contain serialized BatchList protobuf messages to add to the GenesisData. The output shows the location of this file. By default, the genesis.batch file is stored in /var/lib/sawtooth. If $SAWTOOTH_HOME is set, the location is $SAWTOOTH_HOME/data/genesis.batch. Use the --output option to change the name of the file.'
, parents=[parent_parser])
parser.add_argument('-o', '--output', type=str, help=
'choose the output file for GenesisData')
parser.add_argument('input_file', nargs='*', type=str, help=
'file or files containing batches to add to the resulting GenesisData')
| add_genesis_parser | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(subparsers, parent_parser):\n parser = subparsers.add_parser('genesis', help=\n 'Creates the genesis.batch file for initializing the validator',\n description=\n 'Generates the genesis.batch file for initializing the validator.',\n epilog=\n 'This command generates a serialized GenesisData protobuf message and stores it in the genesis.batch file. One or more input files (optional) can contain serialized BatchList protobuf messages to add to the GenesisData. The output shows the location of this file. By default, the genesis.batch file is stored in /var/lib/sawtooth. If $SAWTOOTH_HOME is set, the location is $SAWTOOTH_HOME/data/genesis.batch. Use the --output option to change the name of the file.'\n , parents=[parent_parser])\n parser.add_argument('-o', '--output', type=str, help=\n 'choose the output file for GenesisData')\n parser.add_argument('input_file', nargs='*', type=str, help=\n 'file or files containing batches to add to the resulting GenesisData')\n",
"role": "user"
},
{
"content": "add_genesis_parser",
"role": "assistant"
}
] |
def x(args, data_dir=None):
if data_dir is None:
data_dir = get_data_dir()
if not os.path.exists(data_dir):
raise CliException('Data directory does not exist: {}'.format(data_dir)
)
genesis_batches = []
for input_file in args.input_file:
print('Processing {}...'.format(input_file))
input_data = BatchList()
try:
with open(input_file, 'rb') as in_file:
input_data.ParseFromString(in_file.read())
except:
raise CliException('Unable to read {}'.format(input_file))
genesis_batches += input_data.batches
_validate_depedencies(genesis_batches)
_check_required_settings(genesis_batches)
if args.output:
genesis_file = args.output
else:
genesis_file = os.path.join(data_dir, 'genesis.batch')
print('Generating {}'.format(genesis_file))
output_data = GenesisData(batches=genesis_batches)
with open(genesis_file, 'wb') as out_file:
out_file.write(output_data.SerializeToString())
| do_genesis | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(args, data_dir=None):\n if data_dir is None:\n data_dir = get_data_dir()\n if not os.path.exists(data_dir):\n raise CliException('Data directory does not exist: {}'.format(data_dir)\n )\n genesis_batches = []\n for input_file in args.input_file:\n print('Processing {}...'.format(input_file))\n input_data = BatchList()\n try:\n with open(input_file, 'rb') as in_file:\n input_data.ParseFromString(in_file.read())\n except:\n raise CliException('Unable to read {}'.format(input_file))\n genesis_batches += input_data.batches\n _validate_depedencies(genesis_batches)\n _check_required_settings(genesis_batches)\n if args.output:\n genesis_file = args.output\n else:\n genesis_file = os.path.join(data_dir, 'genesis.batch')\n print('Generating {}'.format(genesis_file))\n output_data = GenesisData(batches=genesis_batches)\n with open(genesis_file, 'wb') as out_file:\n out_file.write(output_data.SerializeToString())\n",
"role": "user"
},
{
"content": "do_genesis",
"role": "assistant"
}
] |
def x(batches):
transaction_ids = set()
for batch in batches:
for txn in batch.transactions:
txn_header = TransactionHeader()
txn_header.ParseFromString(txn.header)
if txn_header.dependencies:
unsatisfied_deps = [id for id in txn_header.dependencies if
id not in transaction_ids]
if unsatisfied_deps:
raise CliException(
'Unsatisfied dependency in given transactions: {}'.
format(unsatisfied_deps))
transaction_ids.add(txn.header_signature)
| validate_depedencies | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(batches):\n transaction_ids = set()\n for batch in batches:\n for txn in batch.transactions:\n txn_header = TransactionHeader()\n txn_header.ParseFromString(txn.header)\n if txn_header.dependencies:\n unsatisfied_deps = [id for id in txn_header.dependencies if\n id not in transaction_ids]\n if unsatisfied_deps:\n raise CliException(\n 'Unsatisfied dependency in given transactions: {}'.\n format(unsatisfied_deps))\n transaction_ids.add(txn.header_signature)\n",
"role": "user"
},
{
"content": "validate_depedencies",
"role": "assistant"
}
] |
def x(batches):
required_settings = ['sawtooth.consensus.algorithm.name',
'sawtooth.consensus.algorithm.version']
for batch in batches:
for txn in batch.transactions:
txn_header = TransactionHeader()
txn_header.ParseFromString(txn.header)
if txn_header.family_name == 'sawtooth_settings':
settings_payload = SettingsPayload()
settings_payload.ParseFromString(txn.payload)
if settings_payload.action == SettingsPayload.PROPOSE:
proposal = SettingProposal()
proposal.ParseFromString(settings_payload.data)
if proposal.setting in required_settings:
required_settings.remove(proposal.setting)
if required_settings:
raise CliException(
'The following setting(s) are required at genesis, but were not included in the genesis batches: {}'
.format(required_settings))
| check_required_settings | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(batches):\n required_settings = ['sawtooth.consensus.algorithm.name',\n 'sawtooth.consensus.algorithm.version']\n for batch in batches:\n for txn in batch.transactions:\n txn_header = TransactionHeader()\n txn_header.ParseFromString(txn.header)\n if txn_header.family_name == 'sawtooth_settings':\n settings_payload = SettingsPayload()\n settings_payload.ParseFromString(txn.payload)\n if settings_payload.action == SettingsPayload.PROPOSE:\n proposal = SettingProposal()\n proposal.ParseFromString(settings_payload.data)\n if proposal.setting in required_settings:\n required_settings.remove(proposal.setting)\n if required_settings:\n raise CliException(\n 'The following setting(s) are required at genesis, but were not included in the genesis batches: {}'\n .format(required_settings))\n",
"role": "user"
},
{
"content": "check_required_settings",
"role": "assistant"
}
] |
def x(self, key, index=None):
records = self.get_multi([key], index=index)
try:
return records[0][1]
except IndexError:
return None
| get | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, key, index=None):\n records = self.get_multi([key], index=index)\n try:\n return records[0][1]\n except IndexError:\n return None\n",
"role": "user"
},
{
"content": "get",
"role": "assistant"
}
] |
async def on_shutdown(self):
"""
Cleans up any outstanding subscriptions.
"""
await self._unregister_subscriptions()
self._accepting = False
for ws, _ in self._subscribers:
await ws.close(code=aiohttp.WSCloseCode.GOING_AWAY, message=
'Server shutdown')
| on_shutdown | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def on_shutdown(self):\n \"\"\"\n Cleans up any outstanding subscriptions.\n \"\"\"\n await self._unregister_subscriptions()\n self._accepting = False\n for ws, _ in self._subscribers:\n await ws.close(code=aiohttp.WSCloseCode.GOING_AWAY, message=\n 'Server shutdown')\n",
"role": "user"
},
{
"content": "on_shutdown",
"role": "assistant"
}
] |
async def subscriptions(self, request):
"""
Handles requests for new subscription websockets.
Args:
request (aiohttp.Request): the incoming request
Returns:
aiohttp.web.WebSocketResponse: the websocket response, when the
resulting websocket is closed
"""
if not self._accepting:
return web.Response(status=503)
web_sock = web.WebSocketResponse()
await web_sock.prepare(request)
async for msg in web_sock:
if msg.type == aiohttp.WSMsgType.TEXT:
await self._handle_message(web_sock, msg.data)
elif msg.type == aiohttp.WSMsgType.ERROR:
LOGGER.warning('Web socket connection closed with exception %s',
web_sock.exception())
await web_sock.close()
await self._handle_unsubscribe(web_sock)
return web_sock
| subscriptions | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def subscriptions(self, request):\n \"\"\"\n Handles requests for new subscription websockets.\n\n Args:\n request (aiohttp.Request): the incoming request\n\n Returns:\n aiohttp.web.WebSocketResponse: the websocket response, when the\n resulting websocket is closed\n \"\"\"\n if not self._accepting:\n return web.Response(status=503)\n web_sock = web.WebSocketResponse()\n await web_sock.prepare(request)\n async for msg in web_sock:\n if msg.type == aiohttp.WSMsgType.TEXT:\n await self._handle_message(web_sock, msg.data)\n elif msg.type == aiohttp.WSMsgType.ERROR:\n LOGGER.warning('Web socket connection closed with exception %s',\n web_sock.exception())\n await web_sock.close()\n await self._handle_unsubscribe(web_sock)\n return web_sock\n",
"role": "user"
},
{
"content": "subscriptions",
"role": "assistant"
}
] |
def x(self, request, result):
self._in_process_transactions_count.dec()
req = processor_pb2.TpProcessRequest()
req.ParseFromString(request)
response = processor_pb2.TpProcessResponse()
response.ParseFromString(result.content)
processor_type = ProcessorType(req.header.family_name, req.header.
family_version)
self._processor_manager[processor_type].get_processor(result.connection_id
).dec_occupancy()
self._processor_manager.notify()
self._get_tp_process_response_counter(response.Status.Name(response.status)
).inc()
if (result.connection_id in self._open_futures and req.signature in
self._open_futures[result.connection_id]):
del self._open_futures[result.connection_id][req.signature]
if response.status == processor_pb2.TpProcessResponse.OK:
state_sets, state_deletes, events, data = (self._context_manager.
get_execution_results(req.context_id))
state_changes = [transaction_receipt_pb2.StateChange(address=addr,
value=value, type=transaction_receipt_pb2.StateChange.SET) for
addr, value in state_sets.items()] + [transaction_receipt_pb2.
StateChange(address=addr, type=transaction_receipt_pb2.
StateChange.DELETE) for addr in state_deletes]
self._scheduler.set_transaction_execution_result(txn_signature=req.
signature, is_valid=True, context_id=req.context_id,
state_changes=state_changes, events=events, data=data)
elif response.status == processor_pb2.TpProcessResponse.INTERNAL_ERROR:
LOGGER.error(
'Transaction processor internal error: %s (transaction: %s, name: %s, version: %s)'
, response.message, req.signature, req.header.family_name, req.
header.family_version)
if self._scheduler.is_transaction_in_schedule(req.signature):
self._execute(processor_type=processor_type, content=request,
signature=req.signature)
else:
self._context_manager.delete_contexts(context_id_list=[req.context_id])
self._fail_transaction(txn_signature=req.signature, context_id=req.
context_id, error_message=response.message, error_data=response
.extended_data)
| future_done_callback | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, request, result):\n self._in_process_transactions_count.dec()\n req = processor_pb2.TpProcessRequest()\n req.ParseFromString(request)\n response = processor_pb2.TpProcessResponse()\n response.ParseFromString(result.content)\n processor_type = ProcessorType(req.header.family_name, req.header.\n family_version)\n self._processor_manager[processor_type].get_processor(result.connection_id\n ).dec_occupancy()\n self._processor_manager.notify()\n self._get_tp_process_response_counter(response.Status.Name(response.status)\n ).inc()\n if (result.connection_id in self._open_futures and req.signature in\n self._open_futures[result.connection_id]):\n del self._open_futures[result.connection_id][req.signature]\n if response.status == processor_pb2.TpProcessResponse.OK:\n state_sets, state_deletes, events, data = (self._context_manager.\n get_execution_results(req.context_id))\n state_changes = [transaction_receipt_pb2.StateChange(address=addr,\n value=value, type=transaction_receipt_pb2.StateChange.SET) for \n addr, value in state_sets.items()] + [transaction_receipt_pb2.\n StateChange(address=addr, type=transaction_receipt_pb2.\n StateChange.DELETE) for addr in state_deletes]\n self._scheduler.set_transaction_execution_result(txn_signature=req.\n signature, is_valid=True, context_id=req.context_id,\n state_changes=state_changes, events=events, data=data)\n elif response.status == processor_pb2.TpProcessResponse.INTERNAL_ERROR:\n LOGGER.error(\n 'Transaction processor internal error: %s (transaction: %s, name: %s, version: %s)'\n , response.message, req.signature, req.header.family_name, req.\n header.family_version)\n if self._scheduler.is_transaction_in_schedule(req.signature):\n self._execute(processor_type=processor_type, content=request,\n signature=req.signature)\n else:\n self._context_manager.delete_contexts(context_id_list=[req.context_id])\n self._fail_transaction(txn_signature=req.signature, context_id=req.\n context_id, error_message=response.message, error_data=response\n .extended_data)\n",
"role": "user"
},
{
"content": "future_done_callback",
"role": "assistant"
}
] |
def x(subparsers, parent_parser):
parser = subparsers.add_parser('batch', help=
'Displays information about batches and submit new batches',
description=
'Provides subcommands to display Batch information and submit Batches to the validator via the REST API.'
)
grand_parsers = parser.add_subparsers(title='subcommands', dest=
'subcommand')
grand_parsers.required = True
add_batch_list_parser(grand_parsers, parent_parser)
add_batch_show_parser(grand_parsers, parent_parser)
add_batch_status_parser(grand_parsers, parent_parser)
add_batch_submit_parser(grand_parsers, parent_parser)
| add_batch_parser | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(subparsers, parent_parser):\n parser = subparsers.add_parser('batch', help=\n 'Displays information about batches and submit new batches',\n description=\n 'Provides subcommands to display Batch information and submit Batches to the validator via the REST API.'\n )\n grand_parsers = parser.add_subparsers(title='subcommands', dest=\n 'subcommand')\n grand_parsers.required = True\n add_batch_list_parser(grand_parsers, parent_parser)\n add_batch_show_parser(grand_parsers, parent_parser)\n add_batch_status_parser(grand_parsers, parent_parser)\n add_batch_submit_parser(grand_parsers, parent_parser)\n",
"role": "user"
},
{
"content": "add_batch_parser",
"role": "assistant"
}
] |
def x(args):
if args.subcommand == 'list':
do_batch_list(args)
if args.subcommand == 'show':
do_batch_show(args)
if args.subcommand == 'status':
do_batch_status(args)
if args.subcommand == 'submit':
do_batch_submit(args)
| do_batch | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(args):\n if args.subcommand == 'list':\n do_batch_list(args)\n if args.subcommand == 'show':\n do_batch_show(args)\n if args.subcommand == 'status':\n do_batch_status(args)\n if args.subcommand == 'submit':\n do_batch_submit(args)\n",
"role": "user"
},
{
"content": "do_batch",
"role": "assistant"
}
] |
def x(args):
rest_client = RestClient(args.url, args.user)
batch_ids = args.batch_ids.split(',')
if args.wait and args.wait > 0:
statuses = rest_client.get_statuses(batch_ids, args.wait)
else:
statuses = rest_client.get_statuses(batch_ids)
if args.format == 'yaml':
fmt.print_yaml(statuses)
elif args.format == 'json':
fmt.print_json(statuses)
else:
raise AssertionError('Missing handler: {}'.format(args.format))
| do_batch_status | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(args):\n rest_client = RestClient(args.url, args.user)\n batch_ids = args.batch_ids.split(',')\n if args.wait and args.wait > 0:\n statuses = rest_client.get_statuses(batch_ids, args.wait)\n else:\n statuses = rest_client.get_statuses(batch_ids)\n if args.format == 'yaml':\n fmt.print_yaml(statuses)\n elif args.format == 'json':\n fmt.print_json(statuses)\n else:\n raise AssertionError('Missing handler: {}'.format(args.format))\n",
"role": "user"
},
{
"content": "do_batch_status",
"role": "assistant"
}
] |
def x(self, block):
if block.header_signature in self._block_manager:
LOGGER.debug('Drop duplicate block: %s', block)
return None
if block.previous_block_id not in self._block_manager:
return self._request_previous_if_not_already_requested(block)
if len(block.batches) > len(block.header.batch_ids):
LOGGER.debug('Block has extra batches. Dropping %s', block)
return None
temp_batches = {}
for batch in block.batches:
temp_batches[batch.header_signature] = batch
if len(block.batches) != len(block.header.batch_ids):
building = True
for batch_id in block.header.batch_ids:
if (batch_id not in self._batch_cache and batch_id not in
temp_batches):
if batch_id not in self._incomplete_blocks:
self._incomplete_blocks[batch_id] = [block]
elif block not in self._incomplete_blocks[batch_id]:
self._incomplete_blocks[batch_id] += [block]
if batch_id in self._requested:
return None
self._requested[batch_id] = None
self._gossip.broadcast_batch_by_batch_id_request(batch_id)
building = False
if not building:
return None
batches = self._finalize_batch_list(block, temp_batches)
del block.batches[:]
block.batches.extend(batches)
if block.header_signature in self._requested:
del self._requested[block.header_signature]
return self._put_or_request_if_missing_predecessor(block)
batch_id_list = [x.header_signature for x in block.batches]
if batch_id_list == list(block.header.batch_ids):
if block.header_signature in self._requested:
del self._requested[block.header_signature]
return self._put_or_request_if_missing_predecessor(block)
if sorted(batch_id_list) == sorted(list(block.header.batch_ids)):
batches = self._finalize_batch_list(block, temp_batches)
del block.batches[:]
if batches is not None:
block.batches.extend(batches)
else:
return None
if block.header_signature in self._requested:
del self._requested[block.header_signature]
return self._put_or_request_if_missing_predecessor(block)
LOGGER.debug(
'Block.header.batch_ids does not match set of batches in block.batches Dropping %s'
, block)
return None
| complete_block | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, block):\n if block.header_signature in self._block_manager:\n LOGGER.debug('Drop duplicate block: %s', block)\n return None\n if block.previous_block_id not in self._block_manager:\n return self._request_previous_if_not_already_requested(block)\n if len(block.batches) > len(block.header.batch_ids):\n LOGGER.debug('Block has extra batches. Dropping %s', block)\n return None\n temp_batches = {}\n for batch in block.batches:\n temp_batches[batch.header_signature] = batch\n if len(block.batches) != len(block.header.batch_ids):\n building = True\n for batch_id in block.header.batch_ids:\n if (batch_id not in self._batch_cache and batch_id not in\n temp_batches):\n if batch_id not in self._incomplete_blocks:\n self._incomplete_blocks[batch_id] = [block]\n elif block not in self._incomplete_blocks[batch_id]:\n self._incomplete_blocks[batch_id] += [block]\n if batch_id in self._requested:\n return None\n self._requested[batch_id] = None\n self._gossip.broadcast_batch_by_batch_id_request(batch_id)\n building = False\n if not building:\n return None\n batches = self._finalize_batch_list(block, temp_batches)\n del block.batches[:]\n block.batches.extend(batches)\n if block.header_signature in self._requested:\n del self._requested[block.header_signature]\n return self._put_or_request_if_missing_predecessor(block)\n batch_id_list = [x.header_signature for x in block.batches]\n if batch_id_list == list(block.header.batch_ids):\n if block.header_signature in self._requested:\n del self._requested[block.header_signature]\n return self._put_or_request_if_missing_predecessor(block)\n if sorted(batch_id_list) == sorted(list(block.header.batch_ids)):\n batches = self._finalize_batch_list(block, temp_batches)\n del block.batches[:]\n if batches is not None:\n block.batches.extend(batches)\n else:\n return None\n if block.header_signature in self._requested:\n del self._requested[block.header_signature]\n return self._put_or_request_if_missing_predecessor(block)\n LOGGER.debug(\n 'Block.header.batch_ids does not match set of batches in block.batches Dropping %s'\n , block)\n return None\n",
"role": "user"
},
{
"content": "complete_block",
"role": "assistant"
}
] |
def x(args):
rest_client = RestClient(args.url, args.user)
if args.subcommand == 'list':
response = rest_client.list_state(args.subtree, args.head)
leaves = response['data']
head = response['head']
keys = 'address', 'size', 'data'
headers = tuple(k.upper() for k in keys)
def x(leaf, decode=True):
decoded = b64decode(leaf['data'])
return leaf['address'], len(decoded), str(decoded
) if decode else leaf['data']
if args.format == 'default':
fmt.print_terminal_table(headers, leaves, parse_leaf_row)
print('HEAD BLOCK: "{}"'.format(head))
elif args.format == 'csv':
fmt.print_csv(headers, leaves, parse_leaf_row)
print('(data for head block: "{}")'.format(head))
elif args.format == 'json' or args.format == 'yaml':
state_data = {'head': head, 'data': [{k: d for k, d in zip(keys,
parse_leaf_row(l, False))} for l in leaves]}
if args.format == 'yaml':
fmt.print_yaml(state_data)
elif args.format == 'json':
fmt.print_json(state_data)
else:
raise AssertionError('Missing handler: {}'.format(args.format))
else:
raise AssertionError('Missing handler: {}'.format(args.format))
if args.subcommand == 'show':
output = rest_client.get_leaf(args.address, args.head)
if output is not None:
print('DATA: "{}"'.format(b64decode(output['data'])))
print('HEAD: "{}"'.format(output['head']))
else:
raise CliException('No data available at {}'.format(args.address))
| do_state | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(args):\n rest_client = RestClient(args.url, args.user)\n if args.subcommand == 'list':\n response = rest_client.list_state(args.subtree, args.head)\n leaves = response['data']\n head = response['head']\n keys = 'address', 'size', 'data'\n headers = tuple(k.upper() for k in keys)\n\n def x(leaf, decode=True):\n decoded = b64decode(leaf['data'])\n return leaf['address'], len(decoded), str(decoded\n ) if decode else leaf['data']\n if args.format == 'default':\n fmt.print_terminal_table(headers, leaves, parse_leaf_row)\n print('HEAD BLOCK: \"{}\"'.format(head))\n elif args.format == 'csv':\n fmt.print_csv(headers, leaves, parse_leaf_row)\n print('(data for head block: \"{}\")'.format(head))\n elif args.format == 'json' or args.format == 'yaml':\n state_data = {'head': head, 'data': [{k: d for k, d in zip(keys,\n parse_leaf_row(l, False))} for l in leaves]}\n if args.format == 'yaml':\n fmt.print_yaml(state_data)\n elif args.format == 'json':\n fmt.print_json(state_data)\n else:\n raise AssertionError('Missing handler: {}'.format(args.format))\n else:\n raise AssertionError('Missing handler: {}'.format(args.format))\n if args.subcommand == 'show':\n output = rest_client.get_leaf(args.address, args.head)\n if output is not None:\n print('DATA: \"{}\"'.format(b64decode(output['data'])))\n print('HEAD: \"{}\"'.format(output['head']))\n else:\n raise CliException('No data available at {}'.format(args.address))\n",
"role": "user"
},
{
"content": "do_state",
"role": "assistant"
}
] |
def x(self, required_state_root):
state_hash = None
if self._previous_valid_batch_c_id is not None:
publishing_or_genesis = (self._always_persist or
required_state_root is None)
state_hash = self._squash(state_root=self._previous_state_hash,
context_ids=[self._previous_valid_batch_c_id], persist=self.
_always_persist, clean_up=publishing_or_genesis)
if self._always_persist is True:
return state_hash
if state_hash == required_state_root:
self._squash(state_root=self._previous_state_hash, context_ids=
[self._previous_valid_batch_c_id], persist=True, clean_up=True)
return state_hash
| compute_merkle_root | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, required_state_root):\n state_hash = None\n if self._previous_valid_batch_c_id is not None:\n publishing_or_genesis = (self._always_persist or \n required_state_root is None)\n state_hash = self._squash(state_root=self._previous_state_hash,\n context_ids=[self._previous_valid_batch_c_id], persist=self.\n _always_persist, clean_up=publishing_or_genesis)\n if self._always_persist is True:\n return state_hash\n if state_hash == required_state_root:\n self._squash(state_root=self._previous_state_hash, context_ids=\n [self._previous_valid_batch_c_id], persist=True, clean_up=True)\n return state_hash\n",
"role": "user"
},
{
"content": "compute_merkle_root",
"role": "assistant"
}
] |
def x(value):
block = Block()
block.ParseFromString(value)
return BlockWrapper(block=block)
| deserialize_block | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(value):\n block = Block()\n block.ParseFromString(value)\n return BlockWrapper(block=block)\n",
"role": "user"
},
{
"content": "deserialize_block",
"role": "assistant"
}
] |
def x(self):
vec_ptr, vec_len, vec_cap = ffi.prepare_vec_result()
try:
_libexec('commit_store_get_chain_head', self.pointer, ctypes.byref(
vec_ptr), ctypes.byref(vec_len), ctypes.byref(vec_cap))
except ValueError:
return None
return self.deserialize_block(ffi.from_rust_vec(vec_ptr, vec_len, vec_cap))
| chain_head | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n vec_ptr, vec_len, vec_cap = ffi.prepare_vec_result()\n try:\n _libexec('commit_store_get_chain_head', self.pointer, ctypes.byref(\n vec_ptr), ctypes.byref(vec_len), ctypes.byref(vec_cap))\n except ValueError:\n return None\n return self.deserialize_block(ffi.from_rust_vec(vec_ptr, vec_len, vec_cap))\n",
"role": "user"
},
{
"content": "chain_head",
"role": "assistant"
}
] |
def x(self, start_block=None, start_block_num=None, reverse=True):
start = None
if start_block_num:
if len(start_block_num) < 2:
raise ValueError('Invalid start block num')
if start_block_num[:2] != '0x':
raise ValueError('Invalid start block num')
start = int(start_block_num, 16)
elif start_block:
start = start_block.block_num
return _BlockStoreIter(self.pointer, start, reverse)
| get_block_iter | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, start_block=None, start_block_num=None, reverse=True):\n start = None\n if start_block_num:\n if len(start_block_num) < 2:\n raise ValueError('Invalid start block num')\n if start_block_num[:2] != '0x':\n raise ValueError('Invalid start block num')\n start = int(start_block_num, 16)\n elif start_block:\n start = start_block.block_num\n return _BlockStoreIter(self.pointer, start, reverse)\n",
"role": "user"
},
{
"content": "get_block_iter",
"role": "assistant"
}
] |
def x(self, block_ids):
return list(filter(lambda b: b is not None, map(self.
_get_block_by_id_or_none, block_ids)))
| get_blocks | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, block_ids):\n return list(filter(lambda b: b is not None, map(self.\n _get_block_by_id_or_none, block_ids)))\n",
"role": "user"
},
{
"content": "get_blocks",
"role": "assistant"
}
] |
def x(self, transaction_id):
payload = self._get_data_by_id(transaction_id,
'commit_store_get_batch_by_transaction')
batch = Batch()
batch.ParseFromString(payload)
return batch
| get_batch_by_transaction | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, transaction_id):\n payload = self._get_data_by_id(transaction_id,\n 'commit_store_get_batch_by_transaction')\n batch = Batch()\n batch.ParseFromString(payload)\n return batch\n",
"role": "user"
},
{
"content": "get_batch_by_transaction",
"role": "assistant"
}
] |
def x(self, batch_id):
payload = self._get_data_by_id(batch_id, 'commit_store_get_batch')
batch = Batch()
batch.ParseFromString(payload)
return batch
| get_batch | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, batch_id):\n payload = self._get_data_by_id(batch_id, 'commit_store_get_batch')\n batch = Batch()\n batch.ParseFromString(payload)\n return batch\n",
"role": "user"
},
{
"content": "get_batch",
"role": "assistant"
}
] |
def x(self, transaction_id):
payload = self._get_data_by_id(transaction_id,
'commit_store_get_transaction')
txn = Transaction()
txn.ParseFromString(payload)
return txn
| get_transaction | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, transaction_id):\n payload = self._get_data_by_id(transaction_id,\n 'commit_store_get_transaction')\n txn = Transaction()\n txn.ParseFromString(payload)\n return txn\n",
"role": "user"
},
{
"content": "get_transaction",
"role": "assistant"
}
] |
def x(headers, example_row):
def x(col):
if isinstance(col, str):
return '{{:{w}.{w}}}'
return '{{:<{w}}}'
widths = [max(len(h), len(str(d))) for h, d in zip(headers, example_row)]
original_last_width = widths[-1]
if sys.stdout.isatty():
widths[-1] = max(len(headers[-1]), tty.width() - sum(w + 2 for w in
widths[0:-1]) - 3)
cols = [format_column(c).format(w=w) for c, w in zip(example_row, widths)]
format_string = ' '.join(cols)
if original_last_width > widths[-1]:
format_string += '...'
return format_string
| format_terminal_row | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(headers, example_row):\n\n def x(col):\n if isinstance(col, str):\n return '{{:{w}.{w}}}'\n return '{{:<{w}}}'\n widths = [max(len(h), len(str(d))) for h, d in zip(headers, example_row)]\n original_last_width = widths[-1]\n if sys.stdout.isatty():\n widths[-1] = max(len(headers[-1]), tty.width() - sum(w + 2 for w in\n widths[0:-1]) - 3)\n cols = [format_column(c).format(w=w) for c, w in zip(example_row, widths)]\n format_string = ' '.join(cols)\n if original_last_width > widths[-1]:\n format_string += '...'\n return format_string\n",
"role": "user"
},
{
"content": "format_terminal_row",
"role": "assistant"
}
] |
def x(headers, data_list, parse_row_fn):
data_iter = iter(data_list)
try:
example = next(data_iter)
example_row = parse_row_fn(example)
data_iter = itertools.chain([example], data_iter)
except StopIteration:
example_row = [''] * len(headers)
format_string = format_terminal_row(headers, example_row)
top_row = format_string.format(*headers)
print(top_row[0:-3] if top_row.endswith('...') else top_row)
for data in data_iter:
print(format_string.format(*parse_row_fn(data)))
| print_terminal_table | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(headers, data_list, parse_row_fn):\n data_iter = iter(data_list)\n try:\n example = next(data_iter)\n example_row = parse_row_fn(example)\n data_iter = itertools.chain([example], data_iter)\n except StopIteration:\n example_row = [''] * len(headers)\n format_string = format_terminal_row(headers, example_row)\n top_row = format_string.format(*headers)\n print(top_row[0:-3] if top_row.endswith('...') else top_row)\n for data in data_iter:\n print(format_string.format(*parse_row_fn(data)))\n",
"role": "user"
},
{
"content": "print_terminal_table",
"role": "assistant"
}
] |
def x(headers, data_list, parse_row_fn):
try:
writer = csv.writer(sys.stdout)
writer.writerow(headers)
for data in data_list:
writer.writerow(parse_row_fn(data))
except csv.Error as e:
raise CliException('Error writing CSV: {}'.format(e))
| print_csv | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(headers, data_list, parse_row_fn):\n try:\n writer = csv.writer(sys.stdout)\n writer.writerow(headers)\n for data in data_list:\n writer.writerow(parse_row_fn(data))\n except csv.Error as e:\n raise CliException('Error writing CSV: {}'.format(e))\n",
"role": "user"
},
{
"content": "print_csv",
"role": "assistant"
}
] |
def x(self, key, item):
with self._lock:
if key in self._dict:
self._dict[key].append(item)
else:
self._dict[key] = [item]
| append | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, key, item):\n with self._lock:\n if key in self._dict:\n self._dict[key].append(item)\n else:\n self._dict[key] = [item]\n",
"role": "user"
},
{
"content": "append",
"role": "assistant"
}
] |
def x(self, key, items):
if not isinstance(items, list):
raise ValueError('items must be a list')
with self._lock:
self._dict[key] = items.copy()
| set | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, key, items):\n if not isinstance(items, list):\n raise ValueError('items must be a list')\n with self._lock:\n self._dict[key] = items.copy()\n",
"role": "user"
},
{
"content": "set",
"role": "assistant"
}
] |
def x(self, key, items):
if not isinstance(items, list):
raise ValueError('items must be a list')
return_value = []
with self._lock:
if key in self._dict:
return_value = self._dict[key]
self._dict[key] = items.copy()
return return_value
| swap | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, key, items):\n if not isinstance(items, list):\n raise ValueError('items must be a list')\n return_value = []\n with self._lock:\n if key in self._dict:\n return_value = self._dict[key]\n self._dict[key] = items.copy()\n return return_value\n",
"role": "user"
},
{
"content": "swap",
"role": "assistant"
}
] |
def x(self, key, default):
with self._lock:
return self._dict.pop(key, default)
| pop | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, key, default):\n with self._lock:\n return self._dict.pop(key, default)\n",
"role": "user"
},
{
"content": "pop",
"role": "assistant"
}
] |
def x(self, key, default):
with self._lock:
try:
return self._dict[key].copy()
except KeyError:
return default
| get | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, key, default):\n with self._lock:\n try:\n return self._dict[key].copy()\n except KeyError:\n return default\n",
"role": "user"
},
{
"content": "get",
"role": "assistant"
}
] |
def x(self, block, receipts):
with self._lock:
for batch_id in self._pending.copy():
if self._batch_committed(batch_id):
self._pending.remove(batch_id)
self._update_observers(batch_id, ClientBatchStatus.COMMITTED)
| chain_update | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, block, receipts):\n with self._lock:\n for batch_id in self._pending.copy():\n if self._batch_committed(batch_id):\n self._pending.remove(batch_id)\n self._update_observers(batch_id, ClientBatchStatus.COMMITTED)\n",
"role": "user"
},
{
"content": "chain_update",
"role": "assistant"
}
] |
def x(self, txn_id, message=None, extended_data=None):
invalid_txn_info = {'id': txn_id}
if message is not None:
invalid_txn_info['message'] = message
if extended_data is not None:
invalid_txn_info['extended_data'] = extended_data
with self._lock:
for batch_id, txn_ids in self._batch_info.items():
if txn_id in txn_ids:
if batch_id not in self._invalid:
self._invalid[batch_id] = [invalid_txn_info]
else:
self._invalid[batch_id].append(invalid_txn_info)
self._pending.discard(batch_id)
self._update_observers(batch_id, ClientBatchStatus.INVALID)
return
| notify_txn_invalid | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, txn_id, message=None, extended_data=None):\n invalid_txn_info = {'id': txn_id}\n if message is not None:\n invalid_txn_info['message'] = message\n if extended_data is not None:\n invalid_txn_info['extended_data'] = extended_data\n with self._lock:\n for batch_id, txn_ids in self._batch_info.items():\n if txn_id in txn_ids:\n if batch_id not in self._invalid:\n self._invalid[batch_id] = [invalid_txn_info]\n else:\n self._invalid[batch_id].append(invalid_txn_info)\n self._pending.discard(batch_id)\n self._update_observers(batch_id, ClientBatchStatus.INVALID)\n return\n",
"role": "user"
},
{
"content": "notify_txn_invalid",
"role": "assistant"
}
] |
def x(self, batch):
txn_ids = {t.header_signature for t in batch.transactions}
with self._lock:
self._pending.add(batch.header_signature)
self._batch_info[batch.header_signature] = txn_ids
self._update_observers(batch.header_signature, ClientBatchStatus.
PENDING)
| notify_batch_pending | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, batch):\n txn_ids = {t.header_signature for t in batch.transactions}\n with self._lock:\n self._pending.add(batch.header_signature)\n self._batch_info[batch.header_signature] = txn_ids\n self._update_observers(batch.header_signature, ClientBatchStatus.\n PENDING)\n",
"role": "user"
},
{
"content": "notify_batch_pending",
"role": "assistant"
}
] |
def x(self, batch_id):
with self._lock:
if self._batch_committed(batch_id):
return ClientBatchStatus.COMMITTED
if batch_id in self._invalid:
return ClientBatchStatus.INVALID
if batch_id in self._pending:
return ClientBatchStatus.PENDING
return ClientBatchStatus.UNKNOWN
| get_status | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, batch_id):\n with self._lock:\n if self._batch_committed(batch_id):\n return ClientBatchStatus.COMMITTED\n if batch_id in self._invalid:\n return ClientBatchStatus.INVALID\n if batch_id in self._pending:\n return ClientBatchStatus.PENDING\n return ClientBatchStatus.UNKNOWN\n",
"role": "user"
},
{
"content": "get_status",
"role": "assistant"
}
] |
def x(self, batch_ids):
with self._lock:
return {b: self.get_status(b) for b in batch_ids}
| get_statuses | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, batch_ids):\n with self._lock:\n return {b: self.get_status(b) for b in batch_ids}\n",
"role": "user"
},
{
"content": "get_statuses",
"role": "assistant"
}
] |
def x(self, batch_id):
with self._lock:
return [info.copy() for info in self._invalid.get(batch_id, [])]
| get_invalid_txn_info | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, batch_id):\n with self._lock:\n return [info.copy() for info in self._invalid.get(batch_id, [])]\n",
"role": "user"
},
{
"content": "get_invalid_txn_info",
"role": "assistant"
}
] |
def x(self, observer, batch_ids):
with self._lock:
statuses = self.get_statuses(batch_ids)
if self._has_no_pendings(statuses):
observer.notify_batches_finished(statuses)
else:
self._observers[observer] = statuses
| watch_statuses | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, observer, batch_ids):\n with self._lock:\n statuses = self.get_statuses(batch_ids)\n if self._has_no_pendings(statuses):\n observer.notify_batches_finished(statuses)\n else:\n self._observers[observer] = statuses\n",
"role": "user"
},
{
"content": "watch_statuses",
"role": "assistant"
}
] |
def x(self, batch_id, status):
for observer, statuses in self._observers.copy().items():
if batch_id in statuses:
statuses[batch_id] = status
if self._has_no_pendings(statuses):
observer.notify_batches_finished(statuses)
self._observers.pop(observer)
| update_observers | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, batch_id, status):\n for observer, statuses in self._observers.copy().items():\n if batch_id in statuses:\n statuses[batch_id] = status\n if self._has_no_pendings(statuses):\n observer.notify_batches_finished(statuses)\n self._observers.pop(observer)\n",
"role": "user"
},
{
"content": "update_observers",
"role": "assistant"
}
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.