source_code
stringlengths
45
7.24k
function_name
stringlengths
1
111
messages
list
def x(self, name): address = _create_role_address(name) role_list_bytes = None try: role_list_bytes = self._state_view.get(address=address) except KeyError: return None if role_list_bytes is not None: role_list = _create_from_bytes(role_list_bytes, identity_pb2.RoleList) for role in role_list.roles: if role.name == name: return role return None
get_role
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, name):\n address = _create_role_address(name)\n role_list_bytes = None\n try:\n role_list_bytes = self._state_view.get(address=address)\n except KeyError:\n return None\n if role_list_bytes is not None:\n role_list = _create_from_bytes(role_list_bytes, identity_pb2.RoleList)\n for role in role_list.roles:\n if role.name == name:\n return role\n return None\n", "role": "user" }, { "content": "get_role", "role": "assistant" } ]
def x(self): prefix = _IDENTITY_NS + _ROLE_NS rolelist_list = [_create_from_bytes(d, identity_pb2.RoleList) for _, d in self._state_view.leaves(prefix=prefix)] roles = [] for role_list in rolelist_list: for role in role_list.roles: roles.append(role) return sorted(roles, key=lambda r: r.name)
get_roles
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n prefix = _IDENTITY_NS + _ROLE_NS\n rolelist_list = [_create_from_bytes(d, identity_pb2.RoleList) for _, d in\n self._state_view.leaves(prefix=prefix)]\n roles = []\n for role_list in rolelist_list:\n for role in role_list.roles:\n roles.append(role)\n return sorted(roles, key=lambda r: r.name)\n", "role": "user" }, { "content": "get_roles", "role": "assistant" } ]
def x(self, name): address = _create_policy_address(name) policy_list_bytes = None try: policy_list_bytes = self._state_view.get(address=address) except KeyError: return None if policy_list_bytes is not None: policy_list = _create_from_bytes(policy_list_bytes, identity_pb2. PolicyList) for policy in policy_list.policies: if policy.name == name: return policy return None
get_policy
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, name):\n address = _create_policy_address(name)\n policy_list_bytes = None\n try:\n policy_list_bytes = self._state_view.get(address=address)\n except KeyError:\n return None\n if policy_list_bytes is not None:\n policy_list = _create_from_bytes(policy_list_bytes, identity_pb2.\n PolicyList)\n for policy in policy_list.policies:\n if policy.name == name:\n return policy\n return None\n", "role": "user" }, { "content": "get_policy", "role": "assistant" } ]
def x(self): prefix = _IDENTITY_NS + _POLICY_NS policylist_list = [_create_from_bytes(d, identity_pb2.PolicyList) for _, d in self._state_view.leaves(prefix=prefix)] policies = [] for policy_list in policylist_list: for policy in policy_list.policies: policies.append(policy) return sorted(policies, key=lambda p: p.name)
get_policies
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n prefix = _IDENTITY_NS + _POLICY_NS\n policylist_list = [_create_from_bytes(d, identity_pb2.PolicyList) for _,\n d in self._state_view.leaves(prefix=prefix)]\n policies = []\n for policy_list in policylist_list:\n for policy in policy_list.policies:\n policies.append(policy)\n return sorted(policies, key=lambda p: p.name)\n", "role": "user" }, { "content": "get_policies", "role": "assistant" } ]
def x(self, previous_block): previous_header_bytes = previous_block.header previous_header = BlockHeader() previous_header.ParseFromString(previous_header_bytes) block_info = BlockInfo(block_num=previous_header.block_num, previous_block_id=previous_header.previous_block_id, signer_public_key=previous_header.signer_public_key, header_signature=previous_block.header_signature, timestamp=int( time.time())) return [self.create_batch(block_info)]
block_start
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, previous_block):\n previous_header_bytes = previous_block.header\n previous_header = BlockHeader()\n previous_header.ParseFromString(previous_header_bytes)\n block_info = BlockInfo(block_num=previous_header.block_num,\n previous_block_id=previous_header.previous_block_id,\n signer_public_key=previous_header.signer_public_key,\n header_signature=previous_block.header_signature, timestamp=int(\n time.time()))\n return [self.create_batch(block_info)]\n", "role": "user" }, { "content": "block_start", "role": "assistant" } ]
def x(settings_view, expected_signer, batches): rules = settings_view.get_setting( 'sawtooth.validator.block_validation_rules') if rules is None: return True transactions = [] for batch in batches: transactions += batch.transactions rules = rules.split(';') valid = True for rule in rules: try: rule_type, arguments = rule.split(':') except ValueError: LOGGER.warning( 'Validation Rule Ignored, not in the correct format: %s', rule) continue rule_type = rule_type.strip() if rule_type == 'NofX': valid = _do_nofx(transactions, arguments) elif rule_type == 'XatY': valid = _do_xaty(transactions, arguments) elif rule_type == 'local': valid = _do_local(transactions, expected_signer, arguments) if not valid: return False return valid
enforce_validation_rules
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(settings_view, expected_signer, batches):\n rules = settings_view.get_setting(\n 'sawtooth.validator.block_validation_rules')\n if rules is None:\n return True\n transactions = []\n for batch in batches:\n transactions += batch.transactions\n rules = rules.split(';')\n valid = True\n for rule in rules:\n try:\n rule_type, arguments = rule.split(':')\n except ValueError:\n LOGGER.warning(\n 'Validation Rule Ignored, not in the correct format: %s', rule)\n continue\n rule_type = rule_type.strip()\n if rule_type == 'NofX':\n valid = _do_nofx(transactions, arguments)\n elif rule_type == 'XatY':\n valid = _do_xaty(transactions, arguments)\n elif rule_type == 'local':\n valid = _do_local(transactions, expected_signer, arguments)\n if not valid:\n return False\n return valid\n", "role": "user" }, { "content": "enforce_validation_rules", "role": "assistant" } ]
def x(transactions, arguments): try: num, family = arguments.split(',') limit = int(num.strip()) except ValueError: LOGGER.warning( 'Ignore, NofX requires arguments in the format int,family not %s', arguments) return True count = 0 family = family.strip() for txn in transactions: header = TransactionHeader() header.ParseFromString(txn.header) if header.family_name == family: count += 1 if count > limit: LOGGER.debug('Too many transactions of type %s', family) return False return True
do_nofx
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(transactions, arguments):\n try:\n num, family = arguments.split(',')\n limit = int(num.strip())\n except ValueError:\n LOGGER.warning(\n 'Ignore, NofX requires arguments in the format int,family not %s',\n arguments)\n return True\n count = 0\n family = family.strip()\n for txn in transactions:\n header = TransactionHeader()\n header.ParseFromString(txn.header)\n if header.family_name == family:\n count += 1\n if count > limit:\n LOGGER.debug('Too many transactions of type %s', family)\n return False\n return True\n", "role": "user" }, { "content": "do_nofx", "role": "assistant" } ]
def x(transactions, arguments): try: family, num = arguments.split(',') position = int(num.strip()) except ValueError: LOGGER.warning( 'Ignore, XatY requires arguments in the format family,position not %s' , arguments) return True family = family.strip() if abs(position) >= len(transactions): LOGGER.debug( 'Block does not have enough transactions to validate this rule XatY:%s' , arguments) return False txn = transactions[position] header = TransactionHeader() header.ParseFromString(txn.header) if header.family_name != family: LOGGER.debug('Transaction at postion %s is not of type %s', position, family) return False return True
do_xaty
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(transactions, arguments):\n try:\n family, num = arguments.split(',')\n position = int(num.strip())\n except ValueError:\n LOGGER.warning(\n 'Ignore, XatY requires arguments in the format family,position not %s'\n , arguments)\n return True\n family = family.strip()\n if abs(position) >= len(transactions):\n LOGGER.debug(\n 'Block does not have enough transactions to validate this rule XatY:%s'\n , arguments)\n return False\n txn = transactions[position]\n header = TransactionHeader()\n header.ParseFromString(txn.header)\n if header.family_name != family:\n LOGGER.debug('Transaction at postion %s is not of type %s',\n position, family)\n return False\n return True\n", "role": "user" }, { "content": "do_xaty", "role": "assistant" } ]
def x(transactions, expected_signer, arguments): indices = arguments.split(',') txns_len = len(transactions) for index in indices: try: index = int(index.strip()) except ValueError: LOGGER.warning( 'Ignore, local requries one or more comma seperated integers that represent indices, not %s' , arguments) return True if abs(index) >= txns_len: LOGGER.debug( 'Ignore, Block does not have enough transactions to validate this rule local:%s' , index) continue txn = transactions[index] header = TransactionHeader() header.ParseFromString(txn.header) if header.signer_public_key != expected_signer: LOGGER.debug( 'Transaction at postion %s was not signed by the same key as the block.' , index) return False return True
do_local
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(transactions, expected_signer, arguments):\n indices = arguments.split(',')\n txns_len = len(transactions)\n for index in indices:\n try:\n index = int(index.strip())\n except ValueError:\n LOGGER.warning(\n 'Ignore, local requries one or more comma seperated integers that represent indices, not %s'\n , arguments)\n return True\n if abs(index) >= txns_len:\n LOGGER.debug(\n 'Ignore, Block does not have enough transactions to validate this rule local:%s'\n , index)\n continue\n txn = transactions[index]\n header = TransactionHeader()\n header.ParseFromString(txn.header)\n if header.signer_public_key != expected_signer:\n LOGGER.debug(\n 'Transaction at postion %s was not signed by the same key as the block.'\n , index)\n return False\n return True\n", "role": "user" }, { "content": "do_local", "role": "assistant" } ]
def x(filename): if not os.path.exists(filename): LOGGER.info( 'Skipping validator config loading from non-existent config file: %s' , filename) return ValidatorConfig() LOGGER.info('Loading validator information from config: %s', filename) try: with open(filename) as fd: raw_config = fd.read() except IOError as e: raise LocalConfigurationError( 'Unable to load validator configuration file: {}'.format(str(e))) toml_config = toml.loads(raw_config) invalid_keys = set(toml_config.keys()).difference(['bind', 'endpoint', 'peering', 'seeds', 'peers', 'network_public_key', 'network_private_key', 'scheduler', 'permissions', 'roles', 'opentsdb_url', 'opentsdb_db', 'opentsdb_username', 'opentsdb_password', 'minimum_peer_connectivity', 'maximum_peer_connectivity', 'state_pruning_block_depth', 'fork_cache_keep_time', 'component_thread_pool_workers', 'network_thread_pool_workers', 'signature_thread_pool_workers']) if invalid_keys: raise LocalConfigurationError('Invalid keys in validator config: {}' .format(', '.join(sorted(list(invalid_keys))))) bind_network = None bind_component = None bind_consensus = None for bind in toml_config.get('bind', []): if 'network' in bind: bind_network = bind[bind.find(':') + 1:] if 'component' in bind: bind_component = bind[bind.find(':') + 1:] if 'consensus' in bind: bind_consensus = bind[bind.find(':') + 1:] network_public_key = None network_private_key = None if toml_config.get('network_public_key') is not None: network_public_key = toml_config.get('network_public_key').encode() if toml_config.get('network_private_key') is not None: network_private_key = toml_config.get('network_private_key').encode() config = ValidatorConfig(bind_network=bind_network, bind_component= bind_component, bind_consensus=bind_consensus, endpoint=toml_config .get('endpoint', None), peering=toml_config.get('peering', None), seeds=toml_config.get('seeds', None), peers=toml_config.get('peers', None), network_public_key=network_public_key, network_private_key= network_private_key, scheduler=toml_config.get('scheduler', None), permissions=parse_permissions(toml_config.get('permissions', None)), roles=toml_config.get('roles', None), opentsdb_url=toml_config.get( 'opentsdb_url', None), opentsdb_db=toml_config.get('opentsdb_db', None), opentsdb_username=toml_config.get('opentsdb_username', None), opentsdb_password=toml_config.get('opentsdb_password', None), minimum_peer_connectivity=toml_config.get( 'minimum_peer_connectivity', None), maximum_peer_connectivity= toml_config.get('maximum_peer_connectivity', None), state_pruning_block_depth=toml_config.get( 'state_pruning_block_depth', None), fork_cache_keep_time= toml_config.get('fork_cache_keep_time', None), component_thread_pool_workers=toml_config.get( 'component_thread_pool_workers', None), network_thread_pool_workers =toml_config.get('network_thread_pool_workers', None), signature_thread_pool_workers=toml_config.get( 'signature_thread_pool_workers', None)) return config
load_toml_validator_config
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(filename):\n if not os.path.exists(filename):\n LOGGER.info(\n 'Skipping validator config loading from non-existent config file: %s'\n , filename)\n return ValidatorConfig()\n LOGGER.info('Loading validator information from config: %s', filename)\n try:\n with open(filename) as fd:\n raw_config = fd.read()\n except IOError as e:\n raise LocalConfigurationError(\n 'Unable to load validator configuration file: {}'.format(str(e)))\n toml_config = toml.loads(raw_config)\n invalid_keys = set(toml_config.keys()).difference(['bind', 'endpoint',\n 'peering', 'seeds', 'peers', 'network_public_key',\n 'network_private_key', 'scheduler', 'permissions', 'roles',\n 'opentsdb_url', 'opentsdb_db', 'opentsdb_username',\n 'opentsdb_password', 'minimum_peer_connectivity',\n 'maximum_peer_connectivity', 'state_pruning_block_depth',\n 'fork_cache_keep_time', 'component_thread_pool_workers',\n 'network_thread_pool_workers', 'signature_thread_pool_workers'])\n if invalid_keys:\n raise LocalConfigurationError('Invalid keys in validator config: {}'\n .format(', '.join(sorted(list(invalid_keys)))))\n bind_network = None\n bind_component = None\n bind_consensus = None\n for bind in toml_config.get('bind', []):\n if 'network' in bind:\n bind_network = bind[bind.find(':') + 1:]\n if 'component' in bind:\n bind_component = bind[bind.find(':') + 1:]\n if 'consensus' in bind:\n bind_consensus = bind[bind.find(':') + 1:]\n network_public_key = None\n network_private_key = None\n if toml_config.get('network_public_key') is not None:\n network_public_key = toml_config.get('network_public_key').encode()\n if toml_config.get('network_private_key') is not None:\n network_private_key = toml_config.get('network_private_key').encode()\n config = ValidatorConfig(bind_network=bind_network, bind_component=\n bind_component, bind_consensus=bind_consensus, endpoint=toml_config\n .get('endpoint', None), peering=toml_config.get('peering', None),\n seeds=toml_config.get('seeds', None), peers=toml_config.get('peers',\n None), network_public_key=network_public_key, network_private_key=\n network_private_key, scheduler=toml_config.get('scheduler', None),\n permissions=parse_permissions(toml_config.get('permissions', None)),\n roles=toml_config.get('roles', None), opentsdb_url=toml_config.get(\n 'opentsdb_url', None), opentsdb_db=toml_config.get('opentsdb_db',\n None), opentsdb_username=toml_config.get('opentsdb_username', None),\n opentsdb_password=toml_config.get('opentsdb_password', None),\n minimum_peer_connectivity=toml_config.get(\n 'minimum_peer_connectivity', None), maximum_peer_connectivity=\n toml_config.get('maximum_peer_connectivity', None),\n state_pruning_block_depth=toml_config.get(\n 'state_pruning_block_depth', None), fork_cache_keep_time=\n toml_config.get('fork_cache_keep_time', None),\n component_thread_pool_workers=toml_config.get(\n 'component_thread_pool_workers', None), network_thread_pool_workers\n =toml_config.get('network_thread_pool_workers', None),\n signature_thread_pool_workers=toml_config.get(\n 'signature_thread_pool_workers', None))\n return config\n", "role": "user" }, { "content": "load_toml_validator_config", "role": "assistant" } ]
def x(configs): bind_network = None bind_component = None bind_consensus = None endpoint = None peering = None seeds = None peers = None network_public_key = None network_private_key = None scheduler = None permissions = None roles = None opentsdb_url = None opentsdb_db = None opentsdb_username = None opentsdb_password = None minimum_peer_connectivity = None maximum_peer_connectivity = None state_pruning_block_depth = None fork_cache_keep_time = None component_thread_pool_workers = None network_thread_pool_workers = None signature_thread_pool_workers = None for config in reversed(configs): if config.bind_network is not None: bind_network = config.bind_network if config.bind_component is not None: bind_component = config.bind_component if config.bind_consensus is not None: bind_consensus = config.bind_consensus if config.endpoint is not None: endpoint = config.endpoint if config.peering is not None: peering = config.peering if config.seeds is not None: seeds = config.seeds if config.peers is not None: peers = config.peers if config.network_public_key is not None: network_public_key = config.network_public_key if config.network_private_key is not None: network_private_key = config.network_private_key if config.scheduler is not None: scheduler = config.scheduler if config.permissions is not None or config.permissions == {}: permissions = config.permissions if config.roles is not None: roles = config.roles if config.opentsdb_url is not None: opentsdb_url = config.opentsdb_url if config.opentsdb_db is not None: opentsdb_db = config.opentsdb_db if config.opentsdb_username is not None: opentsdb_username = config.opentsdb_username if config.opentsdb_password is not None: opentsdb_password = config.opentsdb_password if config.minimum_peer_connectivity is not None: minimum_peer_connectivity = config.minimum_peer_connectivity if config.maximum_peer_connectivity is not None: maximum_peer_connectivity = config.maximum_peer_connectivity if config.state_pruning_block_depth is not None: state_pruning_block_depth = config.state_pruning_block_depth if config.fork_cache_keep_time is not None: fork_cache_keep_time = config.fork_cache_keep_time if config.component_thread_pool_workers is not None: component_thread_pool_workers = (config. component_thread_pool_workers) if config.network_thread_pool_workers is not None: network_thread_pool_workers = config.network_thread_pool_workers if config.signature_thread_pool_workers is not None: signature_thread_pool_workers = (config. signature_thread_pool_workers) return ValidatorConfig(bind_network=bind_network, bind_component= bind_component, bind_consensus=bind_consensus, endpoint=endpoint, peering=peering, seeds=seeds, peers=peers, network_public_key= network_public_key, network_private_key=network_private_key, scheduler=scheduler, permissions=permissions, roles=roles, opentsdb_url=opentsdb_url, opentsdb_db=opentsdb_db, opentsdb_username=opentsdb_username, opentsdb_password= opentsdb_password, minimum_peer_connectivity= minimum_peer_connectivity, maximum_peer_connectivity= maximum_peer_connectivity, state_pruning_block_depth= state_pruning_block_depth, fork_cache_keep_time= fork_cache_keep_time, component_thread_pool_workers= component_thread_pool_workers, network_thread_pool_workers= network_thread_pool_workers, signature_thread_pool_workers= signature_thread_pool_workers)
merge_validator_config
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(configs):\n bind_network = None\n bind_component = None\n bind_consensus = None\n endpoint = None\n peering = None\n seeds = None\n peers = None\n network_public_key = None\n network_private_key = None\n scheduler = None\n permissions = None\n roles = None\n opentsdb_url = None\n opentsdb_db = None\n opentsdb_username = None\n opentsdb_password = None\n minimum_peer_connectivity = None\n maximum_peer_connectivity = None\n state_pruning_block_depth = None\n fork_cache_keep_time = None\n component_thread_pool_workers = None\n network_thread_pool_workers = None\n signature_thread_pool_workers = None\n for config in reversed(configs):\n if config.bind_network is not None:\n bind_network = config.bind_network\n if config.bind_component is not None:\n bind_component = config.bind_component\n if config.bind_consensus is not None:\n bind_consensus = config.bind_consensus\n if config.endpoint is not None:\n endpoint = config.endpoint\n if config.peering is not None:\n peering = config.peering\n if config.seeds is not None:\n seeds = config.seeds\n if config.peers is not None:\n peers = config.peers\n if config.network_public_key is not None:\n network_public_key = config.network_public_key\n if config.network_private_key is not None:\n network_private_key = config.network_private_key\n if config.scheduler is not None:\n scheduler = config.scheduler\n if config.permissions is not None or config.permissions == {}:\n permissions = config.permissions\n if config.roles is not None:\n roles = config.roles\n if config.opentsdb_url is not None:\n opentsdb_url = config.opentsdb_url\n if config.opentsdb_db is not None:\n opentsdb_db = config.opentsdb_db\n if config.opentsdb_username is not None:\n opentsdb_username = config.opentsdb_username\n if config.opentsdb_password is not None:\n opentsdb_password = config.opentsdb_password\n if config.minimum_peer_connectivity is not None:\n minimum_peer_connectivity = config.minimum_peer_connectivity\n if config.maximum_peer_connectivity is not None:\n maximum_peer_connectivity = config.maximum_peer_connectivity\n if config.state_pruning_block_depth is not None:\n state_pruning_block_depth = config.state_pruning_block_depth\n if config.fork_cache_keep_time is not None:\n fork_cache_keep_time = config.fork_cache_keep_time\n if config.component_thread_pool_workers is not None:\n component_thread_pool_workers = (config.\n component_thread_pool_workers)\n if config.network_thread_pool_workers is not None:\n network_thread_pool_workers = config.network_thread_pool_workers\n if config.signature_thread_pool_workers is not None:\n signature_thread_pool_workers = (config.\n signature_thread_pool_workers)\n return ValidatorConfig(bind_network=bind_network, bind_component=\n bind_component, bind_consensus=bind_consensus, endpoint=endpoint,\n peering=peering, seeds=seeds, peers=peers, network_public_key=\n network_public_key, network_private_key=network_private_key,\n scheduler=scheduler, permissions=permissions, roles=roles,\n opentsdb_url=opentsdb_url, opentsdb_db=opentsdb_db,\n opentsdb_username=opentsdb_username, opentsdb_password=\n opentsdb_password, minimum_peer_connectivity=\n minimum_peer_connectivity, maximum_peer_connectivity=\n maximum_peer_connectivity, state_pruning_block_depth=\n state_pruning_block_depth, fork_cache_keep_time=\n fork_cache_keep_time, component_thread_pool_workers=\n component_thread_pool_workers, network_thread_pool_workers=\n network_thread_pool_workers, signature_thread_pool_workers=\n signature_thread_pool_workers)\n", "role": "user" }, { "content": "merge_validator_config", "role": "assistant" } ]
def x(key_dir, key_name): key_path = os.path.join(key_dir, '{}.priv'.format(key_name)) if not os.path.exists(key_path): raise LocalConfigurationError('No such signing key file: {}'.format (key_path)) if not os.access(key_path, os.R_OK): raise LocalConfigurationError('Key file is not readable: {}'.format (key_path)) LOGGER.info('Loading signing key: %s', key_path) try: with open(key_path, 'r') as key_file: private_key_str = key_file.read().strip() except IOError as e: raise LocalConfigurationError('Could not load key file: {}'.format( str(e))) try: private_key = Secp256k1PrivateKey.from_hex(private_key_str) except signing.ParseError as e: raise LocalConfigurationError('Invalid key in file {}: {}'.format( key_path, str(e))) context = signing.create_context('secp256k1') crypto_factory = CryptoFactory(context) return crypto_factory.new_signer(private_key)
load_identity_signer
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(key_dir, key_name):\n key_path = os.path.join(key_dir, '{}.priv'.format(key_name))\n if not os.path.exists(key_path):\n raise LocalConfigurationError('No such signing key file: {}'.format\n (key_path))\n if not os.access(key_path, os.R_OK):\n raise LocalConfigurationError('Key file is not readable: {}'.format\n (key_path))\n LOGGER.info('Loading signing key: %s', key_path)\n try:\n with open(key_path, 'r') as key_file:\n private_key_str = key_file.read().strip()\n except IOError as e:\n raise LocalConfigurationError('Could not load key file: {}'.format(\n str(e)))\n try:\n private_key = Secp256k1PrivateKey.from_hex(private_key_str)\n except signing.ParseError as e:\n raise LocalConfigurationError('Invalid key in file {}: {}'.format(\n key_path, str(e)))\n context = signing.create_context('secp256k1')\n crypto_factory = CryptoFactory(context)\n return crypto_factory.new_signer(private_key)\n", "role": "user" }, { "content": "load_identity_signer", "role": "assistant" } ]
def x(subparsers, parent_parser): parser = subparsers.add_parser('status', help= 'Displays information about validator status', description= "Provides a subcommand to show a validator's status") grand_parsers = parser.add_subparsers(title='subcommands', dest= 'subcommand') grand_parsers.required = True add_status_show_parser(grand_parsers, parent_parser)
add_status_parser
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(subparsers, parent_parser):\n parser = subparsers.add_parser('status', help=\n 'Displays information about validator status', description=\n \"Provides a subcommand to show a validator's status\")\n grand_parsers = parser.add_subparsers(title='subcommands', dest=\n 'subcommand')\n grand_parsers.required = True\n add_status_show_parser(grand_parsers, parent_parser)\n", "role": "user" }, { "content": "add_status_parser", "role": "assistant" } ]
def x(self, processor_type): with self._condition: if processor_type not in self: self.wait_for_registration(processor_type) try: processor = self[processor_type].next_processor() except NoProcessorVacancyError: processor = self.wait_for_vacancy(processor_type) processor.inc_occupancy() return processor
get_next_of_type
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, processor_type):\n with self._condition:\n if processor_type not in self:\n self.wait_for_registration(processor_type)\n try:\n processor = self[processor_type].next_processor()\n except NoProcessorVacancyError:\n processor = self.wait_for_vacancy(processor_type)\n processor.inc_occupancy()\n return processor\n", "role": "user" }, { "content": "get_next_of_type", "role": "assistant" } ]
def x(self, processor_identity): with self._condition: processor_types = self._identities.get(processor_identity) if processor_types is None: LOGGER.warning( 'transaction processor with identity %s tried to unregister but was not registered' , processor_identity) return for processor_type in processor_types: if processor_type not in self._processors: LOGGER.warning( 'processor type %s not a known processor type but is associated with identity %s' , processor_type, processor_identity) continue self._processors[processor_type].remove_processor( processor_identity=processor_identity) if not self._processors[processor_type]: del self._processors[processor_type]
remove
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, processor_identity):\n with self._condition:\n processor_types = self._identities.get(processor_identity)\n if processor_types is None:\n LOGGER.warning(\n 'transaction processor with identity %s tried to unregister but was not registered'\n , processor_identity)\n return\n for processor_type in processor_types:\n if processor_type not in self._processors:\n LOGGER.warning(\n 'processor type %s not a known processor type but is associated with identity %s'\n , processor_type, processor_identity)\n continue\n self._processors[processor_type].remove_processor(\n processor_identity=processor_identity)\n if not self._processors[processor_type]:\n del self._processors[processor_type]\n", "role": "user" }, { "content": "remove", "role": "assistant" } ]
def x(self, processor_type): with self._condition: self._condition.wait_for(lambda : processor_type in self or self. _cancelled_event.is_set()) if self._cancelled_event.is_set(): raise WaitCancelledException()
wait_for_registration
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, processor_type):\n with self._condition:\n self._condition.wait_for(lambda : processor_type in self or self.\n _cancelled_event.is_set())\n if self._cancelled_event.is_set():\n raise WaitCancelledException()\n", "role": "user" }, { "content": "wait_for_registration", "role": "assistant" } ]
def x(self, processor_type): with self._condition: self._condition.wait_for(lambda : self._processor_available( processor_type) or self._cancelled_event.is_set()) if self._cancelled_event.is_set(): raise WaitCancelledException() processor = self[processor_type].next_processor() return processor
wait_for_vacancy
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, processor_type):\n with self._condition:\n self._condition.wait_for(lambda : self._processor_available(\n processor_type) or self._cancelled_event.is_set())\n if self._cancelled_event.is_set():\n raise WaitCancelledException()\n processor = self[processor_type].next_processor()\n return processor\n", "role": "user" }, { "content": "wait_for_vacancy", "role": "assistant" } ]
def x(self, key, default_value=None, value_type=str): try: state_entry = self._state_view.get(SettingsView.setting_address(key)) except KeyError: return default_value if state_entry is not None: setting = Setting() setting.ParseFromString(state_entry) for setting_entry in setting.entries: if setting_entry.key == key: return value_type(setting_entry.value) return default_value
get_setting
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, key, default_value=None, value_type=str):\n try:\n state_entry = self._state_view.get(SettingsView.setting_address(key))\n except KeyError:\n return default_value\n if state_entry is not None:\n setting = Setting()\n setting.ParseFromString(state_entry)\n for setting_entry in setting.entries:\n if setting_entry.key == key:\n return value_type(setting_entry.value)\n return default_value\n", "role": "user" }, { "content": "get_setting", "role": "assistant" } ]
def x(self, key, default_value=None, delimiter=',', value_type=str): value = self.get_setting(key) if value is not None: setting_list = [value_type(v) for v in value.split(delimiter)] else: setting_list = default_value return setting_list
get_setting_list
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, key, default_value=None, delimiter=',', value_type=str):\n value = self.get_setting(key)\n if value is not None:\n setting_list = [value_type(v) for v in value.split(delimiter)]\n else:\n setting_list = default_value\n return setting_list\n", "role": "user" }, { "content": "get_setting_list", "role": "assistant" } ]
def x(key): key_parts = key.split('.', maxsplit=_MAX_KEY_PARTS - 1) addr_parts = [_short_hash(x.encode()) for x in key_parts] addr_parts.extend([_EMPTY_PART] * (_MAX_KEY_PARTS - len(addr_parts))) return CONFIG_STATE_NAMESPACE + ''.join(addr_parts)
setting_address
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(key):\n key_parts = key.split('.', maxsplit=_MAX_KEY_PARTS - 1)\n addr_parts = [_short_hash(x.encode()) for x in key_parts]\n addr_parts.extend([_EMPTY_PART] * (_MAX_KEY_PARTS - len(addr_parts)))\n return CONFIG_STATE_NAMESPACE + ''.join(addr_parts)\n", "role": "user" }, { "content": "setting_address", "role": "assistant" } ]
async def submit_batches(self, request): """Accepts a binary encoded BatchList and submits it to the validator. Request: body: octet-stream BatchList of one or more Batches Response: status: - 202: Batches submitted and pending link: /batches or /batch_statuses link for submitted batches """ timer_ctx = self._post_batches_total_time.time() self._post_batches_count.inc() if request.headers['Content-Type'] != 'application/octet-stream': LOGGER.debug('Submission headers had wrong Content-Type: %s', request.headers['Content-Type']) self._post_batches_error.inc() raise errors.SubmissionWrongContentType() body = await request.read() if not body: LOGGER.debug('Submission contained an empty body') self._post_batches_error.inc() raise errors.NoBatchesSubmitted() try: batch_list = BatchList() batch_list.ParseFromString(body) except DecodeError: LOGGER.debug('Submission body could not be decoded: %s', body) self._post_batches_error.inc() raise errors.BadProtobufSubmitted() error_traps = [error_handlers.BatchInvalidTrap, error_handlers. BatchQueueFullTrap] validator_query = client_batch_submit_pb2.ClientBatchSubmitRequest(batches =batch_list.batches) with self._post_batches_validator_time.time(): await self._query_validator(Message.CLIENT_BATCH_SUBMIT_REQUEST, client_batch_submit_pb2.ClientBatchSubmitResponse, validator_query, error_traps) id_string = ','.join(b.header_signature for b in batch_list.batches) status = 202 link = self._build_url(request, path='/batch_statuses', id=id_string) retval = self._wrap_response(request, metadata={'link': link}, status= status) timer_ctx.stop() return retval
submit_batches
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def submit_batches(self, request):\n \"\"\"Accepts a binary encoded BatchList and submits it to the validator.\n\n Request:\n body: octet-stream BatchList of one or more Batches\n Response:\n status:\n - 202: Batches submitted and pending\n link: /batches or /batch_statuses link for submitted batches\n\n \"\"\"\n timer_ctx = self._post_batches_total_time.time()\n self._post_batches_count.inc()\n if request.headers['Content-Type'] != 'application/octet-stream':\n LOGGER.debug('Submission headers had wrong Content-Type: %s',\n request.headers['Content-Type'])\n self._post_batches_error.inc()\n raise errors.SubmissionWrongContentType()\n body = await request.read()\n if not body:\n LOGGER.debug('Submission contained an empty body')\n self._post_batches_error.inc()\n raise errors.NoBatchesSubmitted()\n try:\n batch_list = BatchList()\n batch_list.ParseFromString(body)\n except DecodeError:\n LOGGER.debug('Submission body could not be decoded: %s', body)\n self._post_batches_error.inc()\n raise errors.BadProtobufSubmitted()\n error_traps = [error_handlers.BatchInvalidTrap, error_handlers.\n BatchQueueFullTrap]\n validator_query = client_batch_submit_pb2.ClientBatchSubmitRequest(batches\n =batch_list.batches)\n with self._post_batches_validator_time.time():\n await self._query_validator(Message.CLIENT_BATCH_SUBMIT_REQUEST,\n client_batch_submit_pb2.ClientBatchSubmitResponse,\n validator_query, error_traps)\n id_string = ','.join(b.header_signature for b in batch_list.batches)\n status = 202\n link = self._build_url(request, path='/batch_statuses', id=id_string)\n retval = self._wrap_response(request, metadata={'link': link}, status=\n status)\n timer_ctx.stop()\n return retval\n", "role": "user" }, { "content": "submit_batches", "role": "assistant" } ]
async def list_statuses(self, request): """Fetches the committed status of batches by either a POST or GET. Request: body: A JSON array of one or more id strings (if POST) query: - id: A comma separated list of up to 15 ids (if GET) - wait: Request should not return until all batches committed Response: data: A JSON object, with batch ids as keys, and statuses as values link: The /batch_statuses link queried (if GET) """ error_traps = [error_handlers.StatusResponseMissing] if request.method == 'POST': if request.headers['Content-Type'] != 'application/json': LOGGER.debug('Request headers had wrong Content-Type: %s', request.headers['Content-Type']) raise errors.StatusWrongContentType() ids = await request.json() if not ids or not isinstance(ids, list) or not all(isinstance(i, str) for i in ids): LOGGER.debug('Request body was invalid: %s', ids) raise errors.StatusBodyInvalid() for i in ids: self._validate_id(i) else: ids = self._get_filter_ids(request) if not ids: LOGGER.debug('Request for statuses missing id query') raise errors.StatusIdQueryInvalid() validator_query = client_batch_submit_pb2.ClientBatchStatusRequest( batch_ids=ids) self._set_wait(request, validator_query) response = await self._query_validator(Message. CLIENT_BATCH_STATUS_REQUEST, client_batch_submit_pb2. ClientBatchStatusResponse, validator_query, error_traps) if request.method != 'POST': metadata = self._get_metadata(request, response) else: metadata = None data = self._drop_id_prefixes(self._drop_empty_props(response[ 'batch_statuses'])) return self._wrap_response(request, data=data, metadata=metadata)
list_statuses
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def list_statuses(self, request):\n \"\"\"Fetches the committed status of batches by either a POST or GET.\n\n Request:\n body: A JSON array of one or more id strings (if POST)\n query:\n - id: A comma separated list of up to 15 ids (if GET)\n - wait: Request should not return until all batches committed\n\n Response:\n data: A JSON object, with batch ids as keys, and statuses as values\n link: The /batch_statuses link queried (if GET)\n \"\"\"\n error_traps = [error_handlers.StatusResponseMissing]\n if request.method == 'POST':\n if request.headers['Content-Type'] != 'application/json':\n LOGGER.debug('Request headers had wrong Content-Type: %s',\n request.headers['Content-Type'])\n raise errors.StatusWrongContentType()\n ids = await request.json()\n if not ids or not isinstance(ids, list) or not all(isinstance(i,\n str) for i in ids):\n LOGGER.debug('Request body was invalid: %s', ids)\n raise errors.StatusBodyInvalid()\n for i in ids:\n self._validate_id(i)\n else:\n ids = self._get_filter_ids(request)\n if not ids:\n LOGGER.debug('Request for statuses missing id query')\n raise errors.StatusIdQueryInvalid()\n validator_query = client_batch_submit_pb2.ClientBatchStatusRequest(\n batch_ids=ids)\n self._set_wait(request, validator_query)\n response = await self._query_validator(Message.\n CLIENT_BATCH_STATUS_REQUEST, client_batch_submit_pb2.\n ClientBatchStatusResponse, validator_query, error_traps)\n if request.method != 'POST':\n metadata = self._get_metadata(request, response)\n else:\n metadata = None\n data = self._drop_id_prefixes(self._drop_empty_props(response[\n 'batch_statuses']))\n return self._wrap_response(request, data=data, metadata=metadata)\n", "role": "user" }, { "content": "list_statuses", "role": "assistant" } ]
async def list_state(self, request): """Fetches list of data entries, optionally filtered by address prefix. Request: query: - head: The id of the block to use as the head of the chain - address: Return entries whose addresses begin with this prefix Response: data: An array of leaf objects with address and data keys head: The head used for this query (most recent if unspecified) link: The link to this exact query, including head block paging: Paging info and nav, like total resources and a next link """ paging_controls = self._get_paging_controls(request) head, root = await self._head_to_root(request.url.query.get('head', None)) validator_query = client_state_pb2.ClientStateListRequest(state_root= root, address=request.url.query.get('address', None), sorting=self. _get_sorting_message(request, 'default'), paging=self. _make_paging_message(paging_controls)) response = await self._query_validator(Message. CLIENT_STATE_LIST_REQUEST, client_state_pb2.ClientStateListResponse, validator_query) return self._wrap_paginated_response(request=request, response=response, controls=paging_controls, data=response.get('entries', []), head=head)
list_state
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def list_state(self, request):\n \"\"\"Fetches list of data entries, optionally filtered by address prefix.\n\n Request:\n query:\n - head: The id of the block to use as the head of the chain\n - address: Return entries whose addresses begin with this\n prefix\n\n Response:\n data: An array of leaf objects with address and data keys\n head: The head used for this query (most recent if unspecified)\n link: The link to this exact query, including head block\n paging: Paging info and nav, like total resources and a next link\n \"\"\"\n paging_controls = self._get_paging_controls(request)\n head, root = await self._head_to_root(request.url.query.get('head', None))\n validator_query = client_state_pb2.ClientStateListRequest(state_root=\n root, address=request.url.query.get('address', None), sorting=self.\n _get_sorting_message(request, 'default'), paging=self.\n _make_paging_message(paging_controls))\n response = await self._query_validator(Message.\n CLIENT_STATE_LIST_REQUEST, client_state_pb2.ClientStateListResponse,\n validator_query)\n return self._wrap_paginated_response(request=request, response=response,\n controls=paging_controls, data=response.get('entries', []), head=head)\n", "role": "user" }, { "content": "list_state", "role": "assistant" } ]
async def fetch_state(self, request): """Fetches data from a specific address in the validator's state tree. Request: query: - head: The id of the block to use as the head of the chain - address: The 70 character address of the data to be fetched Response: data: The base64 encoded binary data stored at that address head: The head used for this query (most recent if unspecified) link: The link to this exact query, including head block """ error_traps = [error_handlers.InvalidAddressTrap, error_handlers. StateNotFoundTrap] address = request.match_info.get('address', '') head = request.url.query.get('head', None) head, root = await self._head_to_root(head) response = await self._query_validator(Message.CLIENT_STATE_GET_REQUEST, client_state_pb2.ClientStateGetResponse, client_state_pb2. ClientStateGetRequest(state_root=root, address=address), error_traps) return self._wrap_response(request, data=response['value'], metadata= self._get_metadata(request, response, head=head))
fetch_state
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def fetch_state(self, request):\n \"\"\"Fetches data from a specific address in the validator's state tree.\n\n Request:\n query:\n - head: The id of the block to use as the head of the chain\n - address: The 70 character address of the data to be fetched\n\n Response:\n data: The base64 encoded binary data stored at that address\n head: The head used for this query (most recent if unspecified)\n link: The link to this exact query, including head block\n \"\"\"\n error_traps = [error_handlers.InvalidAddressTrap, error_handlers.\n StateNotFoundTrap]\n address = request.match_info.get('address', '')\n head = request.url.query.get('head', None)\n head, root = await self._head_to_root(head)\n response = await self._query_validator(Message.CLIENT_STATE_GET_REQUEST,\n client_state_pb2.ClientStateGetResponse, client_state_pb2.\n ClientStateGetRequest(state_root=root, address=address), error_traps)\n return self._wrap_response(request, data=response['value'], metadata=\n self._get_metadata(request, response, head=head))\n", "role": "user" }, { "content": "fetch_state", "role": "assistant" } ]
async def list_blocks(self, request): """Fetches list of blocks from validator, optionally filtered by id. Request: query: - head: The id of the block to use as the head of the chain - id: Comma separated list of block ids to include in results Response: data: JSON array of fully expanded Block objects head: The head used for this query (most recent if unspecified) link: The link to this exact query, including head block paging: Paging info and nav, like total resources and a next link """ paging_controls = self._get_paging_controls(request) validator_query = client_block_pb2.ClientBlockListRequest(head_id=self. _get_head_id(request), block_ids=self._get_filter_ids(request), sorting=self._get_sorting_message(request, 'block_num'), paging= self._make_paging_message(paging_controls)) response = await self._query_validator(Message. CLIENT_BLOCK_LIST_REQUEST, client_block_pb2.ClientBlockListResponse, validator_query) return self._wrap_paginated_response(request=request, response=response, controls=paging_controls, data=[self._expand_block(b) for b in response['blocks']])
list_blocks
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def list_blocks(self, request):\n \"\"\"Fetches list of blocks from validator, optionally filtered by id.\n\n Request:\n query:\n - head: The id of the block to use as the head of the chain\n - id: Comma separated list of block ids to include in results\n\n Response:\n data: JSON array of fully expanded Block objects\n head: The head used for this query (most recent if unspecified)\n link: The link to this exact query, including head block\n paging: Paging info and nav, like total resources and a next link\n \"\"\"\n paging_controls = self._get_paging_controls(request)\n validator_query = client_block_pb2.ClientBlockListRequest(head_id=self.\n _get_head_id(request), block_ids=self._get_filter_ids(request),\n sorting=self._get_sorting_message(request, 'block_num'), paging=\n self._make_paging_message(paging_controls))\n response = await self._query_validator(Message.\n CLIENT_BLOCK_LIST_REQUEST, client_block_pb2.ClientBlockListResponse,\n validator_query)\n return self._wrap_paginated_response(request=request, response=response,\n controls=paging_controls, data=[self._expand_block(b) for b in\n response['blocks']])\n", "role": "user" }, { "content": "list_blocks", "role": "assistant" } ]
async def fetch_block(self, request): """Fetches a specific block from the validator, specified by id. Request: path: - block_id: The 128-character id of the block to be fetched Response: data: A JSON object with the data from the fully expanded Block link: The link to this exact query """ error_traps = [error_handlers.BlockNotFoundTrap] block_id = request.match_info.get('block_id', '') self._validate_id(block_id) response = await self._query_validator(Message. CLIENT_BLOCK_GET_BY_ID_REQUEST, client_block_pb2. ClientBlockGetResponse, client_block_pb2.ClientBlockGetByIdRequest( block_id=block_id), error_traps) return self._wrap_response(request, data=self._expand_block(response[ 'block']), metadata=self._get_metadata(request, response))
fetch_block
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def fetch_block(self, request):\n \"\"\"Fetches a specific block from the validator, specified by id.\n Request:\n path:\n - block_id: The 128-character id of the block to be fetched\n\n Response:\n data: A JSON object with the data from the fully expanded Block\n link: The link to this exact query\n \"\"\"\n error_traps = [error_handlers.BlockNotFoundTrap]\n block_id = request.match_info.get('block_id', '')\n self._validate_id(block_id)\n response = await self._query_validator(Message.\n CLIENT_BLOCK_GET_BY_ID_REQUEST, client_block_pb2.\n ClientBlockGetResponse, client_block_pb2.ClientBlockGetByIdRequest(\n block_id=block_id), error_traps)\n return self._wrap_response(request, data=self._expand_block(response[\n 'block']), metadata=self._get_metadata(request, response))\n", "role": "user" }, { "content": "fetch_block", "role": "assistant" } ]
async def list_batches(self, request): """Fetches list of batches from validator, optionally filtered by id. Request: query: - head: The id of the block to use as the head of the chain - id: Comma separated list of batch ids to include in results Response: data: JSON array of fully expanded Batch objects head: The head used for this query (most recent if unspecified) link: The link to this exact query, including head block paging: Paging info and nav, like total resources and a next link """ paging_controls = self._get_paging_controls(request) validator_query = client_batch_pb2.ClientBatchListRequest(head_id=self. _get_head_id(request), batch_ids=self._get_filter_ids(request), sorting=self._get_sorting_message(request, 'default'), paging=self. _make_paging_message(paging_controls)) response = await self._query_validator(Message. CLIENT_BATCH_LIST_REQUEST, client_batch_pb2.ClientBatchListResponse, validator_query) return self._wrap_paginated_response(request=request, response=response, controls=paging_controls, data=[self._expand_batch(b) for b in response['batches']])
list_batches
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def list_batches(self, request):\n \"\"\"Fetches list of batches from validator, optionally filtered by id.\n\n Request:\n query:\n - head: The id of the block to use as the head of the chain\n - id: Comma separated list of batch ids to include in results\n\n Response:\n data: JSON array of fully expanded Batch objects\n head: The head used for this query (most recent if unspecified)\n link: The link to this exact query, including head block\n paging: Paging info and nav, like total resources and a next link\n \"\"\"\n paging_controls = self._get_paging_controls(request)\n validator_query = client_batch_pb2.ClientBatchListRequest(head_id=self.\n _get_head_id(request), batch_ids=self._get_filter_ids(request),\n sorting=self._get_sorting_message(request, 'default'), paging=self.\n _make_paging_message(paging_controls))\n response = await self._query_validator(Message.\n CLIENT_BATCH_LIST_REQUEST, client_batch_pb2.ClientBatchListResponse,\n validator_query)\n return self._wrap_paginated_response(request=request, response=response,\n controls=paging_controls, data=[self._expand_batch(b) for b in\n response['batches']])\n", "role": "user" }, { "content": "list_batches", "role": "assistant" } ]
async def fetch_batch(self, request): """Fetches a specific batch from the validator, specified by id. Request: path: - batch_id: The 128-character id of the batch to be fetched Response: data: A JSON object with the data from the fully expanded Batch link: The link to this exact query """ error_traps = [error_handlers.BatchNotFoundTrap] batch_id = request.match_info.get('batch_id', '') self._validate_id(batch_id) response = await self._query_validator(Message.CLIENT_BATCH_GET_REQUEST, client_batch_pb2.ClientBatchGetResponse, client_batch_pb2. ClientBatchGetRequest(batch_id=batch_id), error_traps) return self._wrap_response(request, data=self._expand_batch(response[ 'batch']), metadata=self._get_metadata(request, response))
fetch_batch
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def fetch_batch(self, request):\n \"\"\"Fetches a specific batch from the validator, specified by id.\n\n Request:\n path:\n - batch_id: The 128-character id of the batch to be fetched\n\n Response:\n data: A JSON object with the data from the fully expanded Batch\n link: The link to this exact query\n \"\"\"\n error_traps = [error_handlers.BatchNotFoundTrap]\n batch_id = request.match_info.get('batch_id', '')\n self._validate_id(batch_id)\n response = await self._query_validator(Message.CLIENT_BATCH_GET_REQUEST,\n client_batch_pb2.ClientBatchGetResponse, client_batch_pb2.\n ClientBatchGetRequest(batch_id=batch_id), error_traps)\n return self._wrap_response(request, data=self._expand_batch(response[\n 'batch']), metadata=self._get_metadata(request, response))\n", "role": "user" }, { "content": "fetch_batch", "role": "assistant" } ]
async def list_transactions(self, request): """Fetches list of txns from validator, optionally filtered by id. Request: query: - head: The id of the block to use as the head of the chain - id: Comma separated list of txn ids to include in results Response: data: JSON array of Transaction objects with expanded headers head: The head used for this query (most recent if unspecified) link: The link to this exact query, including head block paging: Paging info and nav, like total resources and a next link """ paging_controls = self._get_paging_controls(request) validator_query = client_transaction_pb2.ClientTransactionListRequest( head_id=self._get_head_id(request), transaction_ids=self. _get_filter_ids(request), sorting=self._get_sorting_message(request, 'default'), paging=self._make_paging_message(paging_controls)) response = await self._query_validator(Message. CLIENT_TRANSACTION_LIST_REQUEST, client_transaction_pb2. ClientTransactionListResponse, validator_query) data = [self._expand_transaction(t) for t in response['transactions']] return self._wrap_paginated_response(request=request, response=response, controls=paging_controls, data=data)
list_transactions
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def list_transactions(self, request):\n \"\"\"Fetches list of txns from validator, optionally filtered by id.\n\n Request:\n query:\n - head: The id of the block to use as the head of the chain\n - id: Comma separated list of txn ids to include in results\n\n Response:\n data: JSON array of Transaction objects with expanded headers\n head: The head used for this query (most recent if unspecified)\n link: The link to this exact query, including head block\n paging: Paging info and nav, like total resources and a next link\n \"\"\"\n paging_controls = self._get_paging_controls(request)\n validator_query = client_transaction_pb2.ClientTransactionListRequest(\n head_id=self._get_head_id(request), transaction_ids=self.\n _get_filter_ids(request), sorting=self._get_sorting_message(request,\n 'default'), paging=self._make_paging_message(paging_controls))\n response = await self._query_validator(Message.\n CLIENT_TRANSACTION_LIST_REQUEST, client_transaction_pb2.\n ClientTransactionListResponse, validator_query)\n data = [self._expand_transaction(t) for t in response['transactions']]\n return self._wrap_paginated_response(request=request, response=response,\n controls=paging_controls, data=data)\n", "role": "user" }, { "content": "list_transactions", "role": "assistant" } ]
async def fetch_transaction(self, request): """Fetches a specific transaction from the validator, specified by id. Request: path: - transaction_id: The 128-character id of the txn to be fetched Response: data: A JSON object with the data from the expanded Transaction link: The link to this exact query """ error_traps = [error_handlers.TransactionNotFoundTrap] txn_id = request.match_info.get('transaction_id', '') self._validate_id(txn_id) response = await self._query_validator(Message. CLIENT_TRANSACTION_GET_REQUEST, client_transaction_pb2. ClientTransactionGetResponse, client_transaction_pb2. ClientTransactionGetRequest(transaction_id=txn_id), error_traps) return self._wrap_response(request, data=self._expand_transaction( response['transaction']), metadata=self._get_metadata(request, response))
fetch_transaction
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def fetch_transaction(self, request):\n \"\"\"Fetches a specific transaction from the validator, specified by id.\n\n Request:\n path:\n - transaction_id: The 128-character id of the txn to be fetched\n\n Response:\n data: A JSON object with the data from the expanded Transaction\n link: The link to this exact query\n \"\"\"\n error_traps = [error_handlers.TransactionNotFoundTrap]\n txn_id = request.match_info.get('transaction_id', '')\n self._validate_id(txn_id)\n response = await self._query_validator(Message.\n CLIENT_TRANSACTION_GET_REQUEST, client_transaction_pb2.\n ClientTransactionGetResponse, client_transaction_pb2.\n ClientTransactionGetRequest(transaction_id=txn_id), error_traps)\n return self._wrap_response(request, data=self._expand_transaction(\n response['transaction']), metadata=self._get_metadata(request,\n response))\n", "role": "user" }, { "content": "fetch_transaction", "role": "assistant" } ]
async def list_receipts(self, request): """Fetches the receipts for transaction by either a POST or GET. Request: body: A JSON array of one or more transaction id strings (if POST) query: - id: A comma separated list of up to 15 ids (if GET) - wait: Request should return as soon as some receipts are available Response: data: A JSON object, with transaction ids as keys, and receipts as values link: The /receipts link queried (if GET) """ error_traps = [error_handlers.ReceiptNotFoundTrap] if request.method == 'POST': if request.headers['Content-Type'] != 'application/json': LOGGER.debug('Request headers had wrong Content-Type: %s', request.headers['Content-Type']) raise errors.ReceiptWrongContentType() ids = await request.json() if not ids or not isinstance(ids, list) or not all(isinstance(i, str) for i in ids): LOGGER.debug('Request body was invalid: %s', ids) raise errors.ReceiptBodyInvalid() for i in ids: self._validate_id(i) else: ids = self._get_filter_ids(request) if not ids: LOGGER.debug('Request for receipts missing id query') raise errors.ReceiptIdQueryInvalid() validator_query = client_receipt_pb2.ClientReceiptGetRequest( transaction_ids=ids) self._set_wait(request, validator_query) response = await self._query_validator(Message. CLIENT_RECEIPT_GET_REQUEST, client_receipt_pb2. ClientReceiptGetResponse, validator_query, error_traps) if request.method != 'POST': metadata = self._get_metadata(request, response) else: metadata = None data = self._drop_id_prefixes(self._drop_empty_props(response['receipts'])) return self._wrap_response(request, data=data, metadata=metadata)
list_receipts
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def list_receipts(self, request):\n \"\"\"Fetches the receipts for transaction by either a POST or GET.\n\n Request:\n body: A JSON array of one or more transaction id strings (if POST)\n query:\n - id: A comma separated list of up to 15 ids (if GET)\n - wait: Request should return as soon as some receipts are\n available\n\n Response:\n data: A JSON object, with transaction ids as keys, and receipts as\n values\n link: The /receipts link queried (if GET)\n \"\"\"\n error_traps = [error_handlers.ReceiptNotFoundTrap]\n if request.method == 'POST':\n if request.headers['Content-Type'] != 'application/json':\n LOGGER.debug('Request headers had wrong Content-Type: %s',\n request.headers['Content-Type'])\n raise errors.ReceiptWrongContentType()\n ids = await request.json()\n if not ids or not isinstance(ids, list) or not all(isinstance(i,\n str) for i in ids):\n LOGGER.debug('Request body was invalid: %s', ids)\n raise errors.ReceiptBodyInvalid()\n for i in ids:\n self._validate_id(i)\n else:\n ids = self._get_filter_ids(request)\n if not ids:\n LOGGER.debug('Request for receipts missing id query')\n raise errors.ReceiptIdQueryInvalid()\n validator_query = client_receipt_pb2.ClientReceiptGetRequest(\n transaction_ids=ids)\n self._set_wait(request, validator_query)\n response = await self._query_validator(Message.\n CLIENT_RECEIPT_GET_REQUEST, client_receipt_pb2.\n ClientReceiptGetResponse, validator_query, error_traps)\n if request.method != 'POST':\n metadata = self._get_metadata(request, response)\n else:\n metadata = None\n data = self._drop_id_prefixes(self._drop_empty_props(response['receipts']))\n return self._wrap_response(request, data=data, metadata=metadata)\n", "role": "user" }, { "content": "list_receipts", "role": "assistant" } ]
async def fetch_peers(self, request): """Fetches the peers from the validator. Request: Response: data: JSON array of peer endpoints link: The link to this exact query """ response = await self._query_validator(Message.CLIENT_PEERS_GET_REQUEST, client_peers_pb2.ClientPeersGetResponse, client_peers_pb2. ClientPeersGetRequest()) return self._wrap_response(request, data=response['peers'], metadata= self._get_metadata(request, response))
fetch_peers
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def fetch_peers(self, request):\n \"\"\"Fetches the peers from the validator.\n Request:\n\n Response:\n data: JSON array of peer endpoints\n link: The link to this exact query\n \"\"\"\n response = await self._query_validator(Message.CLIENT_PEERS_GET_REQUEST,\n client_peers_pb2.ClientPeersGetResponse, client_peers_pb2.\n ClientPeersGetRequest())\n return self._wrap_response(request, data=response['peers'], metadata=\n self._get_metadata(request, response))\n", "role": "user" }, { "content": "fetch_peers", "role": "assistant" } ]
async def fetch_status(self, request): """Fetches information pertaining to the valiator's status.""" response = await self._query_validator(Message. CLIENT_STATUS_GET_REQUEST, client_status_pb2. ClientStatusGetResponse, client_status_pb2.ClientStatusGetRequest()) return self._wrap_response(request, data={'peers': response['peers'], 'endpoint': response['endpoint']}, metadata=self._get_metadata( request, response))
fetch_status
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def fetch_status(self, request):\n \"\"\"Fetches information pertaining to the valiator's status.\"\"\"\n response = await self._query_validator(Message.\n CLIENT_STATUS_GET_REQUEST, client_status_pb2.\n ClientStatusGetResponse, client_status_pb2.ClientStatusGetRequest())\n return self._wrap_response(request, data={'peers': response['peers'],\n 'endpoint': response['endpoint']}, metadata=self._get_metadata(\n request, response))\n", "role": "user" }, { "content": "fetch_status", "role": "assistant" } ]
async def _query_validator(self, request_type, response_proto, payload, error_traps=None): """Sends a request to the validator and parses the response. """ LOGGER.debug('Sending %s request to validator', self._get_type_name( request_type)) payload_bytes = payload.SerializeToString() response = await self._send_request(request_type, payload_bytes) content = self._parse_response(response_proto, response) LOGGER.debug('Received %s response from validator with status %s', self ._get_type_name(response.message_type), self._get_status_name( response_proto, content.status)) self._check_status_errors(response_proto, content, error_traps) return self._message_to_dict(content)
query_validator
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def _query_validator(self, request_type, response_proto, payload,\n error_traps=None):\n \"\"\"Sends a request to the validator and parses the response.\n \"\"\"\n LOGGER.debug('Sending %s request to validator', self._get_type_name(\n request_type))\n payload_bytes = payload.SerializeToString()\n response = await self._send_request(request_type, payload_bytes)\n content = self._parse_response(response_proto, response)\n LOGGER.debug('Received %s response from validator with status %s', self\n ._get_type_name(response.message_type), self._get_status_name(\n response_proto, content.status))\n self._check_status_errors(response_proto, content, error_traps)\n return self._message_to_dict(content)\n", "role": "user" }, { "content": "query_validator", "role": "assistant" } ]
async def _send_request(self, request_type, payload): """Uses an executor to send an asynchronous ZMQ request to the validator with the handler's Connection """ try: return await self._connection.send(message_type=request_type, message_content=payload, timeout=self._timeout) except DisconnectError: LOGGER.warning('Validator disconnected while waiting for response') raise errors.ValidatorDisconnected() except asyncio.TimeoutError: LOGGER.warning('Timed out while waiting for validator response') raise errors.ValidatorTimedOut() except SendBackoffTimeoutError: LOGGER.warning('Failed sending message - Backoff timed out') raise errors.SendBackoffTimeout()
send_request
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def _send_request(self, request_type, payload):\n \"\"\"Uses an executor to send an asynchronous ZMQ request to the\n validator with the handler's Connection\n \"\"\"\n try:\n return await self._connection.send(message_type=request_type,\n message_content=payload, timeout=self._timeout)\n except DisconnectError:\n LOGGER.warning('Validator disconnected while waiting for response')\n raise errors.ValidatorDisconnected()\n except asyncio.TimeoutError:\n LOGGER.warning('Timed out while waiting for validator response')\n raise errors.ValidatorTimedOut()\n except SendBackoffTimeoutError:\n LOGGER.warning('Failed sending message - Backoff timed out')\n raise errors.SendBackoffTimeout()\n", "role": "user" }, { "content": "send_request", "role": "assistant" } ]
def x(proto, response): try: content = proto() content.ParseFromString(response.content) return content except (DecodeError, AttributeError): LOGGER.error('Validator response was not parsable: %s', response) raise errors.ValidatorResponseInvalid()
parse_response
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(proto, response):\n try:\n content = proto()\n content.ParseFromString(response.content)\n return content\n except (DecodeError, AttributeError):\n LOGGER.error('Validator response was not parsable: %s', response)\n raise errors.ValidatorResponseInvalid()\n", "role": "user" }, { "content": "parse_response", "role": "assistant" } ]
def x(proto, content, error_traps=None): if content.status == proto.OK: return try: if content.status == proto.INTERNAL_ERROR: raise errors.UnknownValidatorError() except AttributeError: pass try: if content.status == proto.NOT_READY: raise errors.ValidatorNotReady() except AttributeError: pass try: if content.status == proto.NO_ROOT: raise errors.HeadNotFound() except AttributeError: pass try: if content.status == proto.INVALID_PAGING: raise errors.PagingInvalid() except AttributeError: pass try: if content.status == proto.INVALID_SORT: raise errors.SortInvalid() except AttributeError: pass if error_traps is not None: for trap in error_traps: trap.check(content.status)
check_status_errors
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(proto, content, error_traps=None):\n if content.status == proto.OK:\n return\n try:\n if content.status == proto.INTERNAL_ERROR:\n raise errors.UnknownValidatorError()\n except AttributeError:\n pass\n try:\n if content.status == proto.NOT_READY:\n raise errors.ValidatorNotReady()\n except AttributeError:\n pass\n try:\n if content.status == proto.NO_ROOT:\n raise errors.HeadNotFound()\n except AttributeError:\n pass\n try:\n if content.status == proto.INVALID_PAGING:\n raise errors.PagingInvalid()\n except AttributeError:\n pass\n try:\n if content.status == proto.INVALID_SORT:\n raise errors.SortInvalid()\n except AttributeError:\n pass\n if error_traps is not None:\n for trap in error_traps:\n trap.check(content.status)\n", "role": "user" }, { "content": "check_status_errors", "role": "assistant" } ]
def x(request, data=None, metadata=None, status=200): envelope = metadata or {} if data is not None: envelope['data'] = data return web.Response(status=status, content_type='application/json', text=json.dumps(envelope, indent=2, separators=(',', ': '), sort_keys=True))
wrap_response
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(request, data=None, metadata=None, status=200):\n envelope = metadata or {}\n if data is not None:\n envelope['data'] = data\n return web.Response(status=status, content_type='application/json',\n text=json.dumps(envelope, indent=2, separators=(',', ': '),\n sort_keys=True))\n", "role": "user" }, { "content": "wrap_response", "role": "assistant" } ]
def x(cls, request, response, controls, data, head=None): paging_response = response['paging'] if head is None: head = response['head_id'] link = cls._build_url(request, head=head, start=paging_response['start' ], limit=paging_response['limit']) paging = {} limit = controls.get('limit') start = controls.get('start') paging['limit'] = limit paging['start'] = start if paging_response.get('next') == '': return cls._wrap_response(request, data=data, metadata={'head': head, 'link': link, 'paging': paging}) next_id = paging_response['next'] paging['next_position'] = next_id def x(start=None): return cls._build_url(request, head=head, limit=limit, start=start) paging['next'] = build_pg_url(paging_response['next']) return cls._wrap_response(request, data=data, metadata={'head': head, 'link': link, 'paging': paging})
wrap_paginated_response
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, request, response, controls, data, head=None):\n paging_response = response['paging']\n if head is None:\n head = response['head_id']\n link = cls._build_url(request, head=head, start=paging_response['start'\n ], limit=paging_response['limit'])\n paging = {}\n limit = controls.get('limit')\n start = controls.get('start')\n paging['limit'] = limit\n paging['start'] = start\n if paging_response.get('next') == '':\n return cls._wrap_response(request, data=data, metadata={'head':\n head, 'link': link, 'paging': paging})\n next_id = paging_response['next']\n paging['next_position'] = next_id\n\n def x(start=None):\n return cls._build_url(request, head=head, limit=limit, start=start)\n paging['next'] = build_pg_url(paging_response['next'])\n return cls._wrap_response(request, data=data, metadata={'head': head,\n 'link': link, 'paging': paging})\n", "role": "user" }, { "content": "wrap_paginated_response", "role": "assistant" } ]
def x(cls, request, response, head=None): head = response.get('head_id', head) metadata = {'link': cls._build_url(request, head=head)} if head is not None: metadata['head'] = head return metadata
get_metadata
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, request, response, head=None):\n head = response.get('head_id', head)\n metadata = {'link': cls._build_url(request, head=head)}\n if head is not None:\n metadata['head'] = head\n return metadata\n", "role": "user" }, { "content": "get_metadata", "role": "assistant" } ]
def x(cls, request, path=None, **changes): changes = {k: v for k, v in changes.items() if v is not None} queries = {**request.url.query, **changes} queries = {k: v for k, v in queries.items() if v is not False} query_strings = [] def x(key): query_strings.append('{}={}'.format(key, queries[key]) if queries[ key] != '' else key) def x(key): queries.pop(key, None) if 'head' in queries: add_query('head') del_query('head') if 'start' in changes: add_query('start') elif 'start' in queries: add_query('start') del_query('start') if 'limit' in queries: add_query('limit') del_query('limit') for key in sorted(queries): add_query(key) scheme = cls._get_forwarded(request, 'proto') or request.url.scheme host = cls._get_forwarded(request, 'host') or request.host forwarded_path = cls._get_forwarded(request, 'path') path = path if path is not None else request.path query = '?' + '&'.join(query_strings) if query_strings else '' url = '{}://{}{}{}{}'.format(scheme, host, forwarded_path, path, query) return url
build_url
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, request, path=None, **changes):\n changes = {k: v for k, v in changes.items() if v is not None}\n queries = {**request.url.query, **changes}\n queries = {k: v for k, v in queries.items() if v is not False}\n query_strings = []\n\n def x(key):\n query_strings.append('{}={}'.format(key, queries[key]) if queries[\n key] != '' else key)\n\n def x(key):\n queries.pop(key, None)\n if 'head' in queries:\n add_query('head')\n del_query('head')\n if 'start' in changes:\n add_query('start')\n elif 'start' in queries:\n add_query('start')\n del_query('start')\n if 'limit' in queries:\n add_query('limit')\n del_query('limit')\n for key in sorted(queries):\n add_query(key)\n scheme = cls._get_forwarded(request, 'proto') or request.url.scheme\n host = cls._get_forwarded(request, 'host') or request.host\n forwarded_path = cls._get_forwarded(request, 'path')\n path = path if path is not None else request.path\n query = '?' + '&'.join(query_strings) if query_strings else ''\n url = '{}://{}{}{}{}'.format(scheme, host, forwarded_path, path, query)\n return url\n", "role": "user" }, { "content": "build_url", "role": "assistant" } ]
def x(request, key): forwarded = request.headers.get('Forwarded', '') match = re.search('(?<={}=).+?(?=[\\s,;]|$)'.format(key), forwarded, re .IGNORECASE) if match is not None: header = match.group(0) if header[0] == '"' and header[-1] == '"': return header[1:-1] return header return request.headers.get('X-Forwarded-{}'.format(key.title()), '')
get_forwarded
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(request, key):\n forwarded = request.headers.get('Forwarded', '')\n match = re.search('(?<={}=).+?(?=[\\\\s,;]|$)'.format(key), forwarded, re\n .IGNORECASE)\n if match is not None:\n header = match.group(0)\n if header[0] == '\"' and header[-1] == '\"':\n return header[1:-1]\n return header\n return request.headers.get('X-Forwarded-{}'.format(key.title()), '')\n", "role": "user" }, { "content": "get_forwarded", "role": "assistant" } ]
def x(cls, block): cls._parse_header(BlockHeader, block) if 'batches' in block: block['batches'] = [cls._expand_batch(b) for b in block['batches']] return block
expand_block
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, block):\n cls._parse_header(BlockHeader, block)\n if 'batches' in block:\n block['batches'] = [cls._expand_batch(b) for b in block['batches']]\n return block\n", "role": "user" }, { "content": "expand_block", "role": "assistant" } ]
def x(cls, batch): cls._parse_header(BatchHeader, batch) if 'transactions' in batch: batch['transactions'] = [cls._expand_transaction(t) for t in batch[ 'transactions']] return batch
expand_batch
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, batch):\n cls._parse_header(BatchHeader, batch)\n if 'transactions' in batch:\n batch['transactions'] = [cls._expand_transaction(t) for t in batch[\n 'transactions']]\n return batch\n", "role": "user" }, { "content": "expand_batch", "role": "assistant" } ]
def x(cls, header_proto, resource): header = header_proto() try: header_bytes = base64.b64decode(resource['header']) header.ParseFromString(header_bytes) except (KeyError, TypeError, ValueError, DecodeError): header = resource.get('header', None) LOGGER.error('The validator sent a resource with %s %s', 'a missing header' if header is None else 'an invalid header:', header or '') raise errors.ResourceHeaderInvalid() resource['header'] = cls._message_to_dict(header) return resource
parse_header
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, header_proto, resource):\n header = header_proto()\n try:\n header_bytes = base64.b64decode(resource['header'])\n header.ParseFromString(header_bytes)\n except (KeyError, TypeError, ValueError, DecodeError):\n header = resource.get('header', None)\n LOGGER.error('The validator sent a resource with %s %s', \n 'a missing header' if header is None else 'an invalid header:',\n header or '')\n raise errors.ResourceHeaderInvalid()\n resource['header'] = cls._message_to_dict(header)\n return resource\n", "role": "user" }, { "content": "parse_header", "role": "assistant" } ]
def x(request): start = request.url.query.get('start', None) limit = request.url.query.get('limit', None) controls = {} if limit is not None: try: controls['limit'] = int(limit) except ValueError: LOGGER.debug('Request query had an invalid limit: %s', limit) raise errors.CountInvalid() if controls['limit'] <= 0: LOGGER.debug('Request query had an invalid limit: %s', limit) raise errors.CountInvalid() if start is not None: controls['start'] = start return controls
get_paging_controls
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(request):\n start = request.url.query.get('start', None)\n limit = request.url.query.get('limit', None)\n controls = {}\n if limit is not None:\n try:\n controls['limit'] = int(limit)\n except ValueError:\n LOGGER.debug('Request query had an invalid limit: %s', limit)\n raise errors.CountInvalid()\n if controls['limit'] <= 0:\n LOGGER.debug('Request query had an invalid limit: %s', limit)\n raise errors.CountInvalid()\n if start is not None:\n controls['start'] = start\n return controls\n", "role": "user" }, { "content": "get_paging_controls", "role": "assistant" } ]
def x(controls): return client_list_control_pb2.ClientPagingControls(start=controls.get( 'start', None), limit=controls.get('limit', None))
make_paging_message
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(controls):\n return client_list_control_pb2.ClientPagingControls(start=controls.get(\n 'start', None), limit=controls.get('limit', None))\n", "role": "user" }, { "content": "make_paging_message", "role": "assistant" } ]
def x(request, key): control_list = [] reverse = request.url.query.get('reverse', None) if reverse is None: return control_list if reverse.lower() == '': control_list.append(client_list_control_pb2.ClientSortControls( reverse=True, keys=key.split(','))) elif reverse.lower() != 'false': control_list.append(client_list_control_pb2.ClientSortControls( reverse=True, keys=reverse.split(','))) return control_list
get_sorting_message
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(request, key):\n control_list = []\n reverse = request.url.query.get('reverse', None)\n if reverse is None:\n return control_list\n if reverse.lower() == '':\n control_list.append(client_list_control_pb2.ClientSortControls(\n reverse=True, keys=key.split(',')))\n elif reverse.lower() != 'false':\n control_list.append(client_list_control_pb2.ClientSortControls(\n reverse=True, keys=reverse.split(',')))\n return control_list\n", "role": "user" }, { "content": "get_sorting_message", "role": "assistant" } ]
def x(self, request, validator_query): wait = request.url.query.get('wait', 'false') if wait.lower() != 'false': validator_query.wait = True try: validator_query.timeout = int(wait) except ValueError: validator_query.timeout = int(self._timeout * 0.95)
set_wait
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, request, validator_query):\n wait = request.url.query.get('wait', 'false')\n if wait.lower() != 'false':\n validator_query.wait = True\n try:\n validator_query.timeout = int(wait)\n except ValueError:\n validator_query.timeout = int(self._timeout * 0.95)\n", "role": "user" }, { "content": "set_wait", "role": "assistant" } ]
def x(self, item): if isinstance(item, list): return [self._drop_empty_props(i) for i in item] if isinstance(item, dict): return {k: self._drop_empty_props(v) for k, v in item.items() if v != ''} return item
drop_empty_props
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, item):\n if isinstance(item, list):\n return [self._drop_empty_props(i) for i in item]\n if isinstance(item, dict):\n return {k: self._drop_empty_props(v) for k, v in item.items() if v !=\n ''}\n return item\n", "role": "user" }, { "content": "drop_empty_props", "role": "assistant" } ]
def x(self, item): if isinstance(item, list): return [self._drop_id_prefixes(i) for i in item] if isinstance(item, dict): return {('id' if k.endswith('id') else k): self._drop_id_prefixes(v ) for k, v in item.items()} return item
drop_id_prefixes
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, item):\n if isinstance(item, list):\n return [self._drop_id_prefixes(i) for i in item]\n if isinstance(item, dict):\n return {('id' if k.endswith('id') else k): self._drop_id_prefixes(v\n ) for k, v in item.items()}\n return item\n", "role": "user" }, { "content": "drop_id_prefixes", "role": "assistant" } ]
def x(cls, request): head_id = request.url.query.get('head', None) if head_id is not None: cls._validate_id(head_id) return head_id
get_head_id
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, request):\n head_id = request.url.query.get('head', None)\n if head_id is not None:\n cls._validate_id(head_id)\n return head_id\n", "role": "user" }, { "content": "get_head_id", "role": "assistant" } ]
def x(cls, request): id_query = request.url.query.get('id', None) if id_query is None: return None filter_ids = id_query.split(',') for filter_id in filter_ids: cls._validate_id(filter_id) return filter_ids
get_filter_ids
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, request):\n id_query = request.url.query.get('id', None)\n if id_query is None:\n return None\n filter_ids = id_query.split(',')\n for filter_id in filter_ids:\n cls._validate_id(filter_id)\n return filter_ids\n", "role": "user" }, { "content": "get_filter_ids", "role": "assistant" } ]
def x(self, state_root_hash=None): if state_root_hash is None: state_root_hash = INIT_ROOT_KEY merkle_db = MerkleDatabase(self._database, merkle_root=state_root_hash) return StateView(merkle_db)
create_view
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, state_root_hash=None):\n if state_root_hash is None:\n state_root_hash = INIT_ROOT_KEY\n merkle_db = MerkleDatabase(self._database, merkle_root=state_root_hash)\n return StateView(merkle_db)\n", "role": "user" }, { "content": "create_view", "role": "assistant" } ]
def x(self, connection_id, subscriptions, last_known_block_id): with self._subscribers_cv: self._subscribers[connection_id] = EventSubscriber(connection_id, subscriptions, last_known_block_id) LOGGER.debug('Added Subscriber %s for %s', connection_id, subscriptions)
add_subscriber
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, connection_id, subscriptions, last_known_block_id):\n with self._subscribers_cv:\n self._subscribers[connection_id] = EventSubscriber(connection_id,\n subscriptions, last_known_block_id)\n LOGGER.debug('Added Subscriber %s for %s', connection_id, subscriptions)\n", "role": "user" }, { "content": "add_subscriber", "role": "assistant" } ]
def x(self, connection_id): with self._subscribers_cv: subscriber = self._subscribers[connection_id] last_known_block_id = subscriber.get_last_known_block_id() subscriptions = subscriber.subscriptions if last_known_block_id is not None: LOGGER.debug('Catching up Subscriber %s from %s', connection_id, last_known_block_id) for block_id in self.get_catchup_block_ids(last_known_block_id): events = self.get_events_for_block_id(block_id, subscriptions) event_list = EventList(events=events) self._send(connection_id, event_list.SerializeToString())
catchup_subscriber
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, connection_id):\n with self._subscribers_cv:\n subscriber = self._subscribers[connection_id]\n last_known_block_id = subscriber.get_last_known_block_id()\n subscriptions = subscriber.subscriptions\n if last_known_block_id is not None:\n LOGGER.debug('Catching up Subscriber %s from %s', connection_id,\n last_known_block_id)\n for block_id in self.get_catchup_block_ids(last_known_block_id):\n events = self.get_events_for_block_id(block_id, subscriptions)\n event_list = EventList(events=events)\n self._send(connection_id, event_list.SerializeToString())\n", "role": "user" }, { "content": "catchup_subscriber", "role": "assistant" } ]
def x(self, last_known_block_id): catchup_up_blocks = [] chain_head = self._block_store.chain_head if chain_head and last_known_block_id != chain_head.identifier: for block in self._block_store.get_predecessor_iter(): if last_known_block_id != NULL_BLOCK_IDENTIFIER: if block.identifier == last_known_block_id: break catchup_up_blocks.append(block.identifier) return list(reversed(catchup_up_blocks))
get_catchup_block_ids
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, last_known_block_id):\n catchup_up_blocks = []\n chain_head = self._block_store.chain_head\n if chain_head and last_known_block_id != chain_head.identifier:\n for block in self._block_store.get_predecessor_iter():\n if last_known_block_id != NULL_BLOCK_IDENTIFIER:\n if block.identifier == last_known_block_id:\n break\n catchup_up_blocks.append(block.identifier)\n return list(reversed(catchup_up_blocks))\n", "role": "user" }, { "content": "get_catchup_block_ids", "role": "assistant" } ]
def x(self, block_ids, subscriptions): blocks = [self._block_store[block_id] for block_id in block_ids] return self.get_events_for_blocks(blocks, subscriptions)
get_events_for_block_ids
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, block_ids, subscriptions):\n blocks = [self._block_store[block_id] for block_id in block_ids]\n return self.get_events_for_blocks(blocks, subscriptions)\n", "role": "user" }, { "content": "get_events_for_block_ids", "role": "assistant" } ]
def x(self, blocks, subscriptions): events = [] for blkw in blocks: events.extend(self.get_events_for_block(blkw, subscriptions)) return events
get_events_for_blocks
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, blocks, subscriptions):\n events = []\n for blkw in blocks:\n events.extend(self.get_events_for_block(blkw, subscriptions))\n return events\n", "role": "user" }, { "content": "get_events_for_blocks", "role": "assistant" } ]
def x(module_name): module_package = module_name if module_name == 'genesis': module_package = ( 'sawtooth_validator.journal.consensus.genesis.genesis_consensus') elif module_name == 'devmode': module_package = ( 'sawtooth_validator.journal.consensus.dev_mode.dev_mode_consensus') try: return importlib.import_module(module_package) except ImportError: raise UnknownConsensusModuleError( 'Consensus module "{}" does not exist.'.format(module_name))
get_consensus_module
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(module_name):\n module_package = module_name\n if module_name == 'genesis':\n module_package = (\n 'sawtooth_validator.journal.consensus.genesis.genesis_consensus')\n elif module_name == 'devmode':\n module_package = (\n 'sawtooth_validator.journal.consensus.dev_mode.dev_mode_consensus')\n try:\n return importlib.import_module(module_package)\n except ImportError:\n raise UnknownConsensusModuleError(\n 'Consensus module \"{}\" does not exist.'.format(module_name))\n", "role": "user" }, { "content": "get_consensus_module", "role": "assistant" } ]
def x(block_id, state_view): settings_view = SettingsView(state_view) default_consensus = ('genesis' if block_id == NULL_BLOCK_IDENTIFIER else 'devmode') consensus_module_name = settings_view.get_setting( 'sawtooth.consensus.algorithm', default_value=default_consensus) return ConsensusFactory.get_consensus_module(consensus_module_name)
get_configured_consensus_module
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(block_id, state_view):\n settings_view = SettingsView(state_view)\n default_consensus = ('genesis' if block_id == NULL_BLOCK_IDENTIFIER else\n 'devmode')\n consensus_module_name = settings_view.get_setting(\n 'sawtooth.consensus.algorithm', default_value=default_consensus)\n return ConsensusFactory.get_consensus_module(consensus_module_name)\n", "role": "user" }, { "content": "get_configured_consensus_module", "role": "assistant" } ]
def x(args): parser = argparse.ArgumentParser(description= 'Starts the REST API application and connects to a specified validator.' ) parser.add_argument('-B', '--bind', help= 'identify host and port for API to run on default: http://localhost:8008)' , action='append') parser.add_argument('-C', '--connect', help= 'specify URL to connect to a running validator') parser.add_argument('-t', '--timeout', help= 'set time (in seconds) to wait for validator response' ) parser.add_argument('--client-max-size', type=int, help= 'the max size (in bytes) of a request body') parser.add_argument('-v', '--verbose', action='count', default=0, help= 'enable more verbose output to stderr') parser.add_argument('--opentsdb-url', help= 'specify host and port for Open TSDB database used for metrics' ) parser.add_argument('--opentsdb-db', help= 'specify name of database for storing metrics') try: version = pkg_resources.get_distribution(DISTRIBUTION_NAME).version except pkg_resources.DistributionNotFound: version = 'UNKNOWN' parser.add_argument('-V', '--version', action='version', version=( DISTRIBUTION_NAME + ' (Hyperledger Sawtooth) version {}').format( version), help='display version information') return parser.parse_args(args)
parse_args
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(args):\n parser = argparse.ArgumentParser(description=\n 'Starts the REST API application and connects to a specified validator.'\n )\n parser.add_argument('-B', '--bind', help=\n 'identify host and port for API to run on default: http://localhost:8008)'\n , action='append')\n parser.add_argument('-C', '--connect', help=\n 'specify URL to connect to a running validator')\n parser.add_argument('-t', '--timeout', help=\n 'set time (in seconds) to wait for validator response'\n )\n parser.add_argument('--client-max-size', type=int, help=\n 'the max size (in bytes) of a request body')\n parser.add_argument('-v', '--verbose', action='count', default=0, help=\n 'enable more verbose output to stderr')\n parser.add_argument('--opentsdb-url', help=\n 'specify host and port for Open TSDB database used for metrics'\n )\n parser.add_argument('--opentsdb-db', help=\n 'specify name of database for storing metrics')\n try:\n version = pkg_resources.get_distribution(DISTRIBUTION_NAME).version\n except pkg_resources.DistributionNotFound:\n version = 'UNKNOWN'\n parser.add_argument('-V', '--version', action='version', version=(\n DISTRIBUTION_NAME + ' (Hyperledger Sawtooth) version {}').format(\n version), help='display version information')\n return parser.parse_args(args)\n", "role": "user" }, { "content": "parse_args", "role": "assistant" } ]
def x(host, port, connection, timeout, registry, client_max_size=None): loop = asyncio.get_event_loop() connection.open() app = web.Application(loop=loop, client_max_size=client_max_size) app.on_cleanup.append(lambda app: connection.close()) LOGGER.info('Creating handlers for validator at %s', connection.url) handler = RouteHandler(loop, connection, timeout, registry) app.router.add_post('/batches', handler.submit_batches) app.router.add_get('/batch_statuses', handler.list_statuses) app.router.add_post('/batch_statuses', handler.list_statuses) app.router.add_get('/state', handler.list_state) app.router.add_get('/state/{address}', handler.fetch_state) app.router.add_get('/blocks', handler.list_blocks) app.router.add_get('/blocks/{block_id}', handler.fetch_block) app.router.add_get('/batches', handler.list_batches) app.router.add_get('/batches/{batch_id}', handler.fetch_batch) app.router.add_get('/transactions', handler.list_transactions) app.router.add_get('/transactions/{transaction_id}', handler. fetch_transaction) app.router.add_get('/receipts', handler.list_receipts) app.router.add_post('/receipts', handler.list_receipts) app.router.add_get('/peers', handler.fetch_peers) app.router.add_get('/status', handler.fetch_status) subscriber_handler = StateDeltaSubscriberHandler(connection) app.router.add_get('/subscriptions', subscriber_handler.subscriptions) app.on_shutdown.append(lambda app: subscriber_handler.on_shutdown()) LOGGER.info('Starting REST API on %s:%s', host, port) web.run_app(app, host=host, port=port, access_log=LOGGER, access_log_format='%r: %s status, %b size, in %Tf s')
start_rest_api
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(host, port, connection, timeout, registry, client_max_size=None):\n loop = asyncio.get_event_loop()\n connection.open()\n app = web.Application(loop=loop, client_max_size=client_max_size)\n app.on_cleanup.append(lambda app: connection.close())\n LOGGER.info('Creating handlers for validator at %s', connection.url)\n handler = RouteHandler(loop, connection, timeout, registry)\n app.router.add_post('/batches', handler.submit_batches)\n app.router.add_get('/batch_statuses', handler.list_statuses)\n app.router.add_post('/batch_statuses', handler.list_statuses)\n app.router.add_get('/state', handler.list_state)\n app.router.add_get('/state/{address}', handler.fetch_state)\n app.router.add_get('/blocks', handler.list_blocks)\n app.router.add_get('/blocks/{block_id}', handler.fetch_block)\n app.router.add_get('/batches', handler.list_batches)\n app.router.add_get('/batches/{batch_id}', handler.fetch_batch)\n app.router.add_get('/transactions', handler.list_transactions)\n app.router.add_get('/transactions/{transaction_id}', handler.\n fetch_transaction)\n app.router.add_get('/receipts', handler.list_receipts)\n app.router.add_post('/receipts', handler.list_receipts)\n app.router.add_get('/peers', handler.fetch_peers)\n app.router.add_get('/status', handler.fetch_status)\n subscriber_handler = StateDeltaSubscriberHandler(connection)\n app.router.add_get('/subscriptions', subscriber_handler.subscriptions)\n app.on_shutdown.append(lambda app: subscriber_handler.on_shutdown())\n LOGGER.info('Starting REST API on %s:%s', host, port)\n web.run_app(app, host=host, port=port, access_log=LOGGER,\n access_log_format='%r: %s status, %b size, in %Tf s')\n", "role": "user" }, { "content": "start_rest_api", "role": "assistant" } ]
def x(subparsers, parent_parser): parser = subparsers.add_parser('genesis', help= 'Creates the genesis.batch file for initializing the validator', description= 'Generates the genesis.batch file for initializing the validator.', epilog= 'This command generates a serialized GenesisData protobuf message and stores it in the genesis.batch file. One or more input files (optional) can contain serialized BatchList protobuf messages to add to the GenesisData. The output shows the location of this file. By default, the genesis.batch file is stored in /var/lib/sawtooth. If $SAWTOOTH_HOME is set, the location is $SAWTOOTH_HOME/data/genesis.batch. Use the --output option to change the name of the file.' , parents=[parent_parser]) parser.add_argument('-o', '--output', type=str, help= 'choose the output file for GenesisData') parser.add_argument('input_file', nargs='*', type=str, help= 'file or files containing batches to add to the resulting GenesisData')
add_genesis_parser
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(subparsers, parent_parser):\n parser = subparsers.add_parser('genesis', help=\n 'Creates the genesis.batch file for initializing the validator',\n description=\n 'Generates the genesis.batch file for initializing the validator.',\n epilog=\n 'This command generates a serialized GenesisData protobuf message and stores it in the genesis.batch file. One or more input files (optional) can contain serialized BatchList protobuf messages to add to the GenesisData. The output shows the location of this file. By default, the genesis.batch file is stored in /var/lib/sawtooth. If $SAWTOOTH_HOME is set, the location is $SAWTOOTH_HOME/data/genesis.batch. Use the --output option to change the name of the file.'\n , parents=[parent_parser])\n parser.add_argument('-o', '--output', type=str, help=\n 'choose the output file for GenesisData')\n parser.add_argument('input_file', nargs='*', type=str, help=\n 'file or files containing batches to add to the resulting GenesisData')\n", "role": "user" }, { "content": "add_genesis_parser", "role": "assistant" } ]
def x(args, data_dir=None): if data_dir is None: data_dir = get_data_dir() if not os.path.exists(data_dir): raise CliException('Data directory does not exist: {}'.format(data_dir) ) genesis_batches = [] for input_file in args.input_file: print('Processing {}...'.format(input_file)) input_data = BatchList() try: with open(input_file, 'rb') as in_file: input_data.ParseFromString(in_file.read()) except: raise CliException('Unable to read {}'.format(input_file)) genesis_batches += input_data.batches _validate_depedencies(genesis_batches) _check_required_settings(genesis_batches) if args.output: genesis_file = args.output else: genesis_file = os.path.join(data_dir, 'genesis.batch') print('Generating {}'.format(genesis_file)) output_data = GenesisData(batches=genesis_batches) with open(genesis_file, 'wb') as out_file: out_file.write(output_data.SerializeToString())
do_genesis
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(args, data_dir=None):\n if data_dir is None:\n data_dir = get_data_dir()\n if not os.path.exists(data_dir):\n raise CliException('Data directory does not exist: {}'.format(data_dir)\n )\n genesis_batches = []\n for input_file in args.input_file:\n print('Processing {}...'.format(input_file))\n input_data = BatchList()\n try:\n with open(input_file, 'rb') as in_file:\n input_data.ParseFromString(in_file.read())\n except:\n raise CliException('Unable to read {}'.format(input_file))\n genesis_batches += input_data.batches\n _validate_depedencies(genesis_batches)\n _check_required_settings(genesis_batches)\n if args.output:\n genesis_file = args.output\n else:\n genesis_file = os.path.join(data_dir, 'genesis.batch')\n print('Generating {}'.format(genesis_file))\n output_data = GenesisData(batches=genesis_batches)\n with open(genesis_file, 'wb') as out_file:\n out_file.write(output_data.SerializeToString())\n", "role": "user" }, { "content": "do_genesis", "role": "assistant" } ]
def x(batches): transaction_ids = set() for batch in batches: for txn in batch.transactions: txn_header = TransactionHeader() txn_header.ParseFromString(txn.header) if txn_header.dependencies: unsatisfied_deps = [id for id in txn_header.dependencies if id not in transaction_ids] if unsatisfied_deps: raise CliException( 'Unsatisfied dependency in given transactions: {}'. format(unsatisfied_deps)) transaction_ids.add(txn.header_signature)
validate_depedencies
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(batches):\n transaction_ids = set()\n for batch in batches:\n for txn in batch.transactions:\n txn_header = TransactionHeader()\n txn_header.ParseFromString(txn.header)\n if txn_header.dependencies:\n unsatisfied_deps = [id for id in txn_header.dependencies if\n id not in transaction_ids]\n if unsatisfied_deps:\n raise CliException(\n 'Unsatisfied dependency in given transactions: {}'.\n format(unsatisfied_deps))\n transaction_ids.add(txn.header_signature)\n", "role": "user" }, { "content": "validate_depedencies", "role": "assistant" } ]
def x(batches): required_settings = ['sawtooth.consensus.algorithm.name', 'sawtooth.consensus.algorithm.version'] for batch in batches: for txn in batch.transactions: txn_header = TransactionHeader() txn_header.ParseFromString(txn.header) if txn_header.family_name == 'sawtooth_settings': settings_payload = SettingsPayload() settings_payload.ParseFromString(txn.payload) if settings_payload.action == SettingsPayload.PROPOSE: proposal = SettingProposal() proposal.ParseFromString(settings_payload.data) if proposal.setting in required_settings: required_settings.remove(proposal.setting) if required_settings: raise CliException( 'The following setting(s) are required at genesis, but were not included in the genesis batches: {}' .format(required_settings))
check_required_settings
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(batches):\n required_settings = ['sawtooth.consensus.algorithm.name',\n 'sawtooth.consensus.algorithm.version']\n for batch in batches:\n for txn in batch.transactions:\n txn_header = TransactionHeader()\n txn_header.ParseFromString(txn.header)\n if txn_header.family_name == 'sawtooth_settings':\n settings_payload = SettingsPayload()\n settings_payload.ParseFromString(txn.payload)\n if settings_payload.action == SettingsPayload.PROPOSE:\n proposal = SettingProposal()\n proposal.ParseFromString(settings_payload.data)\n if proposal.setting in required_settings:\n required_settings.remove(proposal.setting)\n if required_settings:\n raise CliException(\n 'The following setting(s) are required at genesis, but were not included in the genesis batches: {}'\n .format(required_settings))\n", "role": "user" }, { "content": "check_required_settings", "role": "assistant" } ]
def x(self, key, index=None): records = self.get_multi([key], index=index) try: return records[0][1] except IndexError: return None
get
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, key, index=None):\n records = self.get_multi([key], index=index)\n try:\n return records[0][1]\n except IndexError:\n return None\n", "role": "user" }, { "content": "get", "role": "assistant" } ]
async def on_shutdown(self): """ Cleans up any outstanding subscriptions. """ await self._unregister_subscriptions() self._accepting = False for ws, _ in self._subscribers: await ws.close(code=aiohttp.WSCloseCode.GOING_AWAY, message= 'Server shutdown')
on_shutdown
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def on_shutdown(self):\n \"\"\"\n Cleans up any outstanding subscriptions.\n \"\"\"\n await self._unregister_subscriptions()\n self._accepting = False\n for ws, _ in self._subscribers:\n await ws.close(code=aiohttp.WSCloseCode.GOING_AWAY, message=\n 'Server shutdown')\n", "role": "user" }, { "content": "on_shutdown", "role": "assistant" } ]
async def subscriptions(self, request): """ Handles requests for new subscription websockets. Args: request (aiohttp.Request): the incoming request Returns: aiohttp.web.WebSocketResponse: the websocket response, when the resulting websocket is closed """ if not self._accepting: return web.Response(status=503) web_sock = web.WebSocketResponse() await web_sock.prepare(request) async for msg in web_sock: if msg.type == aiohttp.WSMsgType.TEXT: await self._handle_message(web_sock, msg.data) elif msg.type == aiohttp.WSMsgType.ERROR: LOGGER.warning('Web socket connection closed with exception %s', web_sock.exception()) await web_sock.close() await self._handle_unsubscribe(web_sock) return web_sock
subscriptions
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def subscriptions(self, request):\n \"\"\"\n Handles requests for new subscription websockets.\n\n Args:\n request (aiohttp.Request): the incoming request\n\n Returns:\n aiohttp.web.WebSocketResponse: the websocket response, when the\n resulting websocket is closed\n \"\"\"\n if not self._accepting:\n return web.Response(status=503)\n web_sock = web.WebSocketResponse()\n await web_sock.prepare(request)\n async for msg in web_sock:\n if msg.type == aiohttp.WSMsgType.TEXT:\n await self._handle_message(web_sock, msg.data)\n elif msg.type == aiohttp.WSMsgType.ERROR:\n LOGGER.warning('Web socket connection closed with exception %s',\n web_sock.exception())\n await web_sock.close()\n await self._handle_unsubscribe(web_sock)\n return web_sock\n", "role": "user" }, { "content": "subscriptions", "role": "assistant" } ]
def x(self, request, result): self._in_process_transactions_count.dec() req = processor_pb2.TpProcessRequest() req.ParseFromString(request) response = processor_pb2.TpProcessResponse() response.ParseFromString(result.content) processor_type = ProcessorType(req.header.family_name, req.header. family_version) self._processor_manager[processor_type].get_processor(result.connection_id ).dec_occupancy() self._processor_manager.notify() self._get_tp_process_response_counter(response.Status.Name(response.status) ).inc() if (result.connection_id in self._open_futures and req.signature in self._open_futures[result.connection_id]): del self._open_futures[result.connection_id][req.signature] if response.status == processor_pb2.TpProcessResponse.OK: state_sets, state_deletes, events, data = (self._context_manager. get_execution_results(req.context_id)) state_changes = [transaction_receipt_pb2.StateChange(address=addr, value=value, type=transaction_receipt_pb2.StateChange.SET) for addr, value in state_sets.items()] + [transaction_receipt_pb2. StateChange(address=addr, type=transaction_receipt_pb2. StateChange.DELETE) for addr in state_deletes] self._scheduler.set_transaction_execution_result(txn_signature=req. signature, is_valid=True, context_id=req.context_id, state_changes=state_changes, events=events, data=data) elif response.status == processor_pb2.TpProcessResponse.INTERNAL_ERROR: LOGGER.error( 'Transaction processor internal error: %s (transaction: %s, name: %s, version: %s)' , response.message, req.signature, req.header.family_name, req. header.family_version) if self._scheduler.is_transaction_in_schedule(req.signature): self._execute(processor_type=processor_type, content=request, signature=req.signature) else: self._context_manager.delete_contexts(context_id_list=[req.context_id]) self._fail_transaction(txn_signature=req.signature, context_id=req. context_id, error_message=response.message, error_data=response .extended_data)
future_done_callback
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, request, result):\n self._in_process_transactions_count.dec()\n req = processor_pb2.TpProcessRequest()\n req.ParseFromString(request)\n response = processor_pb2.TpProcessResponse()\n response.ParseFromString(result.content)\n processor_type = ProcessorType(req.header.family_name, req.header.\n family_version)\n self._processor_manager[processor_type].get_processor(result.connection_id\n ).dec_occupancy()\n self._processor_manager.notify()\n self._get_tp_process_response_counter(response.Status.Name(response.status)\n ).inc()\n if (result.connection_id in self._open_futures and req.signature in\n self._open_futures[result.connection_id]):\n del self._open_futures[result.connection_id][req.signature]\n if response.status == processor_pb2.TpProcessResponse.OK:\n state_sets, state_deletes, events, data = (self._context_manager.\n get_execution_results(req.context_id))\n state_changes = [transaction_receipt_pb2.StateChange(address=addr,\n value=value, type=transaction_receipt_pb2.StateChange.SET) for \n addr, value in state_sets.items()] + [transaction_receipt_pb2.\n StateChange(address=addr, type=transaction_receipt_pb2.\n StateChange.DELETE) for addr in state_deletes]\n self._scheduler.set_transaction_execution_result(txn_signature=req.\n signature, is_valid=True, context_id=req.context_id,\n state_changes=state_changes, events=events, data=data)\n elif response.status == processor_pb2.TpProcessResponse.INTERNAL_ERROR:\n LOGGER.error(\n 'Transaction processor internal error: %s (transaction: %s, name: %s, version: %s)'\n , response.message, req.signature, req.header.family_name, req.\n header.family_version)\n if self._scheduler.is_transaction_in_schedule(req.signature):\n self._execute(processor_type=processor_type, content=request,\n signature=req.signature)\n else:\n self._context_manager.delete_contexts(context_id_list=[req.context_id])\n self._fail_transaction(txn_signature=req.signature, context_id=req.\n context_id, error_message=response.message, error_data=response\n .extended_data)\n", "role": "user" }, { "content": "future_done_callback", "role": "assistant" } ]
def x(subparsers, parent_parser): parser = subparsers.add_parser('batch', help= 'Displays information about batches and submit new batches', description= 'Provides subcommands to display Batch information and submit Batches to the validator via the REST API.' ) grand_parsers = parser.add_subparsers(title='subcommands', dest= 'subcommand') grand_parsers.required = True add_batch_list_parser(grand_parsers, parent_parser) add_batch_show_parser(grand_parsers, parent_parser) add_batch_status_parser(grand_parsers, parent_parser) add_batch_submit_parser(grand_parsers, parent_parser)
add_batch_parser
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(subparsers, parent_parser):\n parser = subparsers.add_parser('batch', help=\n 'Displays information about batches and submit new batches',\n description=\n 'Provides subcommands to display Batch information and submit Batches to the validator via the REST API.'\n )\n grand_parsers = parser.add_subparsers(title='subcommands', dest=\n 'subcommand')\n grand_parsers.required = True\n add_batch_list_parser(grand_parsers, parent_parser)\n add_batch_show_parser(grand_parsers, parent_parser)\n add_batch_status_parser(grand_parsers, parent_parser)\n add_batch_submit_parser(grand_parsers, parent_parser)\n", "role": "user" }, { "content": "add_batch_parser", "role": "assistant" } ]
def x(args): if args.subcommand == 'list': do_batch_list(args) if args.subcommand == 'show': do_batch_show(args) if args.subcommand == 'status': do_batch_status(args) if args.subcommand == 'submit': do_batch_submit(args)
do_batch
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(args):\n if args.subcommand == 'list':\n do_batch_list(args)\n if args.subcommand == 'show':\n do_batch_show(args)\n if args.subcommand == 'status':\n do_batch_status(args)\n if args.subcommand == 'submit':\n do_batch_submit(args)\n", "role": "user" }, { "content": "do_batch", "role": "assistant" } ]
def x(args): rest_client = RestClient(args.url, args.user) batch_ids = args.batch_ids.split(',') if args.wait and args.wait > 0: statuses = rest_client.get_statuses(batch_ids, args.wait) else: statuses = rest_client.get_statuses(batch_ids) if args.format == 'yaml': fmt.print_yaml(statuses) elif args.format == 'json': fmt.print_json(statuses) else: raise AssertionError('Missing handler: {}'.format(args.format))
do_batch_status
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(args):\n rest_client = RestClient(args.url, args.user)\n batch_ids = args.batch_ids.split(',')\n if args.wait and args.wait > 0:\n statuses = rest_client.get_statuses(batch_ids, args.wait)\n else:\n statuses = rest_client.get_statuses(batch_ids)\n if args.format == 'yaml':\n fmt.print_yaml(statuses)\n elif args.format == 'json':\n fmt.print_json(statuses)\n else:\n raise AssertionError('Missing handler: {}'.format(args.format))\n", "role": "user" }, { "content": "do_batch_status", "role": "assistant" } ]
def x(self, block): if block.header_signature in self._block_manager: LOGGER.debug('Drop duplicate block: %s', block) return None if block.previous_block_id not in self._block_manager: return self._request_previous_if_not_already_requested(block) if len(block.batches) > len(block.header.batch_ids): LOGGER.debug('Block has extra batches. Dropping %s', block) return None temp_batches = {} for batch in block.batches: temp_batches[batch.header_signature] = batch if len(block.batches) != len(block.header.batch_ids): building = True for batch_id in block.header.batch_ids: if (batch_id not in self._batch_cache and batch_id not in temp_batches): if batch_id not in self._incomplete_blocks: self._incomplete_blocks[batch_id] = [block] elif block not in self._incomplete_blocks[batch_id]: self._incomplete_blocks[batch_id] += [block] if batch_id in self._requested: return None self._requested[batch_id] = None self._gossip.broadcast_batch_by_batch_id_request(batch_id) building = False if not building: return None batches = self._finalize_batch_list(block, temp_batches) del block.batches[:] block.batches.extend(batches) if block.header_signature in self._requested: del self._requested[block.header_signature] return self._put_or_request_if_missing_predecessor(block) batch_id_list = [x.header_signature for x in block.batches] if batch_id_list == list(block.header.batch_ids): if block.header_signature in self._requested: del self._requested[block.header_signature] return self._put_or_request_if_missing_predecessor(block) if sorted(batch_id_list) == sorted(list(block.header.batch_ids)): batches = self._finalize_batch_list(block, temp_batches) del block.batches[:] if batches is not None: block.batches.extend(batches) else: return None if block.header_signature in self._requested: del self._requested[block.header_signature] return self._put_or_request_if_missing_predecessor(block) LOGGER.debug( 'Block.header.batch_ids does not match set of batches in block.batches Dropping %s' , block) return None
complete_block
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, block):\n if block.header_signature in self._block_manager:\n LOGGER.debug('Drop duplicate block: %s', block)\n return None\n if block.previous_block_id not in self._block_manager:\n return self._request_previous_if_not_already_requested(block)\n if len(block.batches) > len(block.header.batch_ids):\n LOGGER.debug('Block has extra batches. Dropping %s', block)\n return None\n temp_batches = {}\n for batch in block.batches:\n temp_batches[batch.header_signature] = batch\n if len(block.batches) != len(block.header.batch_ids):\n building = True\n for batch_id in block.header.batch_ids:\n if (batch_id not in self._batch_cache and batch_id not in\n temp_batches):\n if batch_id not in self._incomplete_blocks:\n self._incomplete_blocks[batch_id] = [block]\n elif block not in self._incomplete_blocks[batch_id]:\n self._incomplete_blocks[batch_id] += [block]\n if batch_id in self._requested:\n return None\n self._requested[batch_id] = None\n self._gossip.broadcast_batch_by_batch_id_request(batch_id)\n building = False\n if not building:\n return None\n batches = self._finalize_batch_list(block, temp_batches)\n del block.batches[:]\n block.batches.extend(batches)\n if block.header_signature in self._requested:\n del self._requested[block.header_signature]\n return self._put_or_request_if_missing_predecessor(block)\n batch_id_list = [x.header_signature for x in block.batches]\n if batch_id_list == list(block.header.batch_ids):\n if block.header_signature in self._requested:\n del self._requested[block.header_signature]\n return self._put_or_request_if_missing_predecessor(block)\n if sorted(batch_id_list) == sorted(list(block.header.batch_ids)):\n batches = self._finalize_batch_list(block, temp_batches)\n del block.batches[:]\n if batches is not None:\n block.batches.extend(batches)\n else:\n return None\n if block.header_signature in self._requested:\n del self._requested[block.header_signature]\n return self._put_or_request_if_missing_predecessor(block)\n LOGGER.debug(\n 'Block.header.batch_ids does not match set of batches in block.batches Dropping %s'\n , block)\n return None\n", "role": "user" }, { "content": "complete_block", "role": "assistant" } ]
def x(args): rest_client = RestClient(args.url, args.user) if args.subcommand == 'list': response = rest_client.list_state(args.subtree, args.head) leaves = response['data'] head = response['head'] keys = 'address', 'size', 'data' headers = tuple(k.upper() for k in keys) def x(leaf, decode=True): decoded = b64decode(leaf['data']) return leaf['address'], len(decoded), str(decoded ) if decode else leaf['data'] if args.format == 'default': fmt.print_terminal_table(headers, leaves, parse_leaf_row) print('HEAD BLOCK: "{}"'.format(head)) elif args.format == 'csv': fmt.print_csv(headers, leaves, parse_leaf_row) print('(data for head block: "{}")'.format(head)) elif args.format == 'json' or args.format == 'yaml': state_data = {'head': head, 'data': [{k: d for k, d in zip(keys, parse_leaf_row(l, False))} for l in leaves]} if args.format == 'yaml': fmt.print_yaml(state_data) elif args.format == 'json': fmt.print_json(state_data) else: raise AssertionError('Missing handler: {}'.format(args.format)) else: raise AssertionError('Missing handler: {}'.format(args.format)) if args.subcommand == 'show': output = rest_client.get_leaf(args.address, args.head) if output is not None: print('DATA: "{}"'.format(b64decode(output['data']))) print('HEAD: "{}"'.format(output['head'])) else: raise CliException('No data available at {}'.format(args.address))
do_state
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(args):\n rest_client = RestClient(args.url, args.user)\n if args.subcommand == 'list':\n response = rest_client.list_state(args.subtree, args.head)\n leaves = response['data']\n head = response['head']\n keys = 'address', 'size', 'data'\n headers = tuple(k.upper() for k in keys)\n\n def x(leaf, decode=True):\n decoded = b64decode(leaf['data'])\n return leaf['address'], len(decoded), str(decoded\n ) if decode else leaf['data']\n if args.format == 'default':\n fmt.print_terminal_table(headers, leaves, parse_leaf_row)\n print('HEAD BLOCK: \"{}\"'.format(head))\n elif args.format == 'csv':\n fmt.print_csv(headers, leaves, parse_leaf_row)\n print('(data for head block: \"{}\")'.format(head))\n elif args.format == 'json' or args.format == 'yaml':\n state_data = {'head': head, 'data': [{k: d for k, d in zip(keys,\n parse_leaf_row(l, False))} for l in leaves]}\n if args.format == 'yaml':\n fmt.print_yaml(state_data)\n elif args.format == 'json':\n fmt.print_json(state_data)\n else:\n raise AssertionError('Missing handler: {}'.format(args.format))\n else:\n raise AssertionError('Missing handler: {}'.format(args.format))\n if args.subcommand == 'show':\n output = rest_client.get_leaf(args.address, args.head)\n if output is not None:\n print('DATA: \"{}\"'.format(b64decode(output['data'])))\n print('HEAD: \"{}\"'.format(output['head']))\n else:\n raise CliException('No data available at {}'.format(args.address))\n", "role": "user" }, { "content": "do_state", "role": "assistant" } ]
def x(self, required_state_root): state_hash = None if self._previous_valid_batch_c_id is not None: publishing_or_genesis = (self._always_persist or required_state_root is None) state_hash = self._squash(state_root=self._previous_state_hash, context_ids=[self._previous_valid_batch_c_id], persist=self. _always_persist, clean_up=publishing_or_genesis) if self._always_persist is True: return state_hash if state_hash == required_state_root: self._squash(state_root=self._previous_state_hash, context_ids= [self._previous_valid_batch_c_id], persist=True, clean_up=True) return state_hash
compute_merkle_root
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, required_state_root):\n state_hash = None\n if self._previous_valid_batch_c_id is not None:\n publishing_or_genesis = (self._always_persist or \n required_state_root is None)\n state_hash = self._squash(state_root=self._previous_state_hash,\n context_ids=[self._previous_valid_batch_c_id], persist=self.\n _always_persist, clean_up=publishing_or_genesis)\n if self._always_persist is True:\n return state_hash\n if state_hash == required_state_root:\n self._squash(state_root=self._previous_state_hash, context_ids=\n [self._previous_valid_batch_c_id], persist=True, clean_up=True)\n return state_hash\n", "role": "user" }, { "content": "compute_merkle_root", "role": "assistant" } ]
def x(value): block = Block() block.ParseFromString(value) return BlockWrapper(block=block)
deserialize_block
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(value):\n block = Block()\n block.ParseFromString(value)\n return BlockWrapper(block=block)\n", "role": "user" }, { "content": "deserialize_block", "role": "assistant" } ]
def x(self): vec_ptr, vec_len, vec_cap = ffi.prepare_vec_result() try: _libexec('commit_store_get_chain_head', self.pointer, ctypes.byref( vec_ptr), ctypes.byref(vec_len), ctypes.byref(vec_cap)) except ValueError: return None return self.deserialize_block(ffi.from_rust_vec(vec_ptr, vec_len, vec_cap))
chain_head
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n vec_ptr, vec_len, vec_cap = ffi.prepare_vec_result()\n try:\n _libexec('commit_store_get_chain_head', self.pointer, ctypes.byref(\n vec_ptr), ctypes.byref(vec_len), ctypes.byref(vec_cap))\n except ValueError:\n return None\n return self.deserialize_block(ffi.from_rust_vec(vec_ptr, vec_len, vec_cap))\n", "role": "user" }, { "content": "chain_head", "role": "assistant" } ]
def x(self, start_block=None, start_block_num=None, reverse=True): start = None if start_block_num: if len(start_block_num) < 2: raise ValueError('Invalid start block num') if start_block_num[:2] != '0x': raise ValueError('Invalid start block num') start = int(start_block_num, 16) elif start_block: start = start_block.block_num return _BlockStoreIter(self.pointer, start, reverse)
get_block_iter
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, start_block=None, start_block_num=None, reverse=True):\n start = None\n if start_block_num:\n if len(start_block_num) < 2:\n raise ValueError('Invalid start block num')\n if start_block_num[:2] != '0x':\n raise ValueError('Invalid start block num')\n start = int(start_block_num, 16)\n elif start_block:\n start = start_block.block_num\n return _BlockStoreIter(self.pointer, start, reverse)\n", "role": "user" }, { "content": "get_block_iter", "role": "assistant" } ]
def x(self, block_ids): return list(filter(lambda b: b is not None, map(self. _get_block_by_id_or_none, block_ids)))
get_blocks
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, block_ids):\n return list(filter(lambda b: b is not None, map(self.\n _get_block_by_id_or_none, block_ids)))\n", "role": "user" }, { "content": "get_blocks", "role": "assistant" } ]
def x(self, transaction_id): payload = self._get_data_by_id(transaction_id, 'commit_store_get_batch_by_transaction') batch = Batch() batch.ParseFromString(payload) return batch
get_batch_by_transaction
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, transaction_id):\n payload = self._get_data_by_id(transaction_id,\n 'commit_store_get_batch_by_transaction')\n batch = Batch()\n batch.ParseFromString(payload)\n return batch\n", "role": "user" }, { "content": "get_batch_by_transaction", "role": "assistant" } ]
def x(self, batch_id): payload = self._get_data_by_id(batch_id, 'commit_store_get_batch') batch = Batch() batch.ParseFromString(payload) return batch
get_batch
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, batch_id):\n payload = self._get_data_by_id(batch_id, 'commit_store_get_batch')\n batch = Batch()\n batch.ParseFromString(payload)\n return batch\n", "role": "user" }, { "content": "get_batch", "role": "assistant" } ]
def x(self, transaction_id): payload = self._get_data_by_id(transaction_id, 'commit_store_get_transaction') txn = Transaction() txn.ParseFromString(payload) return txn
get_transaction
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, transaction_id):\n payload = self._get_data_by_id(transaction_id,\n 'commit_store_get_transaction')\n txn = Transaction()\n txn.ParseFromString(payload)\n return txn\n", "role": "user" }, { "content": "get_transaction", "role": "assistant" } ]
def x(headers, example_row): def x(col): if isinstance(col, str): return '{{:{w}.{w}}}' return '{{:<{w}}}' widths = [max(len(h), len(str(d))) for h, d in zip(headers, example_row)] original_last_width = widths[-1] if sys.stdout.isatty(): widths[-1] = max(len(headers[-1]), tty.width() - sum(w + 2 for w in widths[0:-1]) - 3) cols = [format_column(c).format(w=w) for c, w in zip(example_row, widths)] format_string = ' '.join(cols) if original_last_width > widths[-1]: format_string += '...' return format_string
format_terminal_row
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(headers, example_row):\n\n def x(col):\n if isinstance(col, str):\n return '{{:{w}.{w}}}'\n return '{{:<{w}}}'\n widths = [max(len(h), len(str(d))) for h, d in zip(headers, example_row)]\n original_last_width = widths[-1]\n if sys.stdout.isatty():\n widths[-1] = max(len(headers[-1]), tty.width() - sum(w + 2 for w in\n widths[0:-1]) - 3)\n cols = [format_column(c).format(w=w) for c, w in zip(example_row, widths)]\n format_string = ' '.join(cols)\n if original_last_width > widths[-1]:\n format_string += '...'\n return format_string\n", "role": "user" }, { "content": "format_terminal_row", "role": "assistant" } ]
def x(headers, data_list, parse_row_fn): data_iter = iter(data_list) try: example = next(data_iter) example_row = parse_row_fn(example) data_iter = itertools.chain([example], data_iter) except StopIteration: example_row = [''] * len(headers) format_string = format_terminal_row(headers, example_row) top_row = format_string.format(*headers) print(top_row[0:-3] if top_row.endswith('...') else top_row) for data in data_iter: print(format_string.format(*parse_row_fn(data)))
print_terminal_table
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(headers, data_list, parse_row_fn):\n data_iter = iter(data_list)\n try:\n example = next(data_iter)\n example_row = parse_row_fn(example)\n data_iter = itertools.chain([example], data_iter)\n except StopIteration:\n example_row = [''] * len(headers)\n format_string = format_terminal_row(headers, example_row)\n top_row = format_string.format(*headers)\n print(top_row[0:-3] if top_row.endswith('...') else top_row)\n for data in data_iter:\n print(format_string.format(*parse_row_fn(data)))\n", "role": "user" }, { "content": "print_terminal_table", "role": "assistant" } ]
def x(headers, data_list, parse_row_fn): try: writer = csv.writer(sys.stdout) writer.writerow(headers) for data in data_list: writer.writerow(parse_row_fn(data)) except csv.Error as e: raise CliException('Error writing CSV: {}'.format(e))
print_csv
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(headers, data_list, parse_row_fn):\n try:\n writer = csv.writer(sys.stdout)\n writer.writerow(headers)\n for data in data_list:\n writer.writerow(parse_row_fn(data))\n except csv.Error as e:\n raise CliException('Error writing CSV: {}'.format(e))\n", "role": "user" }, { "content": "print_csv", "role": "assistant" } ]
def x(self, key, item): with self._lock: if key in self._dict: self._dict[key].append(item) else: self._dict[key] = [item]
append
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, key, item):\n with self._lock:\n if key in self._dict:\n self._dict[key].append(item)\n else:\n self._dict[key] = [item]\n", "role": "user" }, { "content": "append", "role": "assistant" } ]
def x(self, key, items): if not isinstance(items, list): raise ValueError('items must be a list') with self._lock: self._dict[key] = items.copy()
set
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, key, items):\n if not isinstance(items, list):\n raise ValueError('items must be a list')\n with self._lock:\n self._dict[key] = items.copy()\n", "role": "user" }, { "content": "set", "role": "assistant" } ]
def x(self, key, items): if not isinstance(items, list): raise ValueError('items must be a list') return_value = [] with self._lock: if key in self._dict: return_value = self._dict[key] self._dict[key] = items.copy() return return_value
swap
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, key, items):\n if not isinstance(items, list):\n raise ValueError('items must be a list')\n return_value = []\n with self._lock:\n if key in self._dict:\n return_value = self._dict[key]\n self._dict[key] = items.copy()\n return return_value\n", "role": "user" }, { "content": "swap", "role": "assistant" } ]
def x(self, key, default): with self._lock: return self._dict.pop(key, default)
pop
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, key, default):\n with self._lock:\n return self._dict.pop(key, default)\n", "role": "user" }, { "content": "pop", "role": "assistant" } ]
def x(self, key, default): with self._lock: try: return self._dict[key].copy() except KeyError: return default
get
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, key, default):\n with self._lock:\n try:\n return self._dict[key].copy()\n except KeyError:\n return default\n", "role": "user" }, { "content": "get", "role": "assistant" } ]
def x(self, block, receipts): with self._lock: for batch_id in self._pending.copy(): if self._batch_committed(batch_id): self._pending.remove(batch_id) self._update_observers(batch_id, ClientBatchStatus.COMMITTED)
chain_update
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, block, receipts):\n with self._lock:\n for batch_id in self._pending.copy():\n if self._batch_committed(batch_id):\n self._pending.remove(batch_id)\n self._update_observers(batch_id, ClientBatchStatus.COMMITTED)\n", "role": "user" }, { "content": "chain_update", "role": "assistant" } ]
def x(self, txn_id, message=None, extended_data=None): invalid_txn_info = {'id': txn_id} if message is not None: invalid_txn_info['message'] = message if extended_data is not None: invalid_txn_info['extended_data'] = extended_data with self._lock: for batch_id, txn_ids in self._batch_info.items(): if txn_id in txn_ids: if batch_id not in self._invalid: self._invalid[batch_id] = [invalid_txn_info] else: self._invalid[batch_id].append(invalid_txn_info) self._pending.discard(batch_id) self._update_observers(batch_id, ClientBatchStatus.INVALID) return
notify_txn_invalid
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, txn_id, message=None, extended_data=None):\n invalid_txn_info = {'id': txn_id}\n if message is not None:\n invalid_txn_info['message'] = message\n if extended_data is not None:\n invalid_txn_info['extended_data'] = extended_data\n with self._lock:\n for batch_id, txn_ids in self._batch_info.items():\n if txn_id in txn_ids:\n if batch_id not in self._invalid:\n self._invalid[batch_id] = [invalid_txn_info]\n else:\n self._invalid[batch_id].append(invalid_txn_info)\n self._pending.discard(batch_id)\n self._update_observers(batch_id, ClientBatchStatus.INVALID)\n return\n", "role": "user" }, { "content": "notify_txn_invalid", "role": "assistant" } ]
def x(self, batch): txn_ids = {t.header_signature for t in batch.transactions} with self._lock: self._pending.add(batch.header_signature) self._batch_info[batch.header_signature] = txn_ids self._update_observers(batch.header_signature, ClientBatchStatus. PENDING)
notify_batch_pending
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, batch):\n txn_ids = {t.header_signature for t in batch.transactions}\n with self._lock:\n self._pending.add(batch.header_signature)\n self._batch_info[batch.header_signature] = txn_ids\n self._update_observers(batch.header_signature, ClientBatchStatus.\n PENDING)\n", "role": "user" }, { "content": "notify_batch_pending", "role": "assistant" } ]
def x(self, batch_id): with self._lock: if self._batch_committed(batch_id): return ClientBatchStatus.COMMITTED if batch_id in self._invalid: return ClientBatchStatus.INVALID if batch_id in self._pending: return ClientBatchStatus.PENDING return ClientBatchStatus.UNKNOWN
get_status
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, batch_id):\n with self._lock:\n if self._batch_committed(batch_id):\n return ClientBatchStatus.COMMITTED\n if batch_id in self._invalid:\n return ClientBatchStatus.INVALID\n if batch_id in self._pending:\n return ClientBatchStatus.PENDING\n return ClientBatchStatus.UNKNOWN\n", "role": "user" }, { "content": "get_status", "role": "assistant" } ]
def x(self, batch_ids): with self._lock: return {b: self.get_status(b) for b in batch_ids}
get_statuses
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, batch_ids):\n with self._lock:\n return {b: self.get_status(b) for b in batch_ids}\n", "role": "user" }, { "content": "get_statuses", "role": "assistant" } ]
def x(self, batch_id): with self._lock: return [info.copy() for info in self._invalid.get(batch_id, [])]
get_invalid_txn_info
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, batch_id):\n with self._lock:\n return [info.copy() for info in self._invalid.get(batch_id, [])]\n", "role": "user" }, { "content": "get_invalid_txn_info", "role": "assistant" } ]
def x(self, observer, batch_ids): with self._lock: statuses = self.get_statuses(batch_ids) if self._has_no_pendings(statuses): observer.notify_batches_finished(statuses) else: self._observers[observer] = statuses
watch_statuses
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, observer, batch_ids):\n with self._lock:\n statuses = self.get_statuses(batch_ids)\n if self._has_no_pendings(statuses):\n observer.notify_batches_finished(statuses)\n else:\n self._observers[observer] = statuses\n", "role": "user" }, { "content": "watch_statuses", "role": "assistant" } ]
def x(self, batch_id, status): for observer, statuses in self._observers.copy().items(): if batch_id in statuses: statuses[batch_id] = status if self._has_no_pendings(statuses): observer.notify_batches_finished(statuses) self._observers.pop(observer)
update_observers
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, batch_id, status):\n for observer, statuses in self._observers.copy().items():\n if batch_id in statuses:\n statuses[batch_id] = status\n if self._has_no_pendings(statuses):\n observer.notify_batches_finished(statuses)\n self._observers.pop(observer)\n", "role": "user" }, { "content": "update_observers", "role": "assistant" } ]