source_code
stringlengths
45
7.24k
function_name
stringlengths
1
111
messages
list
def x(self, block_ids): try: return [(block_id.hex(), self._chain_controller. block_validation_result(block_id.hex())) for block_id in block_ids] except KeyError as key_error: raise UnknownBlock(key_error.args[0])
get_block_statuses
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, block_ids):\n try:\n return [(block_id.hex(), self._chain_controller.\n block_validation_result(block_id.hex())) for block_id in block_ids]\n except KeyError as key_error:\n raise UnknownBlock(key_error.args[0])\n", "role": "user" }, { "content": "get_block_statuses", "role": "assistant" } ]
def x(self, block_id, settings): block = self._get_blocks([block_id.hex()])[0] block_header = BlockHeader() block_header.ParseFromString(block.header) try: settings_view = self._settings_view_factory.create_settings_view( block_header.state_root_hash) except KeyError: LOGGER.error( 'Settings from block %s requested, but root hash %s was missing. Returning no setting values.' , block_id.hex(), block_header.state_root_hash) return [] result = [] for setting in settings: try: value = settings_view.get_setting(setting) except KeyError: continue result.append((setting, value)) return result
settings_get
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, block_id, settings):\n block = self._get_blocks([block_id.hex()])[0]\n block_header = BlockHeader()\n block_header.ParseFromString(block.header)\n try:\n settings_view = self._settings_view_factory.create_settings_view(\n block_header.state_root_hash)\n except KeyError:\n LOGGER.error(\n 'Settings from block %s requested, but root hash %s was missing. Returning no setting values.'\n , block_id.hex(), block_header.state_root_hash)\n return []\n result = []\n for setting in settings:\n try:\n value = settings_view.get_setting(setting)\n except KeyError:\n continue\n result.append((setting, value))\n return result\n", "role": "user" }, { "content": "settings_get", "role": "assistant" } ]
def x(self, block_id, addresses): block = self._get_blocks([block_id.hex()])[0] block_header = BlockHeader() block_header.ParseFromString(block.header) try: state_view = self._state_view_factory.create_view(block_header. state_root_hash) except KeyError: LOGGER.error( 'State from block %s requested, but root hash %s was missing. Returning empty state.' , block_id.hex(), block_header.state_root_hash) return [] result = [] for address in addresses: if len(address) == 70: try: value = state_view.get(address) except KeyError: continue result.append((address, value)) continue leaves = state_view.leaves(address) for leaf in leaves: result.append(leaf) return result
state_get
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, block_id, addresses):\n block = self._get_blocks([block_id.hex()])[0]\n block_header = BlockHeader()\n block_header.ParseFromString(block.header)\n try:\n state_view = self._state_view_factory.create_view(block_header.\n state_root_hash)\n except KeyError:\n LOGGER.error(\n 'State from block %s requested, but root hash %s was missing. Returning empty state.'\n , block_id.hex(), block_header.state_root_hash)\n return []\n result = []\n for address in addresses:\n if len(address) == 70:\n try:\n value = state_view.get(address)\n except KeyError:\n continue\n result.append((address, value))\n continue\n leaves = state_view.leaves(address)\n for leaf in leaves:\n result.append(leaf)\n return result\n", "role": "user" }, { "content": "state_get", "role": "assistant" } ]
def x(subparsers, parent_parser): settings_parser = subparsers.add_parser('settings', help= 'Displays on-chain settings', description= 'Displays the values of currently active on-chain settings.') settings_parsers = settings_parser.add_subparsers(title='settings', dest='settings_cmd') settings_parsers.required = True list_parser = settings_parsers.add_parser('list', help= 'Lists the current keys and values of on-chain settings', description= 'List the current keys and values of on-chain settings. The content can be exported to various formats for external consumption.' ) list_parser.add_argument('--url', type=str, help= "identify the URL of a validator's REST API", default= 'http://localhost:8008') list_parser.add_argument('--filter', type=str, default='', help= 'filters keys that begin with this value') list_parser.add_argument('--format', default='default', choices=[ 'default', 'csv', 'json', 'yaml'], help='choose the output format')
add_settings_parser
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(subparsers, parent_parser):\n settings_parser = subparsers.add_parser('settings', help=\n 'Displays on-chain settings', description=\n 'Displays the values of currently active on-chain settings.')\n settings_parsers = settings_parser.add_subparsers(title='settings',\n dest='settings_cmd')\n settings_parsers.required = True\n list_parser = settings_parsers.add_parser('list', help=\n 'Lists the current keys and values of on-chain settings',\n description=\n 'List the current keys and values of on-chain settings. The content can be exported to various formats for external consumption.'\n )\n list_parser.add_argument('--url', type=str, help=\n \"identify the URL of a validator's REST API\", default=\n 'http://localhost:8008')\n list_parser.add_argument('--filter', type=str, default='', help=\n 'filters keys that begin with this value')\n list_parser.add_argument('--format', default='default', choices=[\n 'default', 'csv', 'json', 'yaml'], help='choose the output format')\n", "role": "user" }, { "content": "add_settings_parser", "role": "assistant" } ]
def x(args): rest_client = RestClient(args.url) state = rest_client.list_state(subtree=SETTINGS_NAMESPACE) prefix = args.filter head = state['head'] state_values = state['data'] printable_settings = [] proposals_address = _key_to_address('sawtooth.settings.vote.proposals') for state_value in state_values: if state_value['address'] == proposals_address: continue decoded = b64decode(state_value['data']) setting = Setting() setting.ParseFromString(decoded) for entry in setting.entries: if entry.key.startswith(prefix): printable_settings.append(entry) printable_settings.sort(key=lambda s: s.key) if args.format == 'default': tty_width = tty.width() for setting in printable_settings: width = tty_width - len(setting.key) - 3 width = width if width > _MIN_PRINT_WIDTH else _MIN_PRINT_WIDTH value = setting.value[:width] + '...' if len(setting.value ) > width else setting.value print('{}: {}'.format(setting.key, value)) elif args.format == 'csv': try: writer = csv.writer(sys.stdout, quoting=csv.QUOTE_ALL) writer.writerow(['KEY', 'VALUE']) for setting in printable_settings: writer.writerow([setting.key, setting.value]) except csv.Error: raise CliException('Error writing CSV') elif args.format == 'json' or args.format == 'yaml': settings_snapshot = {'head': head, 'settings': {setting.key: setting.value for setting in printable_settings}} if args.format == 'json': print(json.dumps(settings_snapshot, indent=2, sort_keys=True)) else: print(yaml.dump(settings_snapshot, default_flow_style=False)[0:-1]) else: raise AssertionError('Unknown format {}'.format(args.format))
do_settings_list
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(args):\n rest_client = RestClient(args.url)\n state = rest_client.list_state(subtree=SETTINGS_NAMESPACE)\n prefix = args.filter\n head = state['head']\n state_values = state['data']\n printable_settings = []\n proposals_address = _key_to_address('sawtooth.settings.vote.proposals')\n for state_value in state_values:\n if state_value['address'] == proposals_address:\n continue\n decoded = b64decode(state_value['data'])\n setting = Setting()\n setting.ParseFromString(decoded)\n for entry in setting.entries:\n if entry.key.startswith(prefix):\n printable_settings.append(entry)\n printable_settings.sort(key=lambda s: s.key)\n if args.format == 'default':\n tty_width = tty.width()\n for setting in printable_settings:\n width = tty_width - len(setting.key) - 3\n width = width if width > _MIN_PRINT_WIDTH else _MIN_PRINT_WIDTH\n value = setting.value[:width] + '...' if len(setting.value\n ) > width else setting.value\n print('{}: {}'.format(setting.key, value))\n elif args.format == 'csv':\n try:\n writer = csv.writer(sys.stdout, quoting=csv.QUOTE_ALL)\n writer.writerow(['KEY', 'VALUE'])\n for setting in printable_settings:\n writer.writerow([setting.key, setting.value])\n except csv.Error:\n raise CliException('Error writing CSV')\n elif args.format == 'json' or args.format == 'yaml':\n settings_snapshot = {'head': head, 'settings': {setting.key:\n setting.value for setting in printable_settings}}\n if args.format == 'json':\n print(json.dumps(settings_snapshot, indent=2, sort_keys=True))\n else:\n print(yaml.dump(settings_snapshot, default_flow_style=False)[0:-1])\n else:\n raise AssertionError('Unknown format {}'.format(args.format))\n", "role": "user" }, { "content": "do_settings_list", "role": "assistant" } ]
def x(key): key_parts = key.split('.', maxsplit=_MAX_KEY_PARTS - 1) key_parts.extend([''] * (_MAX_KEY_PARTS - len(key_parts))) return SETTINGS_NAMESPACE + ''.join(_short_hash(x) for x in key_parts)
key_to_address
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(key):\n key_parts = key.split('.', maxsplit=_MAX_KEY_PARTS - 1)\n key_parts.extend([''] * (_MAX_KEY_PARTS - len(key_parts)))\n return SETTINGS_NAMESPACE + ''.join(_short_hash(x) for x in key_parts)\n", "role": "user" }, { "content": "key_to_address", "role": "assistant" } ]
def x(self, injector): if injector == 'block_info': block_info_injector = importlib.import_module( 'sawtooth_validator.journal.block_info_injector') return block_info_injector.BlockInfoInjector(self. _state_view_factory, self._signer) raise UnknownBatchInjectorError(injector)
create_injector
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, injector):\n if injector == 'block_info':\n block_info_injector = importlib.import_module(\n 'sawtooth_validator.journal.block_info_injector')\n return block_info_injector.BlockInfoInjector(self.\n _state_view_factory, self._signer)\n raise UnknownBatchInjectorError(injector)\n", "role": "user" }, { "content": "create_injector", "role": "assistant" } ]
def x(self, block_id): _libexec('chain_controller_on_block_received', self.pointer, ctypes. c_char_p(block_id.encode('utf-8')))
on_block_received
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, block_id):\n _libexec('chain_controller_on_block_received', self.pointer, ctypes.\n c_char_p(block_id.encode('utf-8')))\n", "role": "user" }, { "content": "on_block_received", "role": "assistant" } ]
def x(subparsers, parent_parser): parser = subparsers.add_parser('compare-chains', help= 'Compare chains from different nodes.', description= 'Compute and display information about how the chains at different nodes differ.' , formatter_class=argparse.RawDescriptionHelpFormatter, epilog= """ By default, prints a table of summary data and a table of per-node data with the following fields. Pass --tree for a fork graph. COMMON ANCESTOR The most recent block that all chains have in common. COMMON HEIGHT Let min_height := the minimum height of any chain across all nodes passed in. COMMON HEIGHT = min_height. HEAD The block id of the most recent block on a given chain. HEIGHT The block number of the most recent block on a given chain. LAG Let max_height := the maximum height of any chain across all nodes passed in. LAG = max_height - HEIGHT for a given chain. DIVERG Let common_ancestor_height := the height of the COMMON ANCESTOR. DIVERG = HEIGHT - common_ancestor_height """ , parents=[parent_parser, base_multinode_parser()]) parser.add_argument('-l', '--limit', default=25, type=int, help= 'the number of blocks to request at a time') parser.add_argument('--table', action='store_true', help= 'Print out a fork table for all nodes since the common ancestor.') parser.add_argument('--tree', action='store_true', help= 'Print out a fork tree for all nodes since the common ancestor.')
add_compare_chains_parser
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(subparsers, parent_parser):\n parser = subparsers.add_parser('compare-chains', help=\n 'Compare chains from different nodes.', description=\n 'Compute and display information about how the chains at different nodes differ.'\n , formatter_class=argparse.RawDescriptionHelpFormatter, epilog=\n \"\"\"\nBy default, prints a table of summary data and a table of per-node data with\nthe following fields. Pass --tree for a fork graph.\n\nCOMMON ANCESTOR\n The most recent block that all chains have in common.\n\nCOMMON HEIGHT\n Let min_height := the minimum height of any chain across all nodes passed\n in. COMMON HEIGHT = min_height.\n\nHEAD\n The block id of the most recent block on a given chain.\n\nHEIGHT\n The block number of the most recent block on a given chain.\n\nLAG\n Let max_height := the maximum height of any chain across all nodes passed\n in. LAG = max_height - HEIGHT for a given chain.\n\nDIVERG\n Let common_ancestor_height := the height of the COMMON ANCESTOR.\n DIVERG = HEIGHT - common_ancestor_height\n\n\"\"\"\n , parents=[parent_parser, base_multinode_parser()])\n parser.add_argument('-l', '--limit', default=25, type=int, help=\n 'the number of blocks to request at a time')\n parser.add_argument('--table', action='store_true', help=\n 'Print out a fork table for all nodes since the common ancestor.')\n parser.add_argument('--tree', action='store_true', help=\n 'Print out a fork tree for all nodes since the common ancestor.')\n", "role": "user" }, { "content": "add_compare_chains_parser", "role": "assistant" } ]
def x(args): urls = split_comma_append_args(args.urls) users = split_comma_append_args(args.users) clients = make_rest_apis(urls, users) broken = [] chains, errors = get_chain_generators(clients, args.limit) broken.extend(errors) for node in errors: print('Error connecting to node %d: %s' % (node, urls[node])) if not chains: print('No nodes reporting') return tails, errors = get_tails(chains) broken.extend(errors) for node in errors: del chains[node] for node in errors: print('Failed to reach common height with node %d: %s' % (node, urls[node])) if not chains: print('Failed to get common height') return graph, errors = build_fork_graph(chains, tails) broken.extend(errors) for node in errors: print('Failed to reach common ancestor with node %d: %s' % (node, urls[node])) if not graph: print('Failed to build fork graph') return broken.sort() node_id_map = get_node_id_map(broken, len(clients)) tails = list(map(lambda item: item[1], filter(lambda item: item[0] not in broken, sorted(tails.items())))) if args.table: print_table(graph, tails, node_id_map) elif args.tree: print_tree(graph, tails, node_id_map) else: print_summary(graph, tails, node_id_map)
do_compare_chains
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(args):\n urls = split_comma_append_args(args.urls)\n users = split_comma_append_args(args.users)\n clients = make_rest_apis(urls, users)\n broken = []\n chains, errors = get_chain_generators(clients, args.limit)\n broken.extend(errors)\n for node in errors:\n print('Error connecting to node %d: %s' % (node, urls[node]))\n if not chains:\n print('No nodes reporting')\n return\n tails, errors = get_tails(chains)\n broken.extend(errors)\n for node in errors:\n del chains[node]\n for node in errors:\n print('Failed to reach common height with node %d: %s' % (node,\n urls[node]))\n if not chains:\n print('Failed to get common height')\n return\n graph, errors = build_fork_graph(chains, tails)\n broken.extend(errors)\n for node in errors:\n print('Failed to reach common ancestor with node %d: %s' % (node,\n urls[node]))\n if not graph:\n print('Failed to build fork graph')\n return\n broken.sort()\n node_id_map = get_node_id_map(broken, len(clients))\n tails = list(map(lambda item: item[1], filter(lambda item: item[0] not in\n broken, sorted(tails.items()))))\n if args.table:\n print_table(graph, tails, node_id_map)\n elif args.tree:\n print_tree(graph, tails, node_id_map)\n else:\n print_summary(graph, tails, node_id_map)\n", "role": "user" }, { "content": "do_compare_chains", "role": "assistant" } ]
def x(graph, tails, node_id_map): heads = get_heads(tails) heights = get_heights(tails) max_height = max(heights) common_height, block_ids_at_common_height = get_common_height(tails) lags = get_lags(heights, max_height) common_ancestor = graph.root divergences = get_divergences(heights, graph.root) col_1 = 8 col_n = 8 format_str = '{:<' + str(col_1) + '} ' + ('{:<' + str(col_n) + '} ') * 2 header = format_str.format('COMMON', 'HEIGHT', 'BLOCKS') print(header) print('-' * len(header)) print(format_str.format('ANCESTOR', common_ancestor.num, common_ancestor.ident[:col_n])) print(format_str.format('HEIGHT', common_height, str( block_ids_at_common_height))) print() node_col_width = get_col_width_for_num(len(tails), len('NODE')) num_col_width = get_col_width_for_num(max_height, len('HEIGHT')) lag_col_width = get_col_width_for_num(max(lags), len('LAG')) diverg_col_width = get_col_width_for_num(max(divergences), len('DIVERG')) format_str = '{:<' + str(node_col_width) + '} {:<8} {:<' + str( num_col_width) + '} {:<' + str(lag_col_width) + '} {:<' + str( diverg_col_width) + '}' header = format_str.format('NODE', 'HEAD', 'HEIGHT', 'LAG', 'DIVERG') print(header) print('-' * len(header)) for i, _ in enumerate(tails): print(format_str.format(node_id_map[i], heads[i], heights[i], lags[ i], divergences[i])) print()
print_summary
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(graph, tails, node_id_map):\n heads = get_heads(tails)\n heights = get_heights(tails)\n max_height = max(heights)\n common_height, block_ids_at_common_height = get_common_height(tails)\n lags = get_lags(heights, max_height)\n common_ancestor = graph.root\n divergences = get_divergences(heights, graph.root)\n col_1 = 8\n col_n = 8\n format_str = '{:<' + str(col_1) + '} ' + ('{:<' + str(col_n) + '} ') * 2\n header = format_str.format('COMMON', 'HEIGHT', 'BLOCKS')\n print(header)\n print('-' * len(header))\n print(format_str.format('ANCESTOR', common_ancestor.num,\n common_ancestor.ident[:col_n]))\n print(format_str.format('HEIGHT', common_height, str(\n block_ids_at_common_height)))\n print()\n node_col_width = get_col_width_for_num(len(tails), len('NODE'))\n num_col_width = get_col_width_for_num(max_height, len('HEIGHT'))\n lag_col_width = get_col_width_for_num(max(lags), len('LAG'))\n diverg_col_width = get_col_width_for_num(max(divergences), len('DIVERG'))\n format_str = '{:<' + str(node_col_width) + '} {:<8} {:<' + str(\n num_col_width) + '} {:<' + str(lag_col_width) + '} {:<' + str(\n diverg_col_width) + '}'\n header = format_str.format('NODE', 'HEAD', 'HEIGHT', 'LAG', 'DIVERG')\n print(header)\n print('-' * len(header))\n for i, _ in enumerate(tails):\n print(format_str.format(node_id_map[i], heads[i], heights[i], lags[\n i], divergences[i]))\n print()\n", "role": "user" }, { "content": "print_summary", "role": "assistant" } ]
def x(graph, tails, node_id_map): node_count = len(tails) num_col_width = max(floor(log(max(get_heights(tails)), 10)) + 1, len('NUM') ) node_col_width = max(floor(log(node_count, 10)) + 1, 8) format_str = '' format_str += '{:<' + str(num_col_width) + '} ' for _ in range(node_count): format_str += '{:<' + str(node_col_width) + '} ' nodes_header = [('NODE ' + str(node_id_map[i])) for i in range(node_count)] header = format_str.format('NUM', *nodes_header) print(header) print('-' * len(header)) prev_block_num = -1 node_list = [''] * node_count for block_num, _, siblings in graph.walk(): if block_num != prev_block_num: if prev_block_num != -1: print(format_str.format(prev_block_num, *node_list)) node_list.clear() node_list.extend([''] * node_count) prev_block_num = block_num for block_id, node_ids in siblings.items(): for node_id in node_ids: node_list[node_id] = block_id[:8] print(format_str.format(prev_block_num, *node_list))
print_table
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(graph, tails, node_id_map):\n node_count = len(tails)\n num_col_width = max(floor(log(max(get_heights(tails)), 10)) + 1, len('NUM')\n )\n node_col_width = max(floor(log(node_count, 10)) + 1, 8)\n format_str = ''\n format_str += '{:<' + str(num_col_width) + '} '\n for _ in range(node_count):\n format_str += '{:<' + str(node_col_width) + '} '\n nodes_header = [('NODE ' + str(node_id_map[i])) for i in range(node_count)]\n header = format_str.format('NUM', *nodes_header)\n print(header)\n print('-' * len(header))\n prev_block_num = -1\n node_list = [''] * node_count\n for block_num, _, siblings in graph.walk():\n if block_num != prev_block_num:\n if prev_block_num != -1:\n print(format_str.format(prev_block_num, *node_list))\n node_list.clear()\n node_list.extend([''] * node_count)\n prev_block_num = block_num\n for block_id, node_ids in siblings.items():\n for node_id in node_ids:\n node_list[node_id] = block_id[:8]\n print(format_str.format(prev_block_num, *node_list))\n", "role": "user" }, { "content": "print_table", "role": "assistant" } ]
def x(graph, tails, node_id_map): walker = graph.walk() next_block_num, next_parent, next_siblings = next(walker) prev_cliques = [] done = False while not done: cliques = {} block_num = next_block_num try: while block_num == next_block_num: cliques[next_parent] = next_siblings next_block_num, next_parent, next_siblings = next(walker) except StopIteration: done = True print_cliques(prev_cliques, cliques, node_id_map) print_block_num_row(block_num, prev_cliques, cliques) print_splits(prev_cliques, cliques) print_folds(prev_cliques, cliques) prev_cliques = build_ordered_cliques(prev_cliques, cliques) print_cliques(prev_cliques, [], node_id_map)
print_tree
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(graph, tails, node_id_map):\n walker = graph.walk()\n next_block_num, next_parent, next_siblings = next(walker)\n prev_cliques = []\n done = False\n while not done:\n cliques = {}\n block_num = next_block_num\n try:\n while block_num == next_block_num:\n cliques[next_parent] = next_siblings\n next_block_num, next_parent, next_siblings = next(walker)\n except StopIteration:\n done = True\n print_cliques(prev_cliques, cliques, node_id_map)\n print_block_num_row(block_num, prev_cliques, cliques)\n print_splits(prev_cliques, cliques)\n print_folds(prev_cliques, cliques)\n prev_cliques = build_ordered_cliques(prev_cliques, cliques)\n print_cliques(prev_cliques, [], node_id_map)\n", "role": "user" }, { "content": "print_tree", "role": "assistant" } ]
def x(cliques, next_cliques): def x(clique): return -len(clique[1]) if not cliques: return list(sorted(list(next_cliques.values())[0].items(), key= sort_key)) ordered_cliques = [] for _, clique in enumerate(cliques): parent, _ = clique if parent in next_cliques: ordered_cliques.extend(sorted(next_cliques[parent].items(), key =sort_key)) return ordered_cliques
build_ordered_cliques
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cliques, next_cliques):\n\n def x(clique):\n return -len(clique[1])\n if not cliques:\n return list(sorted(list(next_cliques.values())[0].items(), key=\n sort_key))\n ordered_cliques = []\n for _, clique in enumerate(cliques):\n parent, _ = clique\n if parent in next_cliques:\n ordered_cliques.extend(sorted(next_cliques[parent].items(), key\n =sort_key))\n return ordered_cliques\n", "role": "user" }, { "content": "build_ordered_cliques", "role": "assistant" } ]
def x(column_to_fold, total_columns, skips): format_str = '{:<2}' * (total_columns - 1) cols = [] for i in range(column_to_fold): if i in skips: cols.append(' ') else: cols.append('| ') for i in range(column_to_fold + 1, total_columns): if i in skips: cols.append(' ') else: cols.append(' /') print(format_str.format(*cols))
print_fold
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(column_to_fold, total_columns, skips):\n format_str = '{:<2}' * (total_columns - 1)\n cols = []\n for i in range(column_to_fold):\n if i in skips:\n cols.append(' ')\n else:\n cols.append('| ')\n for i in range(column_to_fold + 1, total_columns):\n if i in skips:\n cols.append(' ')\n else:\n cols.append(' /')\n print(format_str.format(*cols))\n", "role": "user" }, { "content": "print_fold", "role": "assistant" } ]
def x(block_num, cliques, next_cliques): n_cliques = len(cliques) if n_cliques == 0: print('| {}'.format(block_num)) return def x(clique): block_id, _ = clique if block_id not in next_cliques: return ' ' return '|' format_str = '{:<' + str(n_cliques * 2) + '} {}' branches = list(map(mapper, cliques)) for end in ('', block_num): print(format_str.format(' '.join(branches), end))
print_block_num_row
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(block_num, cliques, next_cliques):\n n_cliques = len(cliques)\n if n_cliques == 0:\n print('| {}'.format(block_num))\n return\n\n def x(clique):\n block_id, _ = clique\n if block_id not in next_cliques:\n return ' '\n return '|'\n format_str = '{:<' + str(n_cliques * 2) + '} {}'\n branches = list(map(mapper, cliques))\n for end in ('', block_num):\n print(format_str.format(' '.join(branches), end))\n", "role": "user" }, { "content": "print_block_num_row", "role": "assistant" } ]
def x(cliques, next_cliques, node_id_map): n_cliques = len(cliques) format_str = '{:<' + str(n_cliques * 2) + '} {} {}' branches = ['|'] * len(cliques) for i, clique in enumerate(cliques): block_id, nodes = clique print(format_str.format(' '.join(branches[:i] + ['*'] + branches[i + 1:]), block_id[:8], format_siblings(nodes, node_id_map))) if block_id not in next_cliques: branches[i] = ' '
print_cliques
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cliques, next_cliques, node_id_map):\n n_cliques = len(cliques)\n format_str = '{:<' + str(n_cliques * 2) + '} {} {}'\n branches = ['|'] * len(cliques)\n for i, clique in enumerate(cliques):\n block_id, nodes = clique\n print(format_str.format(' '.join(branches[:i] + ['*'] + branches[i +\n 1:]), block_id[:8], format_siblings(nodes, node_id_map)))\n if block_id not in next_cliques:\n branches[i] = ' '\n", "role": "user" }, { "content": "print_cliques", "role": "assistant" } ]
def x(cliques, next_cliques): splits = 0 for i, clique in enumerate(cliques): parent, _ = clique if parent in next_cliques: if len(next_cliques[parent]) > 1: print_split(i + splits, len(cliques) + splits) splits += 1
print_splits
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cliques, next_cliques):\n splits = 0\n for i, clique in enumerate(cliques):\n parent, _ = clique\n if parent in next_cliques:\n if len(next_cliques[parent]) > 1:\n print_split(i + splits, len(cliques) + splits)\n splits += 1\n", "role": "user" }, { "content": "print_splits", "role": "assistant" } ]
def x(column_to_split, total_columns): out = '' for _ in range(column_to_split): out += '| ' out += '|\\' for _ in range(column_to_split + 1, total_columns): out += ' \\' print(out)
print_split
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(column_to_split, total_columns):\n out = ''\n for _ in range(column_to_split):\n out += '| '\n out += '|\\\\'\n for _ in range(column_to_split + 1, total_columns):\n out += ' \\\\'\n print(out)\n", "role": "user" }, { "content": "print_split", "role": "assistant" } ]
def x(chains): def x(blocks): return blocks[0].num tails = {} bad_chains = [] for i, chain in chains.items(): try: tails[i] = [next(chain)] except StopIteration: bad_chains.append(i) min_block_num = min(map(get_num_of_oldest, tails.values())) for i, chain in chains.items(): if i not in bad_chains: tail = tails[i] while get_num_of_oldest(tail) > min_block_num: try: block = next(chain) except StopIteration: bad_chains.append(i) break tail.insert(0, block) return tails, bad_chains
get_tails
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(chains):\n\n def x(blocks):\n return blocks[0].num\n tails = {}\n bad_chains = []\n for i, chain in chains.items():\n try:\n tails[i] = [next(chain)]\n except StopIteration:\n bad_chains.append(i)\n min_block_num = min(map(get_num_of_oldest, tails.values()))\n for i, chain in chains.items():\n if i not in bad_chains:\n tail = tails[i]\n while get_num_of_oldest(tail) > min_block_num:\n try:\n block = next(chain)\n except StopIteration:\n bad_chains.append(i)\n break\n tail.insert(0, block)\n return tails, bad_chains\n", "role": "user" }, { "content": "get_tails", "role": "assistant" } ]
def x(collections, key): if len(collections) < 2: return True c0 = key(collections[0]) return all(c0 == key(c) for c in collections[1:])
compare_across
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(collections, key):\n if len(collections) < 2:\n return True\n c0 = key(collections[0])\n return all(c0 == key(c) for c in collections[1:])\n", "role": "user" }, { "content": "compare_across", "role": "assistant" } ]
def x(chains, tails): graph = ForkGraph() bad_chains = [] for i, tail in tails.items(): for block in reversed(tail): graph.add_block(i, block) if _compare_across([tail[0] for tail in tails.values()], key=lambda block: block.ident): return graph, bad_chains while True: heads = [] for i, chain in chains.items(): if i not in bad_chains: try: head = next(chain) except StopIteration: bad_chains.append(i) heads.append((i, head)) for i, block in heads: graph.add_block(i, block) if _compare_across(heads, key=lambda head: head[1].ident): break prune_unreporting_peers(graph, bad_chains) return graph, bad_chains
build_fork_graph
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(chains, tails):\n graph = ForkGraph()\n bad_chains = []\n for i, tail in tails.items():\n for block in reversed(tail):\n graph.add_block(i, block)\n if _compare_across([tail[0] for tail in tails.values()], key=lambda\n block: block.ident):\n return graph, bad_chains\n while True:\n heads = []\n for i, chain in chains.items():\n if i not in bad_chains:\n try:\n head = next(chain)\n except StopIteration:\n bad_chains.append(i)\n heads.append((i, head))\n for i, block in heads:\n graph.add_block(i, block)\n if _compare_across(heads, key=lambda head: head[1].ident):\n break\n prune_unreporting_peers(graph, bad_chains)\n return graph, bad_chains\n", "role": "user" }, { "content": "build_fork_graph", "role": "assistant" } ]
def x(self, identifier, instance=None, tags=None): tag_list = [] if tags is not None: tag_list.extend(tags.items()) tag_list.extend(self._base_tags) return '.'.join(identifier) + ',' + ','.join('{}={}'.format(k, v) for k, v in tag_list)
join
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, identifier, instance=None, tags=None):\n tag_list = []\n if tags is not None:\n tag_list.extend(tags.items())\n tag_list.extend(self._base_tags)\n return '.'.join(identifier) + ',' + ','.join('{}={}'.format(k, v) for k,\n v in tag_list)\n", "role": "user" }, { "content": "join", "role": "assistant" } ]
def x(subparsers, parent_parser): parser = subparsers.add_parser('config', help= 'Changes genesis block settings and create, view, and vote on settings proposals' , description= 'Provides subcommands to change genesis block settings and to view, create, and vote on existing proposals.' ) config_parsers = parser.add_subparsers(title='subcommands', dest= 'subcommand') config_parsers.required = True
add_config_parser
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(subparsers, parent_parser):\n parser = subparsers.add_parser('config', help=\n 'Changes genesis block settings and create, view, and vote on settings proposals'\n , description=\n 'Provides subcommands to change genesis block settings and to view, create, and vote on existing proposals.'\n )\n config_parsers = parser.add_subparsers(title='subcommands', dest=\n 'subcommand')\n config_parsers.required = True\n", "role": "user" }, { "content": "add_config_parser", "role": "assistant" } ]
def x(args): settings = [s.split('=', 1) for s in args.setting] signer = _read_signer(args.key) txns = [_create_propose_txn(signer, setting) for setting in settings] batch = _create_batch(signer, txns) batch_list = BatchList(batches=[batch]) if args.output is not None: try: with open(args.output, 'wb') as batch_file: batch_file.write(batch_list.SerializeToString()) except IOError as e: raise CliException('Unable to write to batch file: {}'.format( str(e))) elif args.sabre_output is not None: for i, txn in enumerate(txns): with open('{}-{}'.format(args.sabre_output, i), 'wb') as outfile: outfile.write(txn.payload) elif args.url is not None: rest_client = RestClient(args.url) rest_client.send_batches(batch_list) else: raise AssertionError('No target for create set.')
do_config_proposal_create
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(args):\n settings = [s.split('=', 1) for s in args.setting]\n signer = _read_signer(args.key)\n txns = [_create_propose_txn(signer, setting) for setting in settings]\n batch = _create_batch(signer, txns)\n batch_list = BatchList(batches=[batch])\n if args.output is not None:\n try:\n with open(args.output, 'wb') as batch_file:\n batch_file.write(batch_list.SerializeToString())\n except IOError as e:\n raise CliException('Unable to write to batch file: {}'.format(\n str(e)))\n elif args.sabre_output is not None:\n for i, txn in enumerate(txns):\n with open('{}-{}'.format(args.sabre_output, i), 'wb') as outfile:\n outfile.write(txn.payload)\n elif args.url is not None:\n rest_client = RestClient(args.url)\n rest_client.send_batches(batch_list)\n else:\n raise AssertionError('No target for create set.')\n", "role": "user" }, { "content": "do_config_proposal_create", "role": "assistant" } ]
def x(args): def x(candidate, public_key, prefix): has_pub_key = not public_key or candidate.votes[0 ].public_key == public_key has_prefix = candidate.proposal.setting.startswith(prefix) return has_prefix and has_pub_key candidates_payload = _get_proposals(RestClient(args.url)) candidates = [c for c in candidates_payload.candidates if _accept(c, args.public_key, args.filter)] if args.format == 'default': for candidate in candidates: print('{}: {} => {}'.format(candidate.proposal_id, candidate. proposal.setting, candidate.proposal.value)) elif args.format == 'csv': writer = csv.writer(sys.stdout, quoting=csv.QUOTE_ALL) writer.writerow(['PROPOSAL_ID', 'KEY', 'VALUE']) for candidate in candidates: writer.writerow([candidate.proposal_id, candidate.proposal. setting, candidate.proposal.value]) elif args.format == 'json' or args.format == 'yaml': candidates_snapshot = {c.proposal_id: {c.proposal.setting: c. proposal.value} for c in candidates} if args.format == 'json': print(json.dumps(candidates_snapshot, indent=2, sort_keys=True)) else: print(yaml.dump(candidates_snapshot, default_flow_style=False)[ 0:-1]) else: raise AssertionError('Unknown format {}'.format(args.format))
do_config_proposal_list
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(args):\n\n def x(candidate, public_key, prefix):\n has_pub_key = not public_key or candidate.votes[0\n ].public_key == public_key\n has_prefix = candidate.proposal.setting.startswith(prefix)\n return has_prefix and has_pub_key\n candidates_payload = _get_proposals(RestClient(args.url))\n candidates = [c for c in candidates_payload.candidates if _accept(c,\n args.public_key, args.filter)]\n if args.format == 'default':\n for candidate in candidates:\n print('{}: {} => {}'.format(candidate.proposal_id, candidate.\n proposal.setting, candidate.proposal.value))\n elif args.format == 'csv':\n writer = csv.writer(sys.stdout, quoting=csv.QUOTE_ALL)\n writer.writerow(['PROPOSAL_ID', 'KEY', 'VALUE'])\n for candidate in candidates:\n writer.writerow([candidate.proposal_id, candidate.proposal.\n setting, candidate.proposal.value])\n elif args.format == 'json' or args.format == 'yaml':\n candidates_snapshot = {c.proposal_id: {c.proposal.setting: c.\n proposal.value} for c in candidates}\n if args.format == 'json':\n print(json.dumps(candidates_snapshot, indent=2, sort_keys=True))\n else:\n print(yaml.dump(candidates_snapshot, default_flow_style=False)[\n 0:-1])\n else:\n raise AssertionError('Unknown format {}'.format(args.format))\n", "role": "user" }, { "content": "do_config_proposal_list", "role": "assistant" } ]
def x(args): signer = _read_signer(args.key) rest_client = RestClient(args.url) proposals = _get_proposals(rest_client) proposal = None for candidate in proposals.candidates: if candidate.proposal_id == args.proposal_id: proposal = candidate break if proposal is None: raise CliException('No proposal exists with the given id') for vote_record in proposal.votes: if vote_record.public_key == signer.get_public_key().as_hex(): raise CliException( 'A vote has already been recorded with this signing key') txn = _create_vote_txn(signer, args.proposal_id, proposal.proposal. setting, args.vote_value) batch = _create_batch(signer, [txn]) batch_list = BatchList(batches=[batch]) rest_client.send_batches(batch_list)
do_config_proposal_vote
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(args):\n signer = _read_signer(args.key)\n rest_client = RestClient(args.url)\n proposals = _get_proposals(rest_client)\n proposal = None\n for candidate in proposals.candidates:\n if candidate.proposal_id == args.proposal_id:\n proposal = candidate\n break\n if proposal is None:\n raise CliException('No proposal exists with the given id')\n for vote_record in proposal.votes:\n if vote_record.public_key == signer.get_public_key().as_hex():\n raise CliException(\n 'A vote has already been recorded with this signing key')\n txn = _create_vote_txn(signer, args.proposal_id, proposal.proposal.\n setting, args.vote_value)\n batch = _create_batch(signer, [txn])\n batch_list = BatchList(batches=[batch])\n rest_client.send_batches(batch_list)\n", "role": "user" }, { "content": "do_config_proposal_vote", "role": "assistant" } ]
def x(signer, setting_key_value): setting_key, setting_value = setting_key_value nonce = hex(random.randint(0, 2 ** 64)) proposal = SettingProposal(setting=setting_key, value=setting_value, nonce=nonce) payload = SettingsPayload(data=proposal.SerializeToString(), action= SettingsPayload.PROPOSE) return _make_txn(signer, setting_key, payload)
create_propose_txn
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(signer, setting_key_value):\n setting_key, setting_value = setting_key_value\n nonce = hex(random.randint(0, 2 ** 64))\n proposal = SettingProposal(setting=setting_key, value=setting_value,\n nonce=nonce)\n payload = SettingsPayload(data=proposal.SerializeToString(), action=\n SettingsPayload.PROPOSE)\n return _make_txn(signer, setting_key, payload)\n", "role": "user" }, { "content": "create_propose_txn", "role": "assistant" } ]
def x(signer, proposal_id, setting_key, vote_value): if vote_value == 'accept': vote_id = SettingVote.ACCEPT else: vote_id = SettingVote.REJECT vote = SettingVote(proposal_id=proposal_id, vote=vote_id) payload = SettingsPayload(data=vote.SerializeToString(), action= SettingsPayload.VOTE) return _make_txn(signer, setting_key, payload)
create_vote_txn
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(signer, proposal_id, setting_key, vote_value):\n if vote_value == 'accept':\n vote_id = SettingVote.ACCEPT\n else:\n vote_id = SettingVote.REJECT\n vote = SettingVote(proposal_id=proposal_id, vote=vote_id)\n payload = SettingsPayload(data=vote.SerializeToString(), action=\n SettingsPayload.VOTE)\n return _make_txn(signer, setting_key, payload)\n", "role": "user" }, { "content": "create_vote_txn", "role": "assistant" } ]
def x(signer, setting_key, payload): serialized_payload = payload.SerializeToString() header = TransactionHeader(signer_public_key=signer.get_public_key(). as_hex(), family_name='sawtooth_settings', family_version='1.0', inputs=_config_inputs(setting_key), outputs=_config_outputs( setting_key), dependencies=[], payload_sha512=hashlib.sha512( serialized_payload).hexdigest(), batcher_public_key=signer. get_public_key().as_hex()).SerializeToString() return Transaction(header=header, header_signature=signer.sign(header), payload=serialized_payload)
make_txn
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(signer, setting_key, payload):\n serialized_payload = payload.SerializeToString()\n header = TransactionHeader(signer_public_key=signer.get_public_key().\n as_hex(), family_name='sawtooth_settings', family_version='1.0',\n inputs=_config_inputs(setting_key), outputs=_config_outputs(\n setting_key), dependencies=[], payload_sha512=hashlib.sha512(\n serialized_payload).hexdigest(), batcher_public_key=signer.\n get_public_key().as_hex()).SerializeToString()\n return Transaction(header=header, header_signature=signer.sign(header),\n payload=serialized_payload)\n", "role": "user" }, { "content": "make_txn", "role": "assistant" } ]
def x(self, message, sender_id): payload = message.SerializeToString() self._notify('consensus_notifier_notify_peer_message', payload, len( payload), sender_id, len(sender_id))
notify_peer_message
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, message, sender_id):\n payload = message.SerializeToString()\n self._notify('consensus_notifier_notify_peer_message', payload, len(\n payload), sender_id, len(sender_id))\n", "role": "user" }, { "content": "notify_peer_message", "role": "assistant" } ]
def x(self, block): payload = block.SerializeToString() self._notify('consensus_notifier_notify_block_new', payload, len(payload))
notify_block_new
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, block):\n payload = block.SerializeToString()\n self._notify('consensus_notifier_notify_block_new', payload, len(payload))\n", "role": "user" }, { "content": "notify_block_new", "role": "assistant" } ]
def x(self, chain_head): chain_head_bytes = chain_head.SerializeToString() self._notify('consensus_notifier_notify_engine_activated', chain_head_bytes, len(chain_head_bytes))
notify_engine_activated
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, chain_head):\n chain_head_bytes = chain_head.SerializeToString()\n self._notify('consensus_notifier_notify_engine_activated',\n chain_head_bytes, len(chain_head_bytes))\n", "role": "user" }, { "content": "notify_engine_activated", "role": "assistant" } ]
def x(self, address, updater, prune=False): node = self._get_or_create(address) node.data = updater(node.data) if prune: node.children.clear()
update
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, address, updater, prune=False):\n node = self._get_or_create(address)\n node.data = updater(node.data)\n if prune:\n node.children.clear()\n", "role": "user" }, { "content": "update", "role": "assistant" } ]
def x(self, address): try: for step in self._walk_to_address(address): node = step except AddressNotInTree: return node.children.clear()
prune
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, address):\n try:\n for step in self._walk_to_address(address):\n node = step\n except AddressNotInTree:\n return\n node.children.clear()\n", "role": "user" }, { "content": "prune", "role": "assistant" } ]
def x(self, address): for step in self._walk_to_address(address): node = step yield node.address, node.data to_process = deque() to_process.extendleft(node.children) while to_process: node = to_process.pop() yield node.address, node.data if node.children: to_process.extendleft(node.children)
walk
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, address):\n for step in self._walk_to_address(address):\n node = step\n yield node.address, node.data\n to_process = deque()\n to_process.extendleft(node.children)\n while to_process:\n node = to_process.pop()\n yield node.address, node.data\n if node.children:\n to_process.extendleft(node.children)\n", "role": "user" }, { "content": "walk", "role": "assistant" } ]
def x(self, address): predecessors = set() enclosing_writer = None node_stream = self._tree.walk(address) address_len = len(address) try: for node_address, node in node_stream: if node is not None: predecessors.update(node.readers) if node.writer is not None: enclosing_writer = node.writer if len(node_address) >= address_len: break except AddressNotInTree as err: if err.match is not None: return self.find_write_predecessors(err.match) return predecessors finally: if enclosing_writer is not None: predecessors.add(enclosing_writer) for _, node in node_stream: if node is not None: if node.writer is not None: predecessors.add(node.writer) predecessors.update(node.readers) return predecessors
find_write_predecessors
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, address):\n predecessors = set()\n enclosing_writer = None\n node_stream = self._tree.walk(address)\n address_len = len(address)\n try:\n for node_address, node in node_stream:\n if node is not None:\n predecessors.update(node.readers)\n if node.writer is not None:\n enclosing_writer = node.writer\n if len(node_address) >= address_len:\n break\n except AddressNotInTree as err:\n if err.match is not None:\n return self.find_write_predecessors(err.match)\n return predecessors\n finally:\n if enclosing_writer is not None:\n predecessors.add(enclosing_writer)\n for _, node in node_stream:\n if node is not None:\n if node.writer is not None:\n predecessors.add(node.writer)\n predecessors.update(node.readers)\n return predecessors\n", "role": "user" }, { "content": "find_write_predecessors", "role": "assistant" } ]
def x(self, txn_id, predecessors): all_pred = set(predecessors) for pred in predecessors: all_pred.update(self._predecessors_by_id[pred]) self._predecessors_by_id[txn_id] = all_pred
add_relationship
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, txn_id, predecessors):\n all_pred = set(predecessors)\n for pred in predecessors:\n all_pred.update(self._predecessors_by_id[pred])\n self._predecessors_by_id[txn_id] = all_pred\n", "role": "user" }, { "content": "add_relationship", "role": "assistant" } ]
def x(self, predecessor, others): return any(predecessor in self._predecessors_by_id[o] for o in others)
is_predecessor_of_other
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, predecessor, others):\n return any(predecessor in self._predecessors_by_id[o] for o in others)\n", "role": "user" }, { "content": "is_predecessor_of_other", "role": "assistant" } ]
def x(self, inputs): dependencies = [] for address in inputs: dependencies.extend(self._predecessor_tree.find_read_predecessors( address)) return dependencies
find_input_dependencies
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, inputs):\n dependencies = []\n for address in inputs:\n dependencies.extend(self._predecessor_tree.find_read_predecessors(\n address))\n return dependencies\n", "role": "user" }, { "content": "find_input_dependencies", "role": "assistant" } ]
def x(self, outputs): dependencies = [] for address in outputs: dependencies.extend(self._predecessor_tree.find_write_predecessors( address)) return dependencies
find_output_dependencies
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, outputs):\n dependencies = []\n for address in outputs:\n dependencies.extend(self._predecessor_tree.find_write_predecessors(\n address))\n return dependencies\n", "role": "user" }, { "content": "find_output_dependencies", "role": "assistant" } ]
def x(self, batch_signature): batch = self._batches_by_id[batch_signature].batch index = self._batches.index(batch) contexts = [] txns_added_predecessors = [] for b in self._batches[index::-1]: batch_is_valid = True contexts_from_batch = [] for txn in b.transactions[::-1]: result = self._txn_results[txn.header_signature] if not result.is_valid: batch_is_valid = False break else: txn_id = txn.header_signature if txn_id not in txns_added_predecessors: txns_added_predecessors.append(self._txn_predecessors[ txn_id]) contexts_from_batch.append(result.context_id) if batch_is_valid: contexts.extend(contexts_from_batch) return contexts
get_contexts_for_squash
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, batch_signature):\n batch = self._batches_by_id[batch_signature].batch\n index = self._batches.index(batch)\n contexts = []\n txns_added_predecessors = []\n for b in self._batches[index::-1]:\n batch_is_valid = True\n contexts_from_batch = []\n for txn in b.transactions[::-1]:\n result = self._txn_results[txn.header_signature]\n if not result.is_valid:\n batch_is_valid = False\n break\n else:\n txn_id = txn.header_signature\n if txn_id not in txns_added_predecessors:\n txns_added_predecessors.append(self._txn_predecessors[\n txn_id])\n contexts_from_batch.append(result.context_id)\n if batch_is_valid:\n contexts.extend(contexts_from_batch)\n return contexts\n", "role": "user" }, { "content": "get_contexts_for_squash", "role": "assistant" } ]
def x(self, txn_id, possible_successor, already_seen): is_successor = self._is_predecessor_of_possible_successor(txn_id, possible_successor) in_different_batch = not self._is_in_same_batch(txn_id, possible_successor) has_not_been_seen = possible_successor not in already_seen return is_successor and in_different_batch and has_not_been_seen
is_txn_to_replay
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, txn_id, possible_successor, already_seen):\n is_successor = self._is_predecessor_of_possible_successor(txn_id,\n possible_successor)\n in_different_batch = not self._is_in_same_batch(txn_id, possible_successor)\n has_not_been_seen = possible_successor not in already_seen\n return is_successor and in_different_batch and has_not_been_seen\n", "role": "user" }, { "content": "is_txn_to_replay", "role": "assistant" } ]
def x(self, sig): batch = self._batches_by_txn_id[sig] seen = [] for txn in batch.transactions: txn_id = txn.header_signature for poss_successor in self._scheduled.copy(): if not self.is_transaction_in_schedule(poss_successor): continue if self._is_txn_to_replay(txn_id, poss_successor, seen): if self._txn_has_result(poss_successor): del self._txn_results[poss_successor] self._scheduled.remove(poss_successor) self._txns_available[poss_successor] = self._transactions[ poss_successor] else: self._outstanding.add(poss_successor) seen.append(poss_successor)
remove_subsequent_result_because_of_batch_failure
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, sig):\n batch = self._batches_by_txn_id[sig]\n seen = []\n for txn in batch.transactions:\n txn_id = txn.header_signature\n for poss_successor in self._scheduled.copy():\n if not self.is_transaction_in_schedule(poss_successor):\n continue\n if self._is_txn_to_replay(txn_id, poss_successor, seen):\n if self._txn_has_result(poss_successor):\n del self._txn_results[poss_successor]\n self._scheduled.remove(poss_successor)\n self._txns_available[poss_successor] = self._transactions[\n poss_successor]\n else:\n self._outstanding.add(poss_successor)\n seen.append(poss_successor)\n", "role": "user" }, { "content": "remove_subsequent_result_because_of_batch_failure", "role": "assistant" } ]
def x(self, txn_signature): batch = self._batches_by_txn_id[txn_signature] least_index = self._index_of_batch(self._batches_by_id[self. _least_batch_id_wo_results].batch) current_index = self._index_of_batch(batch) all_prior = False if current_index <= least_index: return if all(all(t.header_signature in self._txn_results for t in b. transactions) for b in self._batches[least_index:current_index]): all_prior = True if not all_prior: return possible_least = self._batches[current_index].header_signature for b in self._batches[current_index:]: if not all(t.header_signature in self._txn_results for t in b. transactions): possible_least = b.header_signature break self._least_batch_id_wo_results = possible_least
set_least_batch_id
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, txn_signature):\n batch = self._batches_by_txn_id[txn_signature]\n least_index = self._index_of_batch(self._batches_by_id[self.\n _least_batch_id_wo_results].batch)\n current_index = self._index_of_batch(batch)\n all_prior = False\n if current_index <= least_index:\n return\n if all(all(t.header_signature in self._txn_results for t in b.\n transactions) for b in self._batches[least_index:current_index]):\n all_prior = True\n if not all_prior:\n return\n possible_least = self._batches[current_index].header_signature\n for b in self._batches[current_index:]:\n if not all(t.header_signature in self._txn_results for t in b.\n transactions):\n possible_least = b.header_signature\n break\n self._least_batch_id_wo_results = possible_least\n", "role": "user" }, { "content": "set_least_batch_id", "role": "assistant" } ]
def x(self, txn_id): batch = self._batches_by_txn_id[txn_id] return all(self._txn_results[sig].is_valid for sig in set(self. _txn_results).intersection(txn.header_signature for txn in batch. transactions))
txn_is_in_valid_batch
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, txn_id):\n batch = self._batches_by_txn_id[txn_id]\n return all(self._txn_results[sig].is_valid for sig in set(self.\n _txn_results).intersection(txn.header_signature for txn in batch.\n transactions))\n", "role": "user" }, { "content": "txn_is_in_valid_batch", "role": "assistant" } ]
def x(args): if args.key_name is not None: key_name = args.key_name else: key_name = 'validator' key_dir = get_key_dir() if not os.path.exists(key_dir): raise CliException('Key directory does not exist: {}'.format(key_dir)) priv_filename = os.path.join(key_dir, key_name + '.priv') pub_filename = os.path.join(key_dir, key_name + '.pub') if not args.force: file_exists = False for filename in [priv_filename, pub_filename]: if os.path.exists(filename): file_exists = True print('file exists: {}'.format(filename), file=sys.stderr) if file_exists: raise CliException( 'files exist, rerun with --force to overwrite existing files') context = create_context('secp256k1') private_key = context.new_random_private_key() public_key = context.get_public_key(private_key) try: priv_exists = os.path.exists(priv_filename) with open(priv_filename, 'w') as priv_fd: if not args.quiet: if priv_exists: print('overwriting file: {}'.format(priv_filename)) else: print('writing file: {}'.format(priv_filename)) priv_fd.write(private_key.as_hex()) priv_fd.write('\n') keydir_info = os.stat(key_dir) keydir_gid = keydir_info.st_gid keydir_uid = keydir_info.st_uid os.chown(priv_filename, keydir_uid, keydir_gid) os.chmod(priv_filename, 416) pub_exists = os.path.exists(pub_filename) with open(pub_filename, 'w') as pub_fd: if not args.quiet: if pub_exists: print('overwriting file: {}'.format(pub_filename)) else: print('writing file: {}'.format(pub_filename)) pub_fd.write(public_key.as_hex()) pub_fd.write('\n') os.chown(pub_filename, keydir_uid, keydir_gid) os.chmod(pub_filename, 420) except IOError as ioe: raise CliException('IOError: {}'.format(str(ioe)))
do_keygen
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(args):\n if args.key_name is not None:\n key_name = args.key_name\n else:\n key_name = 'validator'\n key_dir = get_key_dir()\n if not os.path.exists(key_dir):\n raise CliException('Key directory does not exist: {}'.format(key_dir))\n priv_filename = os.path.join(key_dir, key_name + '.priv')\n pub_filename = os.path.join(key_dir, key_name + '.pub')\n if not args.force:\n file_exists = False\n for filename in [priv_filename, pub_filename]:\n if os.path.exists(filename):\n file_exists = True\n print('file exists: {}'.format(filename), file=sys.stderr)\n if file_exists:\n raise CliException(\n 'files exist, rerun with --force to overwrite existing files')\n context = create_context('secp256k1')\n private_key = context.new_random_private_key()\n public_key = context.get_public_key(private_key)\n try:\n priv_exists = os.path.exists(priv_filename)\n with open(priv_filename, 'w') as priv_fd:\n if not args.quiet:\n if priv_exists:\n print('overwriting file: {}'.format(priv_filename))\n else:\n print('writing file: {}'.format(priv_filename))\n priv_fd.write(private_key.as_hex())\n priv_fd.write('\\n')\n keydir_info = os.stat(key_dir)\n keydir_gid = keydir_info.st_gid\n keydir_uid = keydir_info.st_uid\n os.chown(priv_filename, keydir_uid, keydir_gid)\n os.chmod(priv_filename, 416)\n pub_exists = os.path.exists(pub_filename)\n with open(pub_filename, 'w') as pub_fd:\n if not args.quiet:\n if pub_exists:\n print('overwriting file: {}'.format(pub_filename))\n else:\n print('writing file: {}'.format(pub_filename))\n pub_fd.write(public_key.as_hex())\n pub_fd.write('\\n')\n os.chown(pub_filename, keydir_uid, keydir_gid)\n os.chmod(pub_filename, 420)\n except IOError as ioe:\n raise CliException('IOError: {}'.format(str(ioe)))\n", "role": "user" }, { "content": "do_keygen", "role": "assistant" } ]
def x(): try: assert os != 'nt' and sys.stdout.isatty() rows, columns = os.popen('stty size', 'r').read().split() except (AssertionError, AttributeError, ValueError): rows, columns = DEFAULT_HEIGHT, DEFAULT_WIDTH return int(rows), int(columns)
size
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n try:\n assert os != 'nt' and sys.stdout.isatty()\n rows, columns = os.popen('stty size', 'r').read().split()\n except (AssertionError, AttributeError, ValueError):\n rows, columns = DEFAULT_HEIGHT, DEFAULT_WIDTH\n return int(rows), int(columns)\n", "role": "user" }, { "content": "size", "role": "assistant" } ]
def x(self, connection_id): with self._lock: peer_endpoints = list(self._peers.values()) if self._endpoint: peer_endpoints.append(self._endpoint) peers_response = GetPeersResponse(peer_endpoints=peer_endpoints) try: self._network.send(validator_pb2.Message. GOSSIP_GET_PEERS_RESPONSE, peers_response.SerializeToString (), connection_id, one_way=True) except ValueError: LOGGER.debug('Connection disconnected: %s', connection_id)
send_peers
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, connection_id):\n with self._lock:\n peer_endpoints = list(self._peers.values())\n if self._endpoint:\n peer_endpoints.append(self._endpoint)\n peers_response = GetPeersResponse(peer_endpoints=peer_endpoints)\n try:\n self._network.send(validator_pb2.Message.\n GOSSIP_GET_PEERS_RESPONSE, peers_response.SerializeToString\n (), connection_id, one_way=True)\n except ValueError:\n LOGGER.debug('Connection disconnected: %s', connection_id)\n", "role": "user" }, { "content": "send_peers", "role": "assistant" } ]
def x(self, peer_endpoints): if self._topology: self._topology.add_candidate_peer_endpoints(peer_endpoints) else: LOGGER.debug( 'Could not add peer endpoints to topology. ConnectionManager does not exist.' )
add_candidate_peer_endpoints
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, peer_endpoints):\n if self._topology:\n self._topology.add_candidate_peer_endpoints(peer_endpoints)\n else:\n LOGGER.debug(\n 'Could not add peer endpoints to topology. ConnectionManager does not exist.'\n )\n", "role": "user" }, { "content": "add_candidate_peer_endpoints", "role": "assistant" } ]
def x(self): with self._lock: return [key for key in (self._network.connection_id_to_public_key( peer) for peer in copy.copy(self._peers)) if key is not None]
get_peers_public_keys
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n with self._lock:\n return [key for key in (self._network.connection_id_to_public_key(\n peer) for peer in copy.copy(self._peers)) if key is not None]\n", "role": "user" }, { "content": "get_peers_public_keys", "role": "assistant" } ]
def x(self, connection_id, endpoint): with self._lock: if len(self._peers) < self._maximum_peer_connectivity: self._peers[connection_id] = endpoint self._topology.set_connection_status(connection_id, PeerStatus.PEER ) LOGGER.debug( 'Added connection_id %s with endpoint %s, connected identities are now %s' , connection_id, endpoint, self._peers) else: raise PeeringException( 'At maximum configured number of peers: {} Rejecting peering request from {}.' .format(self._maximum_peer_connectivity, endpoint)) public_key = self.peer_to_public_key(connection_id) if public_key: self._consensus_notifier.notify_peer_connected(public_key)
register_peer
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, connection_id, endpoint):\n with self._lock:\n if len(self._peers) < self._maximum_peer_connectivity:\n self._peers[connection_id] = endpoint\n self._topology.set_connection_status(connection_id, PeerStatus.PEER\n )\n LOGGER.debug(\n 'Added connection_id %s with endpoint %s, connected identities are now %s'\n , connection_id, endpoint, self._peers)\n else:\n raise PeeringException(\n 'At maximum configured number of peers: {} Rejecting peering request from {}.'\n .format(self._maximum_peer_connectivity, endpoint))\n public_key = self.peer_to_public_key(connection_id)\n if public_key:\n self._consensus_notifier.notify_peer_connected(public_key)\n", "role": "user" }, { "content": "register_peer", "role": "assistant" } ]
def x(self, connection_id): public_key = self.peer_to_public_key(connection_id) if public_key: self._consensus_notifier.notify_peer_disconnected(public_key) with self._lock: if connection_id in self._peers: del self._peers[connection_id] LOGGER.debug( 'Removed connection_id %s, connected identities are now %s', connection_id, self._peers) self._topology.set_connection_status(connection_id, PeerStatus.TEMP ) else: LOGGER.warning( 'Connection unregister failed as connection was not registered: %s' , connection_id)
unregister_peer
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, connection_id):\n public_key = self.peer_to_public_key(connection_id)\n if public_key:\n self._consensus_notifier.notify_peer_disconnected(public_key)\n with self._lock:\n if connection_id in self._peers:\n del self._peers[connection_id]\n LOGGER.debug(\n 'Removed connection_id %s, connected identities are now %s',\n connection_id, self._peers)\n self._topology.set_connection_status(connection_id, PeerStatus.TEMP\n )\n else:\n LOGGER.warning(\n 'Connection unregister failed as connection was not registered: %s'\n , connection_id)\n", "role": "user" }, { "content": "unregister_peer", "role": "assistant" } ]
def x(self, message_type, message, connection_id, one_way=False): try: self._network.send(message_type, message, connection_id, one_way= one_way) except ValueError: LOGGER.debug( 'Connection %s is no longer valid. Removing from list of peers.', connection_id) if connection_id in self._peers: del self._peers[connection_id]
send
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, message_type, message, connection_id, one_way=False):\n try:\n self._network.send(message_type, message, connection_id, one_way=\n one_way)\n except ValueError:\n LOGGER.debug(\n 'Connection %s is no longer valid. Removing from list of peers.',\n connection_id)\n if connection_id in self._peers:\n del self._peers[connection_id]\n", "role": "user" }, { "content": "send", "role": "assistant" } ]
def x(self, gossip_message, message_type, exclude=None): with self._lock: if exclude is None: exclude = [] for connection_id in self._peers.copy(): if (connection_id not in exclude and self._network. is_connection_handshake_complete(connection_id)): self.send(message_type, gossip_message.SerializeToString(), connection_id, one_way=True)
broadcast
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, gossip_message, message_type, exclude=None):\n with self._lock:\n if exclude is None:\n exclude = []\n for connection_id in self._peers.copy():\n if (connection_id not in exclude and self._network.\n is_connection_handshake_complete(connection_id)):\n self.send(message_type, gossip_message.SerializeToString(),\n connection_id, one_way=True)\n", "role": "user" }, { "content": "broadcast", "role": "assistant" } ]
def x(self, peer_endpoints): with self._lock: for endpoint in peer_endpoints: if endpoint not in self._candidate_peer_endpoints: self._candidate_peer_endpoints.append(endpoint)
add_candidate_peer_endpoints
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, peer_endpoints):\n with self._lock:\n for endpoint in peer_endpoints:\n if endpoint not in self._candidate_peer_endpoints:\n self._candidate_peer_endpoints.append(endpoint)\n", "role": "user" }, { "content": "add_candidate_peer_endpoints", "role": "assistant" } ]
def x(self, connection_id): endpoint = self._network.connection_id_to_endpoint(connection_id) endpoint_info = self._temp_endpoints.get(endpoint) LOGGER.debug('Endpoint has completed authorization: %s (id: %s)', endpoint, connection_id) if endpoint_info is None: LOGGER.debug('Received unknown endpoint: %s', endpoint) elif endpoint_info.status == EndpointStatus.PEERING: self._connect_success_peering(connection_id, endpoint) elif endpoint_info.status == EndpointStatus.TOPOLOGY: self._connect_success_topology(connection_id) else: LOGGER.debug('Endpoint has unknown status: %s', endpoint) with self._lock: if endpoint in self._temp_endpoints: del self._temp_endpoints[endpoint]
connect_success
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, connection_id):\n endpoint = self._network.connection_id_to_endpoint(connection_id)\n endpoint_info = self._temp_endpoints.get(endpoint)\n LOGGER.debug('Endpoint has completed authorization: %s (id: %s)',\n endpoint, connection_id)\n if endpoint_info is None:\n LOGGER.debug('Received unknown endpoint: %s', endpoint)\n elif endpoint_info.status == EndpointStatus.PEERING:\n self._connect_success_peering(connection_id, endpoint)\n elif endpoint_info.status == EndpointStatus.TOPOLOGY:\n self._connect_success_topology(connection_id)\n else:\n LOGGER.debug('Endpoint has unknown status: %s', endpoint)\n with self._lock:\n if endpoint in self._temp_endpoints:\n del self._temp_endpoints[endpoint]\n", "role": "user" }, { "content": "connect_success", "role": "assistant" } ]
def x(pointer_type=ctypes.c_uint8): return ctypes.POINTER(pointer_type)(), ctypes.c_size_t(0), ctypes.c_size_t( 0)
prepare_vec_result
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(pointer_type=ctypes.c_uint8):\n return ctypes.POINTER(pointer_type)(), ctypes.c_size_t(0), ctypes.c_size_t(\n 0)\n", "role": "user" }, { "content": "prepare_vec_result", "role": "assistant" } ]
def x(self): if self._ptr: LIBRARY.call(self._drop_ffi_fn, self._ptr) self._ptr = None
drop
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self._ptr:\n LIBRARY.call(self._drop_ffi_fn, self._ptr)\n self._ptr = None\n", "role": "user" }, { "content": "drop", "role": "assistant" } ]
def x(): if 'SAWTOOTH_HOME' in os.environ: home_dir = os.environ['SAWTOOTH_HOME'] return PathConfig(config_dir=os.path.join(home_dir, 'etc'), log_dir =os.path.join(home_dir, 'logs'), data_dir=os.path.join(home_dir, 'data'), key_dir=os.path.join(home_dir, 'keys'), policy_dir=os. path.join(home_dir, 'policy')) if os.name == 'nt': base_dir = os.path.dirname(os.path.dirname(os.path.abspath(sys.argv [0]))) return PathConfig(config_dir=os.path.join(base_dir, 'conf'), log_dir=os.path.join(base_dir, 'logs'), data_dir=os.path.join( base_dir, 'data'), key_dir=os.path.join(base_dir, 'conf', 'keys'), policy_dir=os.path.join(base_dir, 'policy')) return PathConfig(config_dir='/etc/sawtooth', log_dir= '/var/log/sawtooth', data_dir='/var/lib/sawtooth', key_dir= '/etc/sawtooth/keys', policy_dir='/etc/sawtooth/policy')
get_default_path_config
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n if 'SAWTOOTH_HOME' in os.environ:\n home_dir = os.environ['SAWTOOTH_HOME']\n return PathConfig(config_dir=os.path.join(home_dir, 'etc'), log_dir\n =os.path.join(home_dir, 'logs'), data_dir=os.path.join(home_dir,\n 'data'), key_dir=os.path.join(home_dir, 'keys'), policy_dir=os.\n path.join(home_dir, 'policy'))\n if os.name == 'nt':\n base_dir = os.path.dirname(os.path.dirname(os.path.abspath(sys.argv\n [0])))\n return PathConfig(config_dir=os.path.join(base_dir, 'conf'),\n log_dir=os.path.join(base_dir, 'logs'), data_dir=os.path.join(\n base_dir, 'data'), key_dir=os.path.join(base_dir, 'conf',\n 'keys'), policy_dir=os.path.join(base_dir, 'policy'))\n return PathConfig(config_dir='/etc/sawtooth', log_dir=\n '/var/log/sawtooth', data_dir='/var/lib/sawtooth', key_dir=\n '/etc/sawtooth/keys', policy_dir='/etc/sawtooth/policy')\n", "role": "user" }, { "content": "get_default_path_config", "role": "assistant" } ]
def x(filename): if not os.path.exists(filename): LOGGER.info('Skipping path loading from non-existent config file: %s', filename) return PathConfig() LOGGER.info('Loading path information from config: %s', filename) try: with open(filename) as fd: raw_config = fd.read() except IOError as e: raise LocalConfigurationError( 'Unable to load path configuration file: {}'.format(str(e))) toml_config = toml.loads(raw_config) invalid_keys = set(toml_config.keys()).difference(['data_dir', 'key_dir', 'log_dir', 'policy_dir']) if invalid_keys: raise LocalConfigurationError('Invalid keys in path config: {}'. format(', '.join(sorted(list(invalid_keys))))) config = PathConfig(config_dir=None, data_dir=toml_config.get( 'data_dir', None), key_dir=toml_config.get('key_dir', None), log_dir=toml_config.get('log_dir', None), policy_dir=toml_config. get('policy_dir', None)) return config
load_toml_path_config
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(filename):\n if not os.path.exists(filename):\n LOGGER.info('Skipping path loading from non-existent config file: %s',\n filename)\n return PathConfig()\n LOGGER.info('Loading path information from config: %s', filename)\n try:\n with open(filename) as fd:\n raw_config = fd.read()\n except IOError as e:\n raise LocalConfigurationError(\n 'Unable to load path configuration file: {}'.format(str(e)))\n toml_config = toml.loads(raw_config)\n invalid_keys = set(toml_config.keys()).difference(['data_dir',\n 'key_dir', 'log_dir', 'policy_dir'])\n if invalid_keys:\n raise LocalConfigurationError('Invalid keys in path config: {}'.\n format(', '.join(sorted(list(invalid_keys)))))\n config = PathConfig(config_dir=None, data_dir=toml_config.get(\n 'data_dir', None), key_dir=toml_config.get('key_dir', None),\n log_dir=toml_config.get('log_dir', None), policy_dir=toml_config.\n get('policy_dir', None))\n return config\n", "role": "user" }, { "content": "load_toml_path_config", "role": "assistant" } ]
def x(configs, config_dir_override): config_dir = None log_dir = None data_dir = None key_dir = None policy_dir = None for config in reversed(configs): if config.config_dir is not None: config_dir = config.config_dir if config.log_dir is not None: log_dir = config.log_dir if config.data_dir is not None: data_dir = config.data_dir if config.key_dir is not None: key_dir = config.key_dir if config.policy_dir is not None: policy_dir = config.policy_dir if config_dir_override is not None: config_dir = config_dir_override return PathConfig(config_dir=config_dir, log_dir=log_dir, data_dir= data_dir, key_dir=key_dir, policy_dir=policy_dir)
merge_path_config
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(configs, config_dir_override):\n config_dir = None\n log_dir = None\n data_dir = None\n key_dir = None\n policy_dir = None\n for config in reversed(configs):\n if config.config_dir is not None:\n config_dir = config.config_dir\n if config.log_dir is not None:\n log_dir = config.log_dir\n if config.data_dir is not None:\n data_dir = config.data_dir\n if config.key_dir is not None:\n key_dir = config.key_dir\n if config.policy_dir is not None:\n policy_dir = config.policy_dir\n if config_dir_override is not None:\n config_dir = config_dir_override\n return PathConfig(config_dir=config_dir, log_dir=log_dir, data_dir=\n data_dir, key_dir=key_dir, policy_dir=policy_dir)\n", "role": "user" }, { "content": "merge_path_config", "role": "assistant" } ]
def x(self): genesis_file = os.path.join(self._data_dir, 'genesis.batch') has_genesis_batches = Path(genesis_file).is_file() LOGGER.debug('genesis_batch_file: %s', genesis_file if has_genesis_batches else 'not found') chain_head = self._block_store.chain_head has_chain_head = chain_head is not None if has_chain_head: LOGGER.debug('chain_head: %s', chain_head) block_chain_id = self._chain_id_manager.get_block_chain_id() is_genesis_node = block_chain_id is None LOGGER.debug('block_chain_id: %s', block_chain_id if not is_genesis_node else 'not yet specified') if has_genesis_batches and has_chain_head: raise InvalidGenesisStateError( 'Cannot have a genesis_batch_file and an existing chain') if has_genesis_batches and not is_genesis_node: raise InvalidGenesisStateError( 'Cannot have a genesis_batch_file and join an existing network') if not has_genesis_batches and not has_chain_head: LOGGER.info( 'No chain head and not the genesis node: starting in peering mode') return has_genesis_batches and not has_chain_head and is_genesis_node
requires_genesis
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n genesis_file = os.path.join(self._data_dir, 'genesis.batch')\n has_genesis_batches = Path(genesis_file).is_file()\n LOGGER.debug('genesis_batch_file: %s', genesis_file if\n has_genesis_batches else 'not found')\n chain_head = self._block_store.chain_head\n has_chain_head = chain_head is not None\n if has_chain_head:\n LOGGER.debug('chain_head: %s', chain_head)\n block_chain_id = self._chain_id_manager.get_block_chain_id()\n is_genesis_node = block_chain_id is None\n LOGGER.debug('block_chain_id: %s', block_chain_id if not\n is_genesis_node else 'not yet specified')\n if has_genesis_batches and has_chain_head:\n raise InvalidGenesisStateError(\n 'Cannot have a genesis_batch_file and an existing chain')\n if has_genesis_batches and not is_genesis_node:\n raise InvalidGenesisStateError(\n 'Cannot have a genesis_batch_file and join an existing network')\n if not has_genesis_batches and not has_chain_head:\n LOGGER.info(\n 'No chain head and not the genesis node: starting in peering mode')\n return has_genesis_batches and not has_chain_head and is_genesis_node\n", "role": "user" }, { "content": "requires_genesis", "role": "assistant" } ]
def x(self, on_done): genesis_file = os.path.join(self._data_dir, 'genesis.batch') try: with open(genesis_file, 'rb') as batch_file: genesis_data = genesis_pb2.GenesisData() genesis_data.ParseFromString(batch_file.read()) LOGGER.info('Producing genesis block from %s', genesis_file) except IOError: raise InvalidGenesisStateError( 'Genesis File {} specified, but unreadable'.format(genesis_file)) initial_state_root = self._context_manager.get_first_root() genesis_batches = [batch for batch in genesis_data.batches] if genesis_batches: scheduler = SerialScheduler(self._context_manager. get_squash_handler(), initial_state_root, always_persist=True) LOGGER.debug('Adding %s batches', len(genesis_data.batches)) for batch in genesis_data.batches: scheduler.add_batch(batch) self._transaction_executor.execute(scheduler) scheduler.finalize() scheduler.complete(block=True) txn_receipts = [] state_hash = initial_state_root for batch in genesis_batches: result = scheduler.get_batch_execution_result(batch.header_signature) if result is None or not result.is_valid: raise InvalidGenesisStateError( 'Unable to create genesis block, due to batch {}'.format( batch.header_signature)) if result.state_hash is not None: state_hash = result.state_hash txn_results = scheduler.get_transaction_execution_results(batch. header_signature) txn_receipts += self._make_receipts(txn_results) settings_view = SettingsView(self._state_view_factory.create_view( state_hash)) name = settings_view.get_setting('sawtooth.consensus.algorithm.name') version = settings_view.get_setting('sawtooth.consensus.algorithm.version') if name is None or version is None: raise LocalConfigurationError( 'Unable to start validator; sawtooth.consensus.algorithm.name and sawtooth.consensus.algorithm.version must be set in the genesis block.' ) LOGGER.debug('Produced state hash %s for genesis block.', state_hash) block_builder = self._generate_genesis_block() block_builder.add_batches(genesis_batches) block_builder.set_state_hash(state_hash) block_publisher = self._get_block_publisher(initial_state_root) if not block_publisher.initialize_block(block_builder.block_header): LOGGER.error('Consensus refused to initialize consensus block.') raise InvalidGenesisConsensusError( 'Consensus refused to initialize genesis block.') if not block_publisher.finalize_block(block_builder.block_header): LOGGER.error('Consensus refused to finalize genesis block.') raise InvalidGenesisConsensusError( 'Consensus refused to finalize genesis block.') self._sign_block(block_builder) block = block_builder.build_block() blkw = BlockWrapper(block=block) LOGGER.info('Genesis block created: %s', blkw) self._block_manager.put([blkw.block]) self._block_manager.persist(blkw.identifier, 'commit_store') self._txn_receipt_store.chain_update(block, txn_receipts) self._chain_id_manager.save_block_chain_id(block.header_signature) LOGGER.debug('Deleting genesis data.') os.remove(genesis_file) if on_done is not None: on_done()
start
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, on_done):\n genesis_file = os.path.join(self._data_dir, 'genesis.batch')\n try:\n with open(genesis_file, 'rb') as batch_file:\n genesis_data = genesis_pb2.GenesisData()\n genesis_data.ParseFromString(batch_file.read())\n LOGGER.info('Producing genesis block from %s', genesis_file)\n except IOError:\n raise InvalidGenesisStateError(\n 'Genesis File {} specified, but unreadable'.format(genesis_file))\n initial_state_root = self._context_manager.get_first_root()\n genesis_batches = [batch for batch in genesis_data.batches]\n if genesis_batches:\n scheduler = SerialScheduler(self._context_manager.\n get_squash_handler(), initial_state_root, always_persist=True)\n LOGGER.debug('Adding %s batches', len(genesis_data.batches))\n for batch in genesis_data.batches:\n scheduler.add_batch(batch)\n self._transaction_executor.execute(scheduler)\n scheduler.finalize()\n scheduler.complete(block=True)\n txn_receipts = []\n state_hash = initial_state_root\n for batch in genesis_batches:\n result = scheduler.get_batch_execution_result(batch.header_signature)\n if result is None or not result.is_valid:\n raise InvalidGenesisStateError(\n 'Unable to create genesis block, due to batch {}'.format(\n batch.header_signature))\n if result.state_hash is not None:\n state_hash = result.state_hash\n txn_results = scheduler.get_transaction_execution_results(batch.\n header_signature)\n txn_receipts += self._make_receipts(txn_results)\n settings_view = SettingsView(self._state_view_factory.create_view(\n state_hash))\n name = settings_view.get_setting('sawtooth.consensus.algorithm.name')\n version = settings_view.get_setting('sawtooth.consensus.algorithm.version')\n if name is None or version is None:\n raise LocalConfigurationError(\n 'Unable to start validator; sawtooth.consensus.algorithm.name and sawtooth.consensus.algorithm.version must be set in the genesis block.'\n )\n LOGGER.debug('Produced state hash %s for genesis block.', state_hash)\n block_builder = self._generate_genesis_block()\n block_builder.add_batches(genesis_batches)\n block_builder.set_state_hash(state_hash)\n block_publisher = self._get_block_publisher(initial_state_root)\n if not block_publisher.initialize_block(block_builder.block_header):\n LOGGER.error('Consensus refused to initialize consensus block.')\n raise InvalidGenesisConsensusError(\n 'Consensus refused to initialize genesis block.')\n if not block_publisher.finalize_block(block_builder.block_header):\n LOGGER.error('Consensus refused to finalize genesis block.')\n raise InvalidGenesisConsensusError(\n 'Consensus refused to finalize genesis block.')\n self._sign_block(block_builder)\n block = block_builder.build_block()\n blkw = BlockWrapper(block=block)\n LOGGER.info('Genesis block created: %s', blkw)\n self._block_manager.put([blkw.block])\n self._block_manager.persist(blkw.identifier, 'commit_store')\n self._txn_receipt_store.chain_update(block, txn_receipts)\n self._chain_id_manager.save_block_chain_id(block.header_signature)\n LOGGER.debug('Deleting genesis data.')\n os.remove(genesis_file)\n if on_done is not None:\n on_done()\n", "role": "user" }, { "content": "start", "role": "assistant" } ]
def x(self, state_hash): state_view = self._state_view_factory.create_view(state_hash) try: class BatchPublisher: def x(self, transactions): raise InvalidGenesisConsensusError( 'Consensus cannot send transactions during genesis.') consensus = ConsensusFactory.get_configured_consensus_module( NULL_BLOCK_IDENTIFIER, state_view) return consensus.BlockPublisher(BlockCache(self._block_store), state_view_factory=self._state_view_factory, batch_publisher= BatchPublisher(), data_dir=self._data_dir, config_dir=self. _config_dir, validator_id=self._identity_signer.get_public_key( ).as_hex()) except UnknownConsensusModuleError as e: raise InvalidGenesisStateError(e)
get_block_publisher
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, state_hash):\n state_view = self._state_view_factory.create_view(state_hash)\n try:\n\n\n class BatchPublisher:\n\n def x(self, transactions):\n raise InvalidGenesisConsensusError(\n 'Consensus cannot send transactions during genesis.')\n consensus = ConsensusFactory.get_configured_consensus_module(\n NULL_BLOCK_IDENTIFIER, state_view)\n return consensus.BlockPublisher(BlockCache(self._block_store),\n state_view_factory=self._state_view_factory, batch_publisher=\n BatchPublisher(), data_dir=self._data_dir, config_dir=self.\n _config_dir, validator_id=self._identity_signer.get_public_key(\n ).as_hex())\n except UnknownConsensusModuleError as e:\n raise InvalidGenesisStateError(e)\n", "role": "user" }, { "content": "get_block_publisher", "role": "assistant" } ]
def x(self): genesis_header = block_pb2.BlockHeader(block_num=0, previous_block_id= NULL_BLOCK_IDENTIFIER, signer_public_key=self._identity_signer. get_public_key().as_hex()) return BlockBuilder(genesis_header)
generate_genesis_block
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n genesis_header = block_pb2.BlockHeader(block_num=0, previous_block_id=\n NULL_BLOCK_IDENTIFIER, signer_public_key=self._identity_signer.\n get_public_key().as_hex())\n return BlockBuilder(genesis_header)\n", "role": "user" }, { "content": "generate_genesis_block", "role": "assistant" } ]
def x(self, block): block_header = block.block_header header_bytes = block_header.SerializeToString() signature = self._identity_signer.sign(header_bytes) block.set_signature(signature) return block
sign_block
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, block):\n block_header = block.block_header\n header_bytes = block_header.SerializeToString()\n signature = self._identity_signer.sign(header_bytes)\n block.set_signature(signature)\n return block\n", "role": "user" }, { "content": "sign_block", "role": "assistant" } ]
def x(self): time_horizon = time.time() - self._keep_time new_cache = {} for k, v in self._cache.items(): if v.timestamp > time_horizon: new_cache[k] = v self._cache = new_cache
purge_expired
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n time_horizon = time.time() - self._keep_time\n new_cache = {}\n for k, v in self._cache.items():\n if v.timestamp > time_horizon:\n new_cache[k] = v\n self._cache = new_cache\n", "role": "user" }, { "content": "purge_expired", "role": "assistant" } ]
def x(): conf_file = os.path.join(_get_config_dir(), 'log_config.toml') if os.path.exists(conf_file): with open(conf_file) as fd: raw_config = fd.read() log_config = toml.loads(raw_config) return log_config conf_file = os.path.join(_get_config_dir(), 'log_config.yaml') if os.path.exists(conf_file): with open(conf_file) as fd: raw_config = fd.read() log_config = yaml.safe_load(raw_config) return log_config return None
get_config
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n conf_file = os.path.join(_get_config_dir(), 'log_config.toml')\n if os.path.exists(conf_file):\n with open(conf_file) as fd:\n raw_config = fd.read()\n log_config = toml.loads(raw_config)\n return log_config\n conf_file = os.path.join(_get_config_dir(), 'log_config.yaml')\n if os.path.exists(conf_file):\n with open(conf_file) as fd:\n raw_config = fd.read()\n log_config = yaml.safe_load(raw_config)\n return log_config\n return None\n", "role": "user" }, { "content": "get_config", "role": "assistant" } ]
def x(subparsers, parent_parser): parser = subparsers.add_parser('peer', help= 'Displays information about validator peers', description= "Provides a subcommand to list a validator's peers") grand_parsers = parser.add_subparsers(title='subcommands', dest= 'subcommand') grand_parsers.required = True add_peer_list_parser(grand_parsers, parent_parser)
add_peer_parser
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(subparsers, parent_parser):\n parser = subparsers.add_parser('peer', help=\n 'Displays information about validator peers', description=\n \"Provides a subcommand to list a validator's peers\")\n grand_parsers = parser.add_subparsers(title='subcommands', dest=\n 'subcommand')\n grand_parsers.required = True\n add_peer_list_parser(grand_parsers, parent_parser)\n", "role": "user" }, { "content": "add_peer_parser", "role": "assistant" } ]
def x(self): c_length = ctypes.c_int(0) c_limit = ctypes.c_int(0) self._call('pending_batch_info', ctypes.byref(c_length), ctypes.byref( c_limit)) return c_length.value, c_limit.value
pending_batch_info
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n c_length = ctypes.c_int(0)\n c_limit = ctypes.c_int(0)\n self._call('pending_batch_info', ctypes.byref(c_length), ctypes.byref(\n c_limit))\n return c_length.value, c_limit.value\n", "role": "user" }, { "content": "pending_batch_info", "role": "assistant" } ]
def x(self, chain_head, committed_batches=None, uncommitted_batches=None): try: self._py_call('on_chain_updated', ctypes.py_object(chain_head), ctypes.py_object(committed_batches), ctypes.py_object( uncommitted_batches)) except Exception: LOGGER.exception( 'Unhandled exception in BlockPublisher.on_chain_updated')
on_chain_updated
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, chain_head, committed_batches=None, uncommitted_batches=None):\n try:\n self._py_call('on_chain_updated', ctypes.py_object(chain_head),\n ctypes.py_object(committed_batches), ctypes.py_object(\n uncommitted_batches))\n except Exception:\n LOGGER.exception(\n 'Unhandled exception in BlockPublisher.on_chain_updated')\n", "role": "user" }, { "content": "on_chain_updated", "role": "assistant" } ]
def x(self): while True: try: if self._socket.getsockopt(zmq.TYPE) == zmq.ROUTER: zmq_identity, msg_bytes = (yield from self._socket. recv_multipart()) if msg_bytes == b'': LOGGER.debug('ROUTER PROBE FROM %s', zmq_identity) self._socket.send_multipart([bytes(zmq_identity), msg_bytes]) else: self._received_from_identity(zmq_identity) self._dispatcher_queue.put_nowait((zmq_identity, msg_bytes) ) else: msg_bytes = yield from self._socket.recv() self._last_message_time = time.time() self._dispatcher_queue.put_nowait((None, msg_bytes)) self._get_queue_size_gauge(self.connection).set_value(self. _dispatcher_queue.qsize()) except CancelledError: raise except Exception as e: LOGGER.exception( 'Received a message on address %s that caused an error: %s', self._address, e)
receive_message
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n while True:\n try:\n if self._socket.getsockopt(zmq.TYPE) == zmq.ROUTER:\n zmq_identity, msg_bytes = (yield from self._socket.\n recv_multipart())\n if msg_bytes == b'':\n LOGGER.debug('ROUTER PROBE FROM %s', zmq_identity)\n self._socket.send_multipart([bytes(zmq_identity),\n msg_bytes])\n else:\n self._received_from_identity(zmq_identity)\n self._dispatcher_queue.put_nowait((zmq_identity, msg_bytes)\n )\n else:\n msg_bytes = yield from self._socket.recv()\n self._last_message_time = time.time()\n self._dispatcher_queue.put_nowait((None, msg_bytes))\n self._get_queue_size_gauge(self.connection).set_value(self.\n _dispatcher_queue.qsize())\n except CancelledError:\n raise\n except Exception as e:\n LOGGER.exception(\n 'Received a message on address %s that caused an error: %s',\n self._address, e)\n", "role": "user" }, { "content": "receive_message", "role": "assistant" } ]
def x(self, msg, connection_id=None): zmq_identity = None if connection_id is not None and self._connections is not None: if connection_id in self._connections: connection_info = self._connections.get(connection_id) if connection_info.connection_type == ConnectionType.ZMQ_IDENTITY: zmq_identity = connection_info.connection else: LOGGER.debug("Can't send to %s, not in self._connections", connection_id) self._ready.wait() if zmq_identity is None: message_bundle = [msg.SerializeToString()] else: message_bundle = [bytes(zmq_identity), msg.SerializeToString()] try: asyncio.run_coroutine_threadsafe(self._send_message_frame( message_bundle), self._event_loop) except RuntimeError: pass
send_message
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, msg, connection_id=None):\n zmq_identity = None\n if connection_id is not None and self._connections is not None:\n if connection_id in self._connections:\n connection_info = self._connections.get(connection_id)\n if connection_info.connection_type == ConnectionType.ZMQ_IDENTITY:\n zmq_identity = connection_info.connection\n else:\n LOGGER.debug(\"Can't send to %s, not in self._connections\",\n connection_id)\n self._ready.wait()\n if zmq_identity is None:\n message_bundle = [msg.SerializeToString()]\n else:\n message_bundle = [bytes(zmq_identity), msg.SerializeToString()]\n try:\n asyncio.run_coroutine_threadsafe(self._send_message_frame(\n message_bundle), self._event_loop)\n except RuntimeError:\n pass\n", "role": "user" }, { "content": "send_message", "role": "assistant" } ]
def x(self, msg, connection_id=None): zmq_identity = None if connection_id is not None and self._connections is not None: if connection_id in self._connections: connection_info = self._connections.get(connection_id) if connection_info.connection_type == ConnectionType.ZMQ_IDENTITY: zmq_identity = connection_info.connection del self._connections[connection_id] else: LOGGER.debug("Can't send to %s, not in self._connections", connection_id) return self._ready.wait() try: asyncio.run_coroutine_threadsafe(self._send_last_message( zmq_identity, msg), self._event_loop) except RuntimeError: pass
send_last_message
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, msg, connection_id=None):\n zmq_identity = None\n if connection_id is not None and self._connections is not None:\n if connection_id in self._connections:\n connection_info = self._connections.get(connection_id)\n if connection_info.connection_type == ConnectionType.ZMQ_IDENTITY:\n zmq_identity = connection_info.connection\n del self._connections[connection_id]\n else:\n LOGGER.debug(\"Can't send to %s, not in self._connections\",\n connection_id)\n return\n self._ready.wait()\n try:\n asyncio.run_coroutine_threadsafe(self._send_last_message(\n zmq_identity, msg), self._event_loop)\n except RuntimeError:\n pass\n", "role": "user" }, { "content": "send_last_message", "role": "assistant" } ]
def x(self, socket_type, complete_or_error_queue): try: if self._secured: if (self._server_public_key is None or self._server_private_key is None): raise LocalConfigurationError( 'Attempting to start socket in secure mode, but complete server keys were not provided' ) self._event_loop = zmq.asyncio.ZMQEventLoop() asyncio.set_event_loop(self._event_loop) self._context = zmq.asyncio.Context() self._socket = self._context.socket(socket_type) self._socket.set(zmq.TCP_KEEPALIVE, 1) self._socket.set(zmq.TCP_KEEPALIVE_IDLE, self._connection_timeout) self._socket.set(zmq.TCP_KEEPALIVE_INTVL, self._heartbeat_interval) if socket_type == zmq.DEALER: self._socket.identity = '{}-{}'.format(self._zmq_identity, hashlib.sha512(uuid.uuid4().hex.encode()).hexdigest()[:23] ).encode('ascii') if self._secured: public_key, secretkey = zmq.curve_keypair() self._socket.curve_publickey = public_key self._socket.curve_secretkey = secretkey self._socket.curve_serverkey = self._server_public_key self._socket.connect(self._address) elif socket_type == zmq.ROUTER: if self._secured: auth = AsyncioAuthenticator(self._context) self._auth = auth auth.start() auth.configure_curve(domain='*', location=zmq.auth. CURVE_ALLOW_ANY) self._socket.curve_secretkey = self._server_private_key self._socket.curve_publickey = self._server_public_key self._socket.curve_server = True try: self._socket.bind(self._address) except zmq.error.ZMQError as e: raise LocalConfigurationError("Can't bind to {}: {}".format (self._address, str(e))) else: LOGGER.info('Listening on %s', self._address) self._dispatcher.add_send_message(self._connection, self.send_message) self._dispatcher.add_send_last_message(self._connection, self. send_last_message) asyncio.ensure_future(self._remove_expired_futures(), loop=self. _event_loop) asyncio.ensure_future(self._receive_message(), loop=self._event_loop) asyncio.ensure_future(self._dispatch_message(), loop=self._event_loop) self._dispatcher_queue = asyncio.Queue() if self._monitor: self._monitor_fd = 'inproc://monitor.s-{}'.format(_generate_id( )[0:5]) self._monitor_sock = self._socket.get_monitor_socket(zmq. EVENT_DISCONNECTED, addr=self._monitor_fd) asyncio.ensure_future(self._monitor_disconnects(), loop=self. _event_loop) except Exception as e: complete_or_error_queue.put_nowait(e) self._close_sockets() raise if self._heartbeat: asyncio.ensure_future(self._do_heartbeat(), loop=self._event_loop) complete_or_error_queue.put_nowait(_STARTUP_COMPLETE_SENTINEL) asyncio.ensure_future(self._notify_started(), loop=self._event_loop) self._event_loop.run_forever() self._event_loop.close() self._close_sockets()
setup
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, socket_type, complete_or_error_queue):\n try:\n if self._secured:\n if (self._server_public_key is None or self._server_private_key is\n None):\n raise LocalConfigurationError(\n 'Attempting to start socket in secure mode, but complete server keys were not provided'\n )\n self._event_loop = zmq.asyncio.ZMQEventLoop()\n asyncio.set_event_loop(self._event_loop)\n self._context = zmq.asyncio.Context()\n self._socket = self._context.socket(socket_type)\n self._socket.set(zmq.TCP_KEEPALIVE, 1)\n self._socket.set(zmq.TCP_KEEPALIVE_IDLE, self._connection_timeout)\n self._socket.set(zmq.TCP_KEEPALIVE_INTVL, self._heartbeat_interval)\n if socket_type == zmq.DEALER:\n self._socket.identity = '{}-{}'.format(self._zmq_identity,\n hashlib.sha512(uuid.uuid4().hex.encode()).hexdigest()[:23]\n ).encode('ascii')\n if self._secured:\n public_key, secretkey = zmq.curve_keypair()\n self._socket.curve_publickey = public_key\n self._socket.curve_secretkey = secretkey\n self._socket.curve_serverkey = self._server_public_key\n self._socket.connect(self._address)\n elif socket_type == zmq.ROUTER:\n if self._secured:\n auth = AsyncioAuthenticator(self._context)\n self._auth = auth\n auth.start()\n auth.configure_curve(domain='*', location=zmq.auth.\n CURVE_ALLOW_ANY)\n self._socket.curve_secretkey = self._server_private_key\n self._socket.curve_publickey = self._server_public_key\n self._socket.curve_server = True\n try:\n self._socket.bind(self._address)\n except zmq.error.ZMQError as e:\n raise LocalConfigurationError(\"Can't bind to {}: {}\".format\n (self._address, str(e)))\n else:\n LOGGER.info('Listening on %s', self._address)\n self._dispatcher.add_send_message(self._connection, self.send_message)\n self._dispatcher.add_send_last_message(self._connection, self.\n send_last_message)\n asyncio.ensure_future(self._remove_expired_futures(), loop=self.\n _event_loop)\n asyncio.ensure_future(self._receive_message(), loop=self._event_loop)\n asyncio.ensure_future(self._dispatch_message(), loop=self._event_loop)\n self._dispatcher_queue = asyncio.Queue()\n if self._monitor:\n self._monitor_fd = 'inproc://monitor.s-{}'.format(_generate_id(\n )[0:5])\n self._monitor_sock = self._socket.get_monitor_socket(zmq.\n EVENT_DISCONNECTED, addr=self._monitor_fd)\n asyncio.ensure_future(self._monitor_disconnects(), loop=self.\n _event_loop)\n except Exception as e:\n complete_or_error_queue.put_nowait(e)\n self._close_sockets()\n raise\n if self._heartbeat:\n asyncio.ensure_future(self._do_heartbeat(), loop=self._event_loop)\n complete_or_error_queue.put_nowait(_STARTUP_COMPLETE_SENTINEL)\n asyncio.ensure_future(self._notify_started(), loop=self._event_loop)\n self._event_loop.run_forever()\n self._event_loop.close()\n self._close_sockets()\n", "role": "user" }, { "content": "setup", "role": "assistant" } ]
def x(self, connection_id): with self._connections_lock: try: connection_info = self._connections[connection_id] return connection_info.public_key except KeyError: return None
connection_id_to_public_key
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, connection_id):\n with self._connections_lock:\n try:\n connection_info = self._connections[connection_id]\n return connection_info.public_key\n except KeyError:\n return None\n", "role": "user" }, { "content": "connection_id_to_public_key", "role": "assistant" } ]
def x(self, public_key): with self._connections_lock: for connection_id, connection_info in self._connections.items(): if connection_info.public_key == public_key: return connection_id return None
public_key_to_connection_id
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, public_key):\n with self._connections_lock:\n for connection_id, connection_info in self._connections.items():\n if connection_info.public_key == public_key:\n return connection_id\n return None\n", "role": "user" }, { "content": "public_key_to_connection_id", "role": "assistant" } ]
def x(self, connection_id): with self._connections_lock: try: connection_info = self._connections[connection_id] return connection_info.uri except KeyError: return None
connection_id_to_endpoint
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, connection_id):\n with self._connections_lock:\n try:\n connection_info = self._connections[connection_id]\n return connection_info.uri\n except KeyError:\n return None\n", "role": "user" }, { "content": "connection_id_to_endpoint", "role": "assistant" } ]
def x(self, connection_id): with self._connections_lock: try: connection_info = self._connections[connection_id] return connection_info.status except KeyError: return None
get_connection_status
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, connection_id):\n with self._connections_lock:\n try:\n connection_info = self._connections[connection_id]\n return connection_info.status\n except KeyError:\n return None\n", "role": "user" }, { "content": "get_connection_status", "role": "assistant" } ]
def x(self): LOGGER.debug( 'Determining whether inbound connection should be allowed. num connections: %s max %s' , len(self._connections), self._max_incoming_connections) return self._max_incoming_connections >= len(self._connections)
allow_inbound_connection
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n LOGGER.debug(\n 'Determining whether inbound connection should be allowed. num connections: %s max %s'\n , len(self._connections), self._max_incoming_connections)\n return self._max_incoming_connections >= len(self._connections)\n", "role": "user" }, { "content": "allow_inbound_connection", "role": "assistant" } ]
def x(self, uri): LOGGER.debug('Adding connection to %s', uri) conn = OutboundConnection(connections=self._connections, endpoint=uri, dispatcher=self._dispatcher, zmq_identity=self._zmq_identity, secured=self._secured, server_public_key=self._server_public_key, server_private_key=self._server_private_key, future_callback_threadpool=self._future_callback_threadpool, heartbeat=True, connection_timeout=self._connection_timeout) self.outbound_connections[uri] = conn conn.start() self._add_connection(conn, uri) connect_message = ConnectionRequest(endpoint=self._public_endpoint) conn.send(validator_pb2.Message.NETWORK_CONNECT, connect_message. SerializeToString(), callback=partial(self._connect_callback, connection=conn)) return conn
add_outbound_connection
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, uri):\n LOGGER.debug('Adding connection to %s', uri)\n conn = OutboundConnection(connections=self._connections, endpoint=uri,\n dispatcher=self._dispatcher, zmq_identity=self._zmq_identity,\n secured=self._secured, server_public_key=self._server_public_key,\n server_private_key=self._server_private_key,\n future_callback_threadpool=self._future_callback_threadpool,\n heartbeat=True, connection_timeout=self._connection_timeout)\n self.outbound_connections[uri] = conn\n conn.start()\n self._add_connection(conn, uri)\n connect_message = ConnectionRequest(endpoint=self._public_endpoint)\n conn.send(validator_pb2.Message.NETWORK_CONNECT, connect_message.\n SerializeToString(), callback=partial(self._connect_callback,\n connection=conn))\n return conn\n", "role": "user" }, { "content": "add_outbound_connection", "role": "assistant" } ]
def x(self, connection_id): connect_message = ConnectionRequest(endpoint=self._public_endpoint) self._safe_send(validator_pb2.Message.NETWORK_CONNECT, connect_message. SerializeToString(), connection_id, callback=partial(self. _inbound_connection_request_callback, connection_id=connection_id))
send_connect_request
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, connection_id):\n connect_message = ConnectionRequest(endpoint=self._public_endpoint)\n self._safe_send(validator_pb2.Message.NETWORK_CONNECT, connect_message.\n SerializeToString(), connection_id, callback=partial(self.\n _inbound_connection_request_callback, connection_id=connection_id))\n", "role": "user" }, { "content": "send_connect_request", "role": "assistant" } ]
def x(self, message_type, data, connection_id, callback=None, one_way=False): if connection_id not in self._connections: raise ValueError('Unknown connection id: {}'.format(connection_id)) connection_info = self._connections.get(connection_id) if connection_info.connection_type == ConnectionType.ZMQ_IDENTITY: message = validator_pb2.Message(correlation_id=_generate_id(), content=data, message_type=message_type) timer_tag = get_enum_name(message.message_type) timer_ctx = self._get_send_response_timer(timer_tag).time() fut = future.Future(message.correlation_id, message.content, callback, timeout=self._connection_timeout, timer_ctx=timer_ctx) if not one_way: self._futures.put(fut) self._send_receive_thread.send_message(msg=message, connection_id= connection_id) return fut return connection_info.connection.send(message_type, data, callback= callback, one_way=one_way)
send
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, message_type, data, connection_id, callback=None, one_way=False):\n if connection_id not in self._connections:\n raise ValueError('Unknown connection id: {}'.format(connection_id))\n connection_info = self._connections.get(connection_id)\n if connection_info.connection_type == ConnectionType.ZMQ_IDENTITY:\n message = validator_pb2.Message(correlation_id=_generate_id(),\n content=data, message_type=message_type)\n timer_tag = get_enum_name(message.message_type)\n timer_ctx = self._get_send_response_timer(timer_tag).time()\n fut = future.Future(message.correlation_id, message.content,\n callback, timeout=self._connection_timeout, timer_ctx=timer_ctx)\n if not one_way:\n self._futures.put(fut)\n self._send_receive_thread.send_message(msg=message, connection_id=\n connection_id)\n return fut\n return connection_info.connection.send(message_type, data, callback=\n callback, one_way=one_way)\n", "role": "user" }, { "content": "send", "role": "assistant" } ]
def x(self, endpoint): with self._connections_lock: for connection_id in self._connections: connection_info = self._connections[connection_id] if connection_info.uri == endpoint: return connection_id raise KeyError()
get_connection_id_by_endpoint
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, endpoint):\n with self._connections_lock:\n for connection_id in self._connections:\n connection_info = self._connections[connection_id]\n if connection_info.uri == endpoint:\n return connection_id\n raise KeyError()\n", "role": "user" }, { "content": "get_connection_id_by_endpoint", "role": "assistant" } ]
def x(self, connection_id, endpoint): if connection_id in self._connections: connection_info = self._connections[connection_id] self._connections[connection_id] = ConnectionInfo(connection_info. connection_type, connection_info.connection, endpoint, connection_info.status, connection_info.public_key) else: LOGGER.debug( 'Could not update the endpoint %s for connection_id %s. The connection does not exist.' , endpoint, connection_id)
update_connection_endpoint
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, connection_id, endpoint):\n if connection_id in self._connections:\n connection_info = self._connections[connection_id]\n self._connections[connection_id] = ConnectionInfo(connection_info.\n connection_type, connection_info.connection, endpoint,\n connection_info.status, connection_info.public_key)\n else:\n LOGGER.debug(\n 'Could not update the endpoint %s for connection_id %s. The connection does not exist.'\n , endpoint, connection_id)\n", "role": "user" }, { "content": "update_connection_endpoint", "role": "assistant" } ]
def x(self, connection_id, public_key): if connection_id in self._connections: connection_info = self._connections[connection_id] self._connections[connection_id] = ConnectionInfo(connection_info. connection_type, connection_info.connection, connection_info. uri, connection_info.status, public_key) else: LOGGER.debug( 'Could not update the public key %s for connection_id %s. The connection does not exist.' , public_key, connection_id)
update_connection_public_key
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, connection_id, public_key):\n if connection_id in self._connections:\n connection_info = self._connections[connection_id]\n self._connections[connection_id] = ConnectionInfo(connection_info.\n connection_type, connection_info.connection, connection_info.\n uri, connection_info.status, public_key)\n else:\n LOGGER.debug(\n 'Could not update the public key %s for connection_id %s. The connection does not exist.'\n , public_key, connection_id)\n", "role": "user" }, { "content": "update_connection_public_key", "role": "assistant" } ]
def x(self, message_type, data, connection_id, callback=None, one_way=False): if connection_id not in self._connections: raise ValueError('Unknown connection id: {}'.format(connection_id)) connection_info = self._connections.get(connection_id) if connection_info.connection_type == ConnectionType.ZMQ_IDENTITY: message = validator_pb2.Message(correlation_id=_generate_id(), content=data, message_type=message_type) fut = future.Future(message.correlation_id, message.content, callback, timeout=self._connection_timeout) if not one_way: self._futures.put(fut) self._send_receive_thread.send_last_message(msg=message, connection_id=connection_id) return fut del self._connections[connection_id] return connection_info.connection.send_last_message(message_type, data, callback=callback)
send_last_message
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, message_type, data, connection_id, callback=None, one_way=False):\n if connection_id not in self._connections:\n raise ValueError('Unknown connection id: {}'.format(connection_id))\n connection_info = self._connections.get(connection_id)\n if connection_info.connection_type == ConnectionType.ZMQ_IDENTITY:\n message = validator_pb2.Message(correlation_id=_generate_id(),\n content=data, message_type=message_type)\n fut = future.Future(message.correlation_id, message.content,\n callback, timeout=self._connection_timeout)\n if not one_way:\n self._futures.put(fut)\n self._send_receive_thread.send_last_message(msg=message,\n connection_id=connection_id)\n return fut\n del self._connections[connection_id]\n return connection_info.connection.send_last_message(message_type, data,\n callback=callback)\n", "role": "user" }, { "content": "send_last_message", "role": "assistant" } ]
def x(self, message_type, data, callback=None, one_way=False): message = validator_pb2.Message(correlation_id=_generate_id(), content= data, message_type=message_type) fut = future.Future(message.correlation_id, message.content, callback, timeout=self._connection_timeout) if not one_way: self._futures.put(fut) self._send_receive_thread.send_message(message) return fut
send
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, message_type, data, callback=None, one_way=False):\n message = validator_pb2.Message(correlation_id=_generate_id(), content=\n data, message_type=message_type)\n fut = future.Future(message.correlation_id, message.content, callback,\n timeout=self._connection_timeout)\n if not one_way:\n self._futures.put(fut)\n self._send_receive_thread.send_message(message)\n return fut\n", "role": "user" }, { "content": "send", "role": "assistant" } ]
def x(self): if self._block_header is None: self._block_header = BlockHeader() self._block_header.ParseFromString(self.block.header) return self._block_header
header
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self._block_header is None:\n self._block_header = BlockHeader()\n self._block_header.ParseFromString(self.block.header)\n return self._block_header\n", "role": "user" }, { "content": "header", "role": "assistant" } ]
def x(block_wrapper, state_view_factory): state_root_hash = (block_wrapper.state_root_hash if block_wrapper is not None else None) return state_view_factory.create_view(state_root_hash)
state_view_for_block
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(block_wrapper, state_view_factory):\n state_root_hash = (block_wrapper.state_root_hash if block_wrapper is not\n None else None)\n return state_view_factory.create_view(state_root_hash)\n", "role": "user" }, { "content": "state_view_for_block", "role": "assistant" } ]
def x(block_wrapper, settings_view_factory): state_root_hash = (block_wrapper.state_root_hash if block_wrapper is not None else None) return settings_view_factory.create_settings_view(state_root_hash)
settings_view_for_block
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(block_wrapper, settings_view_factory):\n state_root_hash = (block_wrapper.state_root_hash if block_wrapper is not\n None else None)\n return settings_view_factory.create_settings_view(state_root_hash)\n", "role": "user" }, { "content": "settings_view_for_block", "role": "assistant" } ]
def x(remote_addr): if not isinstance(remote_addr, text_type) and isinstance(remote_addr, str): remote_addr = remote_addr.decode('ascii', 'ignore') try: ipv4 = ipaddress.IPv4Address(remote_addr) return u''.join(ipv4.exploded.rsplit('.', 1)[0]) + '.' + '0' except ipaddress.AddressValueError: try: ipv6 = ipaddress.IPv6Address(remote_addr) if ipv6.ipv4_mapped is not None: return anonymize(text_type(ipv6.ipv4_mapped)) return u'' + ipv6.exploded.rsplit(':', 5)[0] + ':' + ':'.join([ '0000'] * 5) except ipaddress.AddressValueError: return u'0.0.0.0'
anonymize
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(remote_addr):\n if not isinstance(remote_addr, text_type) and isinstance(remote_addr, str):\n remote_addr = remote_addr.decode('ascii', 'ignore')\n try:\n ipv4 = ipaddress.IPv4Address(remote_addr)\n return u''.join(ipv4.exploded.rsplit('.', 1)[0]) + '.' + '0'\n except ipaddress.AddressValueError:\n try:\n ipv6 = ipaddress.IPv6Address(remote_addr)\n if ipv6.ipv4_mapped is not None:\n return anonymize(text_type(ipv6.ipv4_mapped))\n return u'' + ipv6.exploded.rsplit(':', 5)[0] + ':' + ':'.join([\n '0000'] * 5)\n except ipaddress.AddressValueError:\n return u'0.0.0.0'\n", "role": "user" }, { "content": "anonymize", "role": "assistant" } ]
def x(data, default=u'Untitled.', id=None): html = html5lib.parse(data, treebuilder='dom') assert html.lastChild.nodeName == 'html' html = html.lastChild el = list(filter(lambda i: i.attributes['id'].value == 'isso-thread', filter(lambda i: 'id' in i.attributes, chain(*map(html. getElementsByTagName, ('div', 'section')))))) if not el: return id, default el = el[0] visited = [] def x(node): for child in node.childNodes: if child.nodeType != child.ELEMENT_NODE: continue if child.nodeName.upper() == 'H1': return child if child not in visited: return recurse(child) def x(rv): for child in rv.childNodes: if child.nodeType == child.TEXT_NODE: yield child.nodeValue if child.nodeType == child.ELEMENT_NODE: for item in gettext(child): yield item try: id = unquote(el.attributes['data-isso-id'].value) except (KeyError, AttributeError): pass try: return id, unquote(el.attributes['data-title'].value) except (KeyError, AttributeError): pass while el is not None: visited.append(el) rv = recurse(el) if rv: return id, ''.join(gettext(rv)).strip() el = el.parentNode return id, default
thread
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(data, default=u'Untitled.', id=None):\n html = html5lib.parse(data, treebuilder='dom')\n assert html.lastChild.nodeName == 'html'\n html = html.lastChild\n el = list(filter(lambda i: i.attributes['id'].value == 'isso-thread',\n filter(lambda i: 'id' in i.attributes, chain(*map(html.\n getElementsByTagName, ('div', 'section'))))))\n if not el:\n return id, default\n el = el[0]\n visited = []\n\n def x(node):\n for child in node.childNodes:\n if child.nodeType != child.ELEMENT_NODE:\n continue\n if child.nodeName.upper() == 'H1':\n return child\n if child not in visited:\n return recurse(child)\n\n def x(rv):\n for child in rv.childNodes:\n if child.nodeType == child.TEXT_NODE:\n yield child.nodeValue\n if child.nodeType == child.ELEMENT_NODE:\n for item in gettext(child):\n yield item\n try:\n id = unquote(el.attributes['data-isso-id'].value)\n except (KeyError, AttributeError):\n pass\n try:\n return id, unquote(el.attributes['data-title'].value)\n except (KeyError, AttributeError):\n pass\n while el is not None:\n visited.append(el)\n rv = recurse(el)\n if rv:\n return id, ''.join(gettext(rv)).strip()\n el = el.parentNode\n return id, default\n", "role": "user" }, { "content": "thread", "role": "assistant" } ]
def x(self, thread): thread_id = thread['id'] title = thread['title'] self.db.threads.new(thread_id, title) comments = list(map(self._build_comment, thread['comments'])) comments.sort(key=lambda comment: comment['id']) self.count += len(comments) for comment in comments: self.db.comments.add(thread_id, comment)
insert
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, thread):\n thread_id = thread['id']\n title = thread['title']\n self.db.threads.new(thread_id, title)\n comments = list(map(self._build_comment, thread['comments']))\n comments.sort(key=lambda comment: comment['id'])\n self.count += len(comments)\n for comment in comments:\n self.db.comments.add(thread_id, comment)\n", "role": "user" }, { "content": "insert", "role": "assistant" } ]
def x(self): with io.open(self.json_file, 'rt', encoding='utf8') as fh: threads = json.load(fh) progress = Progress(len(threads)) for i, thread in enumerate(threads): progress.update(i, str(i)) self.insert(thread) progress.finish('{0} threads, {1} comments'.format(len(threads), self. count))
migrate
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n with io.open(self.json_file, 'rt', encoding='utf8') as fh:\n threads = json.load(fh)\n progress = Progress(len(threads))\n for i, thread in enumerate(threads):\n progress.update(i, str(i))\n self.insert(thread)\n progress.finish('{0} threads, {1} comments'.format(len(threads), self.\n count))\n", "role": "user" }, { "content": "migrate", "role": "assistant" } ]
def x(self, uri, c): if c.get('parent') is not None: ref = self.get(c['parent']) if ref.get('parent') is not None: c['parent'] = ref['parent'] self.db.execute(['INSERT INTO comments (', ' tid, parent, created, modified, mode, remote_addr,', ' text, author, email, website, voters, notification)', 'SELECT', ' threads.id, ?,', ' ?, ?, ?, ?,', ' ?, ?, ?, ?, ?, ?', 'FROM threads WHERE threads.uri = ?;'], (c.get('parent'), c.get( 'created') or time.time(), None, c['mode'], c['remote_addr'], c[ 'text'], c.get('author'), c.get('email'), c.get('website'), buffer( Bloomfilter(iterable=[c['remote_addr']]).array), c.get( 'notification'), uri)) return dict(zip(Comments.fields, self.db.execute( 'SELECT *, MAX(c.id) FROM comments AS c INNER JOIN threads ON threads.uri = ?' , (uri,)).fetchone()))
add
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, uri, c):\n if c.get('parent') is not None:\n ref = self.get(c['parent'])\n if ref.get('parent') is not None:\n c['parent'] = ref['parent']\n self.db.execute(['INSERT INTO comments (',\n ' tid, parent, created, modified, mode, remote_addr,',\n ' text, author, email, website, voters, notification)', 'SELECT',\n ' threads.id, ?,', ' ?, ?, ?, ?,', ' ?, ?, ?, ?, ?, ?',\n 'FROM threads WHERE threads.uri = ?;'], (c.get('parent'), c.get(\n 'created') or time.time(), None, c['mode'], c['remote_addr'], c[\n 'text'], c.get('author'), c.get('email'), c.get('website'), buffer(\n Bloomfilter(iterable=[c['remote_addr']]).array), c.get(\n 'notification'), uri))\n return dict(zip(Comments.fields, self.db.execute(\n 'SELECT *, MAX(c.id) FROM comments AS c INNER JOIN threads ON threads.uri = ?'\n , (uri,)).fetchone()))\n", "role": "user" }, { "content": "add", "role": "assistant" } ]
def x(self, email, id): self.db.execute(['UPDATE comments SET', ' notification=0', 'WHERE email=? AND (id=? OR parent=?);'], (email, id, id))
unsubscribe
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, email, id):\n self.db.execute(['UPDATE comments SET', ' notification=0',\n 'WHERE email=? AND (id=? OR parent=?);'], (email, id, id))\n", "role": "user" }, { "content": "unsubscribe", "role": "assistant" } ]
def x(self, id, data): self.db.execute(['UPDATE comments SET', ','.join(key + '=' + '?' for key in data), 'WHERE id=?;'], list(data.values()) + [id]) return self.get(id)
update
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, id, data):\n self.db.execute(['UPDATE comments SET', ','.join(key + '=' + '?' for\n key in data), 'WHERE id=?;'], list(data.values()) + [id])\n return self.get(id)\n", "role": "user" }, { "content": "update", "role": "assistant" } ]
def x(self, id): rv = self.db.execute('SELECT * FROM comments WHERE id=?', (id,)).fetchone() if rv: return dict(zip(Comments.fields, rv)) return None
get
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, id):\n rv = self.db.execute('SELECT * FROM comments WHERE id=?', (id,)).fetchone()\n if rv:\n return dict(zip(Comments.fields, rv))\n return None\n", "role": "user" }, { "content": "get", "role": "assistant" } ]