source_code
stringlengths
45
7.24k
function_name
stringlengths
1
111
messages
list
def x(self): for node in self.iter_nodes(): for relation, target in self.relations_of(node.obj, True): yield node.obj, relation, target
list_relations
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n for node in self.iter_nodes():\n for relation, target in self.relations_of(node.obj, True):\n yield node.obj, relation, target\n", "role": "user" }, { "content": "list_relations", "role": "assistant" } ]
def x(self): for src_node in self.iter_nodes(): for relation in src_node.outgoing: for dst_node in src_node.outgoing[relation]: print(repr(src_node.obj), '-', relation, '-', repr(dst_node .obj))
show_relations
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n for src_node in self.iter_nodes():\n for relation in src_node.outgoing:\n for dst_node in src_node.outgoing[relation]:\n print(repr(src_node.obj), '-', relation, '-', repr(dst_node\n .obj))\n", "role": "user" }, { "content": "show_relations", "role": "assistant" } ]
def x(self, relation, filter_fn): assert type(relation).__name__ in {'str', 'unicode' }, 'where needs the first arg to be a string' assert callable(filter_fn), 'filter_fn needs to be callable' return VList(i for i in self if relation in i._relations() and any( filter_fn(_()) for _ in i[relation]))
where
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, relation, filter_fn):\n assert type(relation).__name__ in {'str', 'unicode'\n }, 'where needs the first arg to be a string'\n assert callable(filter_fn), 'filter_fn needs to be callable'\n return VList(i for i in self if relation in i._relations() and any(\n filter_fn(_()) for _ in i[relation]))\n", "role": "user" }, { "content": "where", "role": "assistant" } ]
def x(self, filter_fn): assert callable(filter_fn), 'filter_fn needs to be callable' return VList(i for i in self if filter_fn(i()))
where
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, filter_fn):\n assert callable(filter_fn), 'filter_fn needs to be callable'\n return VList(i for i in self if filter_fn(i()))\n", "role": "user" }, { "content": "where", "role": "assistant" } ]
def x(self, **kwargs): out = self for k, v in kwargs.items(): out = out.where(k, lambda i: i == v) return out
where
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, **kwargs):\n out = self\n for k, v in kwargs.items():\n out = out.where(k, lambda i: i == v)\n return out\n", "role": "user" }, { "content": "where", "role": "assistant" } ]
def x(path=''): from os.path import isfile if not isfile(path): from os import chmod from stat import S_IRUSR, S_IWUSR open(path, 'a').close() attempt(lambda : chmod(path, S_IRUSR | S_IWUSR))
create_file
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(path=''):\n from os.path import isfile\n if not isfile(path):\n from os import chmod\n from stat import S_IRUSR, S_IWUSR\n open(path, 'a').close()\n attempt(lambda : chmod(path, S_IRUSR | S_IWUSR))\n", "role": "user" }, { "content": "create_file", "role": "assistant" } ]
def x(self, item): item_id = self._id_of(item) if item_id is None: blob = self.serialize(item) with self._write_lock: self._execute('INSERT into objects (code) values (?);', (blob,)) self.autocommit()
store_item
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, item):\n item_id = self._id_of(item)\n if item_id is None:\n blob = self.serialize(item)\n with self._write_lock:\n self._execute('INSERT into objects (code) values (?);', (blob,))\n self.autocommit()\n", "role": "user" }, { "content": "store_item", "role": "assistant" } ]
def x(self, item): for relation in self.relations_of(item): self.delete_relation(item, relation) for origin, relation in self.relations_to(item, True): self.delete_relation(origin, relation, item) with self._write_lock: self._execute( """ DELETE from objects where code=? """ , (self.serialize(item),)) self.autocommit()
delete_item
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, item):\n for relation in self.relations_of(item):\n self.delete_relation(item, relation)\n for origin, relation in self.relations_to(item, True):\n self.delete_relation(origin, relation, item)\n with self._write_lock:\n self._execute(\n \"\"\"\n DELETE from objects where code=?\n \"\"\"\n , (self.serialize(item),))\n self.autocommit()\n", "role": "user" }, { "content": "delete_item", "role": "assistant" } ]
def x(self, src, name, dst): self.__require_string__(name) self.store_item(src) self.store_item(dst) with self._write_lock: self._execute( 'insert into relations select ob1.id, ?, ob2.id from objects as ob1, objects as ob2 where ob1.code=? and ob2.code=?;' , (name, self.serialize(src), self.serialize(dst))) self.autocommit()
store_relation
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, src, name, dst):\n self.__require_string__(name)\n self.store_item(src)\n self.store_item(dst)\n with self._write_lock:\n self._execute(\n 'insert into relations select ob1.id, ?, ob2.id from objects as ob1, objects as ob2 where ob1.code=? and ob2.code=?;'\n , (name, self.serialize(src), self.serialize(dst)))\n self.autocommit()\n", "role": "user" }, { "content": "store_relation", "role": "assistant" } ]
def x(self, src, relation, dst): self.__require_string__(relation) src_id = self._id_of(src) dst_id = self._id_of(dst) with self._write_lock: self._execute( """ DELETE from relations where src=? and name=? and dst=? """ , (src_id, relation, dst_id)) self.autocommit()
delete_single_relation
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, src, relation, dst):\n self.__require_string__(relation)\n src_id = self._id_of(src)\n dst_id = self._id_of(dst)\n with self._write_lock:\n self._execute(\n \"\"\"\n DELETE from relations where src=? and name=? and dst=?\n \"\"\"\n , (src_id, relation, dst_id))\n self.autocommit()\n", "role": "user" }, { "content": "delete_single_relation", "role": "assistant" } ]
def x(self, src, relation, *targets): self.__require_string__(relation) if len(targets): for i in targets: self._delete_single_relation(src, relation, i) else: for i in list(self.find(src, relation)): self._delete_single_relation(src, relation, i)
delete_relation
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, src, relation, *targets):\n self.__require_string__(relation)\n if len(targets):\n for i in targets:\n self._delete_single_relation(src, relation, i)\n else:\n for i in list(self.find(src, relation)):\n self._delete_single_relation(src, relation, i)\n", "role": "user" }, { "content": "delete_relation", "role": "assistant" } ]
def x(self, target, relation): query = ( 'select ob1.code from objects as ob1, objects as ob2, relations where relations.dst=ob1.id and relations.name=? and relations.src=ob2.id and ob2.code=?' ) for i in self._execute(query, (relation, self.serialize(target))): yield self.deserialize(i[0])
find
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, target, relation):\n query = (\n 'select ob1.code from objects as ob1, objects as ob2, relations where relations.dst=ob1.id and relations.name=? and relations.src=ob2.id and ob2.code=?'\n )\n for i in self._execute(query, (relation, self.serialize(target))):\n yield self.deserialize(i[0])\n", "role": "user" }, { "content": "find", "role": "assistant" } ]
def x(self, target, include_object=False): if include_object: _ = self._execute( """ select relations.name, ob2.code from relations, objects as ob1, objects as ob2 where relations.src=ob1.id and ob2.id=relations.dst and ob1.code=? """ , (self.serialize(target),)) for i in _: yield i[0], self.deserialize(i[1]) else: _ = self._execute( """ select distinct relations.name from relations, objects where relations.src=objects.id and objects.code=? """ , (self.serialize(target),)) for i in _: yield i[0]
relations_of
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, target, include_object=False):\n if include_object:\n _ = self._execute(\n \"\"\"\n select relations.name, ob2.code from relations, objects as ob1, objects as ob2 where relations.src=ob1.id and ob2.id=relations.dst and ob1.code=?\n \"\"\"\n , (self.serialize(target),))\n for i in _:\n yield i[0], self.deserialize(i[1])\n else:\n _ = self._execute(\n \"\"\"\n select distinct relations.name from relations, objects where relations.src=objects.id and objects.code=?\n \"\"\"\n , (self.serialize(target),))\n for i in _:\n yield i[0]\n", "role": "user" }, { "content": "relations_of", "role": "assistant" } ]
def x(self, target, include_object=False): if include_object: _ = self._execute( """ select name, (select code from objects where id=src) from relations where dst=? """ , (self._id_of(target),)) for i in _: yield self.deserialize(i[1]), i[0] else: _ = self._execute( """ select distinct name from relations where dst=? """ , (self._id_of(target),)) for i in _: yield i[0]
relations_to
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, target, include_object=False):\n if include_object:\n _ = self._execute(\n \"\"\"\n select name, (select code from objects where id=src) from relations where dst=?\n \"\"\"\n , (self._id_of(target),))\n for i in _:\n yield self.deserialize(i[1]), i[0]\n else:\n _ = self._execute(\n \"\"\"\n select distinct name from relations where dst=?\n \"\"\"\n , (self._id_of(target),))\n for i in _:\n yield i[0]\n", "role": "user" }, { "content": "relations_to", "role": "assistant" } ]
def x(self, target): return gen.chain(((r, i) for i in self.find(target, r)) for r in self. relations_of(target))
connections_of
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, target):\n return gen.chain(((r, i) for i in self.find(target, r)) for r in self.\n relations_of(target))\n", "role": "user" }, { "content": "connections_of", "role": "assistant" } ]
def x(self): for i in self._execute('select * from objects'): _id, code = i yield _id, code, self.deserialize(code)
list_objects
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n for i in self._execute('select * from objects'):\n _id, code = i\n yield _id, code, self.deserialize(code)\n", "role": "user" }, { "content": "list_objects", "role": "assistant" } ]
def x(self): _ = self._execute('select * from relations').fetchall() for i in _: src, name, dst = i src = self.deserialize(next(self._execute( 'select code from objects where id=?', (src,)))[0]) dst = self.deserialize(next(self._execute( 'select code from objects where id=?', (dst,)))[0]) yield src, name, dst
list_relations
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n _ = self._execute('select * from relations').fetchall()\n for i in _:\n src, name, dst = i\n src = self.deserialize(next(self._execute(\n 'select code from objects where id=?', (src,)))[0])\n dst = self.deserialize(next(self._execute(\n 'select code from objects where id=?', (dst,)))[0])\n yield src, name, dst\n", "role": "user" }, { "content": "list_relations", "role": "assistant" } ]
def x(db, ignore_tables=None): migrator = None if migrator is None: migrator = auto_detect_migrator(db) existing_tables = [unicode(t) for t in db.get_tables()] existing_indexes = {table: get_indexes_by_table(db, table) for table in existing_tables} existing_columns_by_table = get_columns_by_table(db) foreign_keys_by_table = get_foreign_keys_by_table(db) table_names_to_models = {_table_name(cls): cls for cls in all_models.keys() } to_run = [] table_adds, add_fks, table_deletes, table_renames = calc_table_changes( existing_tables, ignore_tables=ignore_tables) table_renamed_from = {v: k for k, v in table_renames.items()} for tbl in table_adds: to_run += create_table(table_names_to_models[tbl]) for field in add_fks: if hasattr(field, '__pwdbev__not_deferred' ) and field.__pwdbev__not_deferred: field.deferred = False to_run += create_foreign_key(field) for k, v in table_renames.items(): to_run += rename_table(migrator, k, v) rename_cols_by_table = {} deleted_cols_by_table = {} for etn, ecols in existing_columns_by_table.items(): if etn in table_deletes: continue ntn = table_renames.get(etn, etn) model = table_names_to_models.get(ntn) if not model: continue defined_fields = model._meta.sorted_fields defined_column_name_to_field = {unicode(_column_name(f)): f for f in defined_fields} existing_fks_by_column = {fk.column: fk for fk in foreign_keys_by_table[etn]} adds, deletes, renames, alter_statements = calc_column_changes(db, migrator, etn, ntn, ecols, defined_fields, existing_fks_by_column) for column_name in adds: field = defined_column_name_to_field[column_name] to_run += alter_add_column(db, migrator, ntn, column_name, field) if not field.null: if field.default is not None: to_run += set_default(db, migrator, ntn, column_name, field ) else: to_run.append(( '-- adding a not null column without a default will fail if the table is not empty' , [])) to_run += add_not_null(db, migrator, ntn, column_name, field) for column_name in deletes: fk = existing_fks_by_column.get(column_name) if fk: to_run += drop_foreign_key(db, migrator, ntn, fk.name) to_run += drop_column(db, migrator, ntn, column_name) for ocn, ncn in renames.items(): field = defined_column_name_to_field[ncn] to_run += rename_column(db, migrator, ntn, ocn, ncn, field) to_run += alter_statements rename_cols_by_table[ntn] = renames deleted_cols_by_table[ntn] = deletes for ntn, model in table_names_to_models.items(): etn = table_renamed_from.get(ntn, ntn) deletes = deleted_cols_by_table.get(ntn, set()) existing_indexes_for_table = [i for i in existing_indexes.get(etn, []) if not any([(c in deletes) for c in i.columns])] to_run += calc_index_changes(db, migrator, existing_indexes_for_table, model, rename_cols_by_table.get(ntn, {})) """ to_run += calc_perms_changes($schema_tables, noop) unless $check_perms_for.empty? """ for tbl in table_deletes: to_run += drop_table(migrator, tbl) return to_run
calc_changes
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(db, ignore_tables=None):\n migrator = None\n if migrator is None:\n migrator = auto_detect_migrator(db)\n existing_tables = [unicode(t) for t in db.get_tables()]\n existing_indexes = {table: get_indexes_by_table(db, table) for table in\n existing_tables}\n existing_columns_by_table = get_columns_by_table(db)\n foreign_keys_by_table = get_foreign_keys_by_table(db)\n table_names_to_models = {_table_name(cls): cls for cls in all_models.keys()\n }\n to_run = []\n table_adds, add_fks, table_deletes, table_renames = calc_table_changes(\n existing_tables, ignore_tables=ignore_tables)\n table_renamed_from = {v: k for k, v in table_renames.items()}\n for tbl in table_adds:\n to_run += create_table(table_names_to_models[tbl])\n for field in add_fks:\n if hasattr(field, '__pwdbev__not_deferred'\n ) and field.__pwdbev__not_deferred:\n field.deferred = False\n to_run += create_foreign_key(field)\n for k, v in table_renames.items():\n to_run += rename_table(migrator, k, v)\n rename_cols_by_table = {}\n deleted_cols_by_table = {}\n for etn, ecols in existing_columns_by_table.items():\n if etn in table_deletes:\n continue\n ntn = table_renames.get(etn, etn)\n model = table_names_to_models.get(ntn)\n if not model:\n continue\n defined_fields = model._meta.sorted_fields\n defined_column_name_to_field = {unicode(_column_name(f)): f for f in\n defined_fields}\n existing_fks_by_column = {fk.column: fk for fk in\n foreign_keys_by_table[etn]}\n adds, deletes, renames, alter_statements = calc_column_changes(db,\n migrator, etn, ntn, ecols, defined_fields, existing_fks_by_column)\n for column_name in adds:\n field = defined_column_name_to_field[column_name]\n to_run += alter_add_column(db, migrator, ntn, column_name, field)\n if not field.null:\n if field.default is not None:\n to_run += set_default(db, migrator, ntn, column_name, field\n )\n else:\n to_run.append((\n '-- adding a not null column without a default will fail if the table is not empty'\n , []))\n to_run += add_not_null(db, migrator, ntn, column_name, field)\n for column_name in deletes:\n fk = existing_fks_by_column.get(column_name)\n if fk:\n to_run += drop_foreign_key(db, migrator, ntn, fk.name)\n to_run += drop_column(db, migrator, ntn, column_name)\n for ocn, ncn in renames.items():\n field = defined_column_name_to_field[ncn]\n to_run += rename_column(db, migrator, ntn, ocn, ncn, field)\n to_run += alter_statements\n rename_cols_by_table[ntn] = renames\n deleted_cols_by_table[ntn] = deletes\n for ntn, model in table_names_to_models.items():\n etn = table_renamed_from.get(ntn, ntn)\n deletes = deleted_cols_by_table.get(ntn, set())\n existing_indexes_for_table = [i for i in existing_indexes.get(etn,\n []) if not any([(c in deletes) for c in i.columns])]\n to_run += calc_index_changes(db, migrator,\n existing_indexes_for_table, model, rename_cols_by_table.get(ntn,\n {}))\n \"\"\"\n to_run += calc_perms_changes($schema_tables, noop) unless $check_perms_for.empty?\n \"\"\"\n for tbl in table_deletes:\n to_run += drop_table(migrator, tbl)\n return to_run\n", "role": "user" }, { "content": "calc_changes", "role": "assistant" } ]
def x(self, sock, after_idle_sec=5, interval_sec=60, max_fails=5): if platform.system() == 'Darwin': TCP_KEEPALIVE = 16 sock.setsockopt(socket.SOL_SOCKET, socket.SO_KEEPALIVE, 1) sock.setsockopt(socket.IPPROTO_TCP, TCP_KEEPALIVE, interval_sec) if platform.system() == 'Windows': sock.ioctl(socket.SIO_KEEPALIVE_VALS, (1, 10000, 3000)) if platform.system() == 'Linux': sock.setsockopt(socket.SOL_SOCKET, socket.SO_KEEPALIVE, 1) sock.setsockopt(socket.IPPROTO_TCP, socket.TCP_KEEPIDLE, after_idle_sec ) sock.setsockopt(socket.IPPROTO_TCP, socket.TCP_KEEPINTVL, interval_sec) sock.setsockopt(socket.IPPROTO_TCP, socket.TCP_KEEPCNT, max_fails)
set_keep_alive
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, sock, after_idle_sec=5, interval_sec=60, max_fails=5):\n if platform.system() == 'Darwin':\n TCP_KEEPALIVE = 16\n sock.setsockopt(socket.SOL_SOCKET, socket.SO_KEEPALIVE, 1)\n sock.setsockopt(socket.IPPROTO_TCP, TCP_KEEPALIVE, interval_sec)\n if platform.system() == 'Windows':\n sock.ioctl(socket.SIO_KEEPALIVE_VALS, (1, 10000, 3000))\n if platform.system() == 'Linux':\n sock.setsockopt(socket.SOL_SOCKET, socket.SO_KEEPALIVE, 1)\n sock.setsockopt(socket.IPPROTO_TCP, socket.TCP_KEEPIDLE, after_idle_sec\n )\n sock.setsockopt(socket.IPPROTO_TCP, socket.TCP_KEEPINTVL, interval_sec)\n sock.setsockopt(socket.IPPROTO_TCP, socket.TCP_KEEPCNT, max_fails)\n", "role": "user" }, { "content": "set_keep_alive", "role": "assistant" } ]
def x(self, encoding='unicode'): buf_len = len(self.buf) replies = [] reply = b'' chop = 0 skip = 0 i = 0 buf_len = len(self.buf) for i in range(0, buf_len): ch = self.buf[i:i + 1] if skip: skip -= 1 i += 1 continue nxt = i + 1 if nxt < buf_len: if ch == b'\r' and self.buf[nxt:nxt + 1] == b'\n': if reply != b'': if encoding == 'unicode': replies.append(encode_str(reply, encoding)) else: replies.append(reply) reply = b'' chop = nxt + 1 skip = 1 i += 1 continue reply += ch i += 1 if chop: self.buf = self.buf[chop:] return replies
parse_buf
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, encoding='unicode'):\n buf_len = len(self.buf)\n replies = []\n reply = b''\n chop = 0\n skip = 0\n i = 0\n buf_len = len(self.buf)\n for i in range(0, buf_len):\n ch = self.buf[i:i + 1]\n if skip:\n skip -= 1\n i += 1\n continue\n nxt = i + 1\n if nxt < buf_len:\n if ch == b'\\r' and self.buf[nxt:nxt + 1] == b'\\n':\n if reply != b'':\n if encoding == 'unicode':\n replies.append(encode_str(reply, encoding))\n else:\n replies.append(reply)\n reply = b''\n chop = nxt + 1\n skip = 1\n i += 1\n continue\n reply += ch\n i += 1\n if chop:\n self.buf = self.buf[chop:]\n return replies\n", "role": "user" }, { "content": "parse_buf", "role": "assistant" } ]
def x(self, fixed_limit=None, encoding='unicode'): if not self.connected: return repeat = 1 wait = 0.2 chunk_no = 0 max_buf = self.max_buf max_chunks = self.max_chunks if fixed_limit is not None: max_buf = fixed_limit max_chunks = fixed_limit while repeat: chunk_size = self.chunk_size while True: buf_len = len(self.buf) if buf_len >= max_buf: break remaining = max_buf - buf_len if remaining < chunk_size: chunk_size = remaining if chunk_no >= max_chunks and not self.blocking: break try: chunk = self.s.recv(chunk_size) except socket.timeout as e: self.debug_print('Get chunks timed out.') self.debug_print(e) err = e.args[0] self.debug_print(err) if err == 'timed out': repeat = 0 break except ssl.SSLError as e: if e.errno == ssl.SSL_ERROR_WANT_READ: self.debug_print('SSL_ERROR_WANT_READ') break else: self.debug_print('Get chunks ssl error') self.close() return except socket.error as e: err = e.args[0] if err == errno.EAGAIN or err == errno.EWOULDBLOCK: break else: self.debug_print('get chunks other closing') self.close() return else: if chunk == b'': self.close() return self.buf += chunk if self.blocking: break chunk_no += 1 if not repeat: break else: repeat = 0 if self.blocking: if fixed_limit is None: if self.delimiter not in self.buf: repeat = 1 time.sleep(wait)
get_chunks
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, fixed_limit=None, encoding='unicode'):\n if not self.connected:\n return\n repeat = 1\n wait = 0.2\n chunk_no = 0\n max_buf = self.max_buf\n max_chunks = self.max_chunks\n if fixed_limit is not None:\n max_buf = fixed_limit\n max_chunks = fixed_limit\n while repeat:\n chunk_size = self.chunk_size\n while True:\n buf_len = len(self.buf)\n if buf_len >= max_buf:\n break\n remaining = max_buf - buf_len\n if remaining < chunk_size:\n chunk_size = remaining\n if chunk_no >= max_chunks and not self.blocking:\n break\n try:\n chunk = self.s.recv(chunk_size)\n except socket.timeout as e:\n self.debug_print('Get chunks timed out.')\n self.debug_print(e)\n err = e.args[0]\n self.debug_print(err)\n if err == 'timed out':\n repeat = 0\n break\n except ssl.SSLError as e:\n if e.errno == ssl.SSL_ERROR_WANT_READ:\n self.debug_print('SSL_ERROR_WANT_READ')\n break\n else:\n self.debug_print('Get chunks ssl error')\n self.close()\n return\n except socket.error as e:\n err = e.args[0]\n if err == errno.EAGAIN or err == errno.EWOULDBLOCK:\n break\n else:\n self.debug_print('get chunks other closing')\n self.close()\n return\n else:\n if chunk == b'':\n self.close()\n return\n self.buf += chunk\n if self.blocking:\n break\n chunk_no += 1\n if not repeat:\n break\n else:\n repeat = 0\n if self.blocking:\n if fixed_limit is None:\n if self.delimiter not in self.buf:\n repeat = 1\n time.sleep(wait)\n", "role": "user" }, { "content": "get_chunks", "role": "assistant" } ]
def x(self, node_ip, node_port=None, same_nodes=1): self.debug_print('Validating: ' + node_ip) if not is_ip_valid(node_ip) or node_ip == '0.0.0.0': self.debug_print('Invalid node ip in validate node') return 0 if node_port != 0 and node_port is not None: if not is_valid_port(node_port): self.debug_print('Invalid node port in validate port') return 0 """ Don't accept connections from self to passive server or connections to already connected nodes. """ if not self.enable_duplicate_ip_cons: if node_ip == '127.0.0.1' or node_ip == get_lan_ip(self.interface ) or node_ip == self.wan_ip: self.debug_print('Cannot connect to ourself.') return 0 if node_ip == self.passive_bind and node_port == self.passive_port: self.debug_print('Error connecting to same listen server.') return 0 if same_nodes: for node in (self.outbound + self.inbound): try: addr, port = node['con'].s.getpeername() if node_ip == addr: self.debug_print('Already connected to this node.') return 0 except Exception as e: print(e) return 0 return 1
validate_node
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, node_ip, node_port=None, same_nodes=1):\n self.debug_print('Validating: ' + node_ip)\n if not is_ip_valid(node_ip) or node_ip == '0.0.0.0':\n self.debug_print('Invalid node ip in validate node')\n return 0\n if node_port != 0 and node_port is not None:\n if not is_valid_port(node_port):\n self.debug_print('Invalid node port in validate port')\n return 0\n \"\"\"\n Don't accept connections from self to passive server\n or connections to already connected nodes.\n \"\"\"\n if not self.enable_duplicate_ip_cons:\n if node_ip == '127.0.0.1' or node_ip == get_lan_ip(self.interface\n ) or node_ip == self.wan_ip:\n self.debug_print('Cannot connect to ourself.')\n return 0\n if node_ip == self.passive_bind and node_port == self.passive_port:\n self.debug_print('Error connecting to same listen server.')\n return 0\n if same_nodes:\n for node in (self.outbound + self.inbound):\n try:\n addr, port = node['con'].s.getpeername()\n if node_ip == addr:\n self.debug_print('Already connected to this node.')\n return 0\n except Exception as e:\n print(e)\n return 0\n return 1\n", "role": "user" }, { "content": "validate_node", "role": "assistant" } ]
def x(self): if not self.enable_bootstrap: return None t = time.time() if self.last_bootstrap is not None: if t - self.last_bootstrap <= rendezvous_interval: self.debug_print('Bootstrapped recently') return None self.last_bootstrap = t self.debug_print('Searching for nodes to connect to.') try: connection_slots = self.max_outbound - len(self.outbound) if connection_slots > 0: rendezvous_con = self.rendezvous.server_connect() rendezvous_con.send_line('BOOTSTRAP ' + str(self.max_outbound * 2)) choices = rendezvous_con.recv_line(timeout=2) if choices == 'NODES EMPTY': rendezvous_con.close() self.debug_print('Node list is empty.') return self else: self.debug_print('Found node list.') choices = re.findall( '(?:(p|s)[:]([0-9]+[.][0-9]+[.][0-9]+[.][0-9]+)[:]([0-9]+))+\\s?' , choices) rendezvous_con.s.close() passive_nodes = [] for node in choices: if not connection_slots: break node_type, node_ip, node_port = node self.debug_print(str(node)) if node_type == 'p': passive_nodes.append(node) i = 0 while i < len(passive_nodes) and connection_slots > 0: node_type, node_ip, node_port = passive_nodes[i] con = self.add_node(node_ip, node_port, 'passive') if con is not None: connection_slots -= 1 self.debug_print('Con successful.') else: self.debug_print('Con failed.') i += 1 except Exception as e: self.debug_print('Unknown error in bootstrap()') error = parse_exception(e) log_exception(self.error_log_path, error) return self
bootstrap
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if not self.enable_bootstrap:\n return None\n t = time.time()\n if self.last_bootstrap is not None:\n if t - self.last_bootstrap <= rendezvous_interval:\n self.debug_print('Bootstrapped recently')\n return None\n self.last_bootstrap = t\n self.debug_print('Searching for nodes to connect to.')\n try:\n connection_slots = self.max_outbound - len(self.outbound)\n if connection_slots > 0:\n rendezvous_con = self.rendezvous.server_connect()\n rendezvous_con.send_line('BOOTSTRAP ' + str(self.max_outbound * 2))\n choices = rendezvous_con.recv_line(timeout=2)\n if choices == 'NODES EMPTY':\n rendezvous_con.close()\n self.debug_print('Node list is empty.')\n return self\n else:\n self.debug_print('Found node list.')\n choices = re.findall(\n '(?:(p|s)[:]([0-9]+[.][0-9]+[.][0-9]+[.][0-9]+)[:]([0-9]+))+\\\\s?'\n , choices)\n rendezvous_con.s.close()\n passive_nodes = []\n for node in choices:\n if not connection_slots:\n break\n node_type, node_ip, node_port = node\n self.debug_print(str(node))\n if node_type == 'p':\n passive_nodes.append(node)\n i = 0\n while i < len(passive_nodes) and connection_slots > 0:\n node_type, node_ip, node_port = passive_nodes[i]\n con = self.add_node(node_ip, node_port, 'passive')\n if con is not None:\n connection_slots -= 1\n self.debug_print('Con successful.')\n else:\n self.debug_print('Con failed.')\n i += 1\n except Exception as e:\n self.debug_print('Unknown error in bootstrap()')\n error = parse_exception(e)\n log_exception(self.error_log_path, error)\n return self\n", "role": "user" }, { "content": "bootstrap", "role": "assistant" } ]
def x(self): if not self.enable_advertise: self.debug_print('Advertise is disbled!') return None if self.net_type == 'direct' and self.node_type == 'passive': return None if not self.is_net_started: raise Exception('Please call start() before you call advertise()') t = time.time() if self.last_advertise is not None: if t - self.last_advertise <= advertise_interval: return None if len(self.inbound) >= self.min_connected: return None self.last_advertise = t try: if (self.node_type == 'passive' and self.passive_port is not None and self.enable_advertise): self.rendezvous.passive_listen(self.passive_port, self.max_inbound) """ Simultaneous open is only used as a fail-safe for connections to nodes on the direct_net and only direct_net can list itself as simultaneous so its safe to leave this enabled. """ if self.node_type == 'simultaneous': self.rendezvous.simultaneous_listen() except Exception as e: error = parse_exception(e) log_exception(self.error_log_path, error) return self
advertise
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if not self.enable_advertise:\n self.debug_print('Advertise is disbled!')\n return None\n if self.net_type == 'direct' and self.node_type == 'passive':\n return None\n if not self.is_net_started:\n raise Exception('Please call start() before you call advertise()')\n t = time.time()\n if self.last_advertise is not None:\n if t - self.last_advertise <= advertise_interval:\n return None\n if len(self.inbound) >= self.min_connected:\n return None\n self.last_advertise = t\n try:\n if (self.node_type == 'passive' and self.passive_port is not None and\n self.enable_advertise):\n self.rendezvous.passive_listen(self.passive_port, self.max_inbound)\n \"\"\"\n Simultaneous open is only used as a fail-safe for connections to\n nodes on the direct_net and only direct_net can list itself as\n simultaneous so its safe to leave this enabled.\n \"\"\"\n if self.node_type == 'simultaneous':\n self.rendezvous.simultaneous_listen()\n except Exception as e:\n error = parse_exception(e)\n log_exception(self.error_log_path, error)\n return self\n", "role": "user" }, { "content": "advertise", "role": "assistant" } ]
def x(self): if self.node_type == 'simultaneous': if self.nat_type != 'unknown': return 'simultaneous' unspecific_bind = ['0.0.0.0', '127.0.0.1', 'localhost'] if self.passive_bind in unspecific_bind: lan_ip = get_lan_ip(self.interface) else: lan_ip = self.passive_bind if (lan_ip is not None and self.passive_port is not None and self. enable_forwarding): self.debug_print('Checking if port is forwarded.') if is_port_forwarded(lan_ip, self.passive_port, 'TCP', self. forwarding_servers): msg = 'Port already forwarded. Skipping NAT traversal.' self.debug_print(msg) self.forwarding_type = 'forwarded' return 'passive' else: self.debug_print('Port is not already forwarded.') try: self.debug_print('Trying UPnP') UPnP(self.interface).forward_port('TCP', self.passive_port, lan_ip) if is_port_forwarded(lan_ip, self.passive_port, 'TCP', self. forwarding_servers): self.forwarding_type = 'UPnP' self.debug_print('Forwarded port with UPnP.') else: self.debug_print('UPnP failed to forward port.') except Exception as e: error = parse_exception(e) log_exception(self.error_log_path, error) self.debug_print('UPnP failed to forward port.') try: self.debug_print('Trying NATPMP.') NatPMP(self.interface).forward_port('TCP', self. passive_port, lan_ip) if is_port_forwarded(lan_ip, self.passive_port, 'TCP', self .forwarding_servers): self.forwarding_type = 'NATPMP' self.debug_print('Port forwarded with NATPMP.') else: self.debug_print('Failed to forward port with NATPMP.') self.debug_print( 'Falling back on TCP hole punching or proxying.') except Exception as e: error = parse_exception(e) log_exception(self.error_log_path, error) self.debug_print('Failed to forward port with NATPMP.') if self.forwarding_type != 'manual': return 'passive' if self.nat_type != 'unknown': return 'simultaneous' else: return 'active'
determine_node
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.node_type == 'simultaneous':\n if self.nat_type != 'unknown':\n return 'simultaneous'\n unspecific_bind = ['0.0.0.0', '127.0.0.1', 'localhost']\n if self.passive_bind in unspecific_bind:\n lan_ip = get_lan_ip(self.interface)\n else:\n lan_ip = self.passive_bind\n if (lan_ip is not None and self.passive_port is not None and self.\n enable_forwarding):\n self.debug_print('Checking if port is forwarded.')\n if is_port_forwarded(lan_ip, self.passive_port, 'TCP', self.\n forwarding_servers):\n msg = 'Port already forwarded. Skipping NAT traversal.'\n self.debug_print(msg)\n self.forwarding_type = 'forwarded'\n return 'passive'\n else:\n self.debug_print('Port is not already forwarded.')\n try:\n self.debug_print('Trying UPnP')\n UPnP(self.interface).forward_port('TCP', self.passive_port, lan_ip)\n if is_port_forwarded(lan_ip, self.passive_port, 'TCP', self.\n forwarding_servers):\n self.forwarding_type = 'UPnP'\n self.debug_print('Forwarded port with UPnP.')\n else:\n self.debug_print('UPnP failed to forward port.')\n except Exception as e:\n error = parse_exception(e)\n log_exception(self.error_log_path, error)\n self.debug_print('UPnP failed to forward port.')\n try:\n self.debug_print('Trying NATPMP.')\n NatPMP(self.interface).forward_port('TCP', self.\n passive_port, lan_ip)\n if is_port_forwarded(lan_ip, self.passive_port, 'TCP', self\n .forwarding_servers):\n self.forwarding_type = 'NATPMP'\n self.debug_print('Port forwarded with NATPMP.')\n else:\n self.debug_print('Failed to forward port with NATPMP.')\n self.debug_print(\n 'Falling back on TCP hole punching or proxying.')\n except Exception as e:\n error = parse_exception(e)\n log_exception(self.error_log_path, error)\n self.debug_print('Failed to forward port with NATPMP.')\n if self.forwarding_type != 'manual':\n return 'passive'\n if self.nat_type != 'unknown':\n return 'simultaneous'\n else:\n return 'active'\n", "role": "user" }, { "content": "determine_node", "role": "assistant" } ]
def x(self): self.debug_print('Starting networking.') self.debug_print( 'Make sure to iterate over replies if you need connection alive management!' ) signal.signal(signal.SIGINT, self.stop) self.debug_print('WAN IP = ' + str(self.wan_ip)) try: rendezvous_con = self.rendezvous.server_connect() rendezvous_con.close() except: raise Exception('Unable to connect to rendezvous server.') self.start_passive_server() if self.nat_type == 'unknown': self.debug_print('Determining NAT type.') nat_type = self.rendezvous.determine_nat() if nat_type is not None and nat_type != 'unknown': self.nat_type = nat_type self.rendezvous.nat_type = nat_type self.debug_print('NAT type = ' + nat_type) else: self.debug_print('Unable to determine NAT type.') if self.node_type == 'simultaneous': if self.nat_type not in self.rendezvous.predictable_nats: self.debug_print( 'Manual setting of simultanous specified but ignored since NAT does not support it.' ) self.node_type = 'active' else: self.debug_print('Determining node type.') if self.node_type == 'unknown': self.node_type = self.determine_node() if self.net_type == 'p2p': """ TCP hole punching is reserved specifically for direct networks (a net object reserved for receiving direct connections -- p2p is for connecting to the main network. The reason for this is you can't do multiple TCP hole punches at the same time so reserved for direct network where it's most needed. """ if self.node_type == 'simultaneous': self.debug_print('Simultaneous is not allowed for P2P') self.node_type = 'active' self.disable_simultaneous() self.debug_print('Node type = ' + self.node_type) self.close_cons() self.is_net_started = 1 self.unl = UNL(net=self, dht_node=self.dht_node, wan_ip=self.wan_ip) return self
start
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n self.debug_print('Starting networking.')\n self.debug_print(\n 'Make sure to iterate over replies if you need connection alive management!'\n )\n signal.signal(signal.SIGINT, self.stop)\n self.debug_print('WAN IP = ' + str(self.wan_ip))\n try:\n rendezvous_con = self.rendezvous.server_connect()\n rendezvous_con.close()\n except:\n raise Exception('Unable to connect to rendezvous server.')\n self.start_passive_server()\n if self.nat_type == 'unknown':\n self.debug_print('Determining NAT type.')\n nat_type = self.rendezvous.determine_nat()\n if nat_type is not None and nat_type != 'unknown':\n self.nat_type = nat_type\n self.rendezvous.nat_type = nat_type\n self.debug_print('NAT type = ' + nat_type)\n else:\n self.debug_print('Unable to determine NAT type.')\n if self.node_type == 'simultaneous':\n if self.nat_type not in self.rendezvous.predictable_nats:\n self.debug_print(\n 'Manual setting of simultanous specified but ignored since NAT does not support it.'\n )\n self.node_type = 'active'\n else:\n self.debug_print('Determining node type.')\n if self.node_type == 'unknown':\n self.node_type = self.determine_node()\n if self.net_type == 'p2p':\n \"\"\"\n TCP hole punching is reserved specifically for direct networks\n (a net object reserved for receiving direct connections\n -- p2p is for connecting to the main network. The reason for this\n is you can't do multiple TCP hole punches at the same time so\n reserved for direct network where it's most needed.\n \"\"\"\n if self.node_type == 'simultaneous':\n self.debug_print('Simultaneous is not allowed for P2P')\n self.node_type = 'active'\n self.disable_simultaneous()\n self.debug_print('Node type = ' + self.node_type)\n self.close_cons()\n self.is_net_started = 1\n self.unl = UNL(net=self, dht_node=self.dht_node, wan_ip=self.wan_ip)\n return self\n", "role": "user" }, { "content": "start", "role": "assistant" } ]
def x(self, signum=None, frame=None): self.debug_print('Stopping networking.') if self.passive is not None: try: self.passive.shutdown(1) except: pass self.passive.close() self.passive = None if self.last_advertise is not None: self.rendezvous.leave_fight() """ Just let the threads timeout by themselves. Otherwise mutex deadlocks could occur. for unl_thread in self.unl.unl_threads: unl_thread.exit() """ for con in self: con.close() if signum is not None: raise Exception('Process was interrupted.')
stop
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, signum=None, frame=None):\n self.debug_print('Stopping networking.')\n if self.passive is not None:\n try:\n self.passive.shutdown(1)\n except:\n pass\n self.passive.close()\n self.passive = None\n if self.last_advertise is not None:\n self.rendezvous.leave_fight()\n \"\"\"\n Just let the threads timeout by themselves.\n Otherwise mutex deadlocks could occur.\n for unl_thread in self.unl.unl_threads:\n unl_thread.exit()\n \"\"\"\n for con in self:\n con.close()\n if signum is not None:\n raise Exception('Process was interrupted.')\n", "role": "user" }, { "content": "stop", "role": "assistant" } ]
def x(self): msg = 'REMOTE TCP %s' % str(self.transport.getPeer().port) self.send_line(msg)
send_remote_port
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n msg = 'REMOTE TCP %s' % str(self.transport.getPeer().port)\n self.send_line(msg)\n", "role": "user" }, { "content": "send_remote_port", "role": "assistant" } ]
def x(self, node_ip): if node_ip in self.factory.candidates: old_candidates = [] for candidate in self.factory.candidates[node_ip]: elapsed = int(time.time() - candidate['time']) if elapsed > self.challege_timeout: old_candidates.append(candidate) for candidate in old_candidates: self.factory.candidates[node_ip].remove(candidate)
cleanup_candidates
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, node_ip):\n if node_ip in self.factory.candidates:\n old_candidates = []\n for candidate in self.factory.candidates[node_ip]:\n elapsed = int(time.time() - candidate['time'])\n if elapsed > self.challege_timeout:\n old_candidates.append(candidate)\n for candidate in old_candidates:\n self.factory.candidates[node_ip].remove(candidate)\n", "role": "user" }, { "content": "cleanup_candidates", "role": "assistant" } ]
def x(self, node_ip): if node_ip in self.factory.candidates: old_candidates = [] for candidate in self.factory.candidates[node_ip]: if not candidate['con'].connected: continue if candidate['propogated']: continue msg = 'CHALLENGE %s %s %s' % (candidate['ip_addr'], ' '.join( map(str, candidate['predictions'])), candidate['proto']) self.factory.nodes['simultaneous'][node_ip]['con'].send_line(msg) old_candidates.append(candidate)
propogate_candidates
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, node_ip):\n if node_ip in self.factory.candidates:\n old_candidates = []\n for candidate in self.factory.candidates[node_ip]:\n if not candidate['con'].connected:\n continue\n if candidate['propogated']:\n continue\n msg = 'CHALLENGE %s %s %s' % (candidate['ip_addr'], ' '.join(\n map(str, candidate['predictions'])), candidate['proto'])\n self.factory.nodes['simultaneous'][node_ip]['con'].send_line(msg)\n old_candidates.append(candidate)\n", "role": "user" }, { "content": "propogate_candidates", "role": "assistant" } ]
def x(self, node_ip): for candidate in self.factory.candidates[node_ip]: if not candidate['con'].connected: continue if candidate['time'] - self.factory.nodes['simultaneous'][node_ip][ 'time'] > self.challege_timeout: msg = 'RECONNECT' self.factory.nodes['simultaneous'][node_ip]['con'].send_line(msg) return self.cleanup_candidates(node_ip) self.propogate_candidates(node_ip)
synchronize_simultaneous
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, node_ip):\n for candidate in self.factory.candidates[node_ip]:\n if not candidate['con'].connected:\n continue\n if candidate['time'] - self.factory.nodes['simultaneous'][node_ip][\n 'time'] > self.challege_timeout:\n msg = 'RECONNECT'\n self.factory.nodes['simultaneous'][node_ip]['con'].send_line(msg)\n return\n self.cleanup_candidates(node_ip)\n self.propogate_candidates(node_ip)\n", "role": "user" }, { "content": "synchronize_simultaneous", "role": "assistant" } ]
def x(self, reason): try: self.connected = False if debug: print(self.log_entry('CLOSED =', 'none')) t = time.time() if time.time() - self.factory.last_cleanup >= self.cleanup: self.factory.last_cleanup = t old_node_ips = [] for node_ip in list(self.factory.nodes['passive']): passive_node = self.factory.nodes['passive'][node_ip] if t - passive_node['time'] >= self.node_lifetime: old_node_ips.append(node_ip) for node_ip in old_node_ips: del self.factory.nodes['passive'][node_ip] old_node_ips = [] for node_ip in list(self.factory.nodes['simultaneous']): simultaneous_node = self.factory.nodes['simultaneous'][node_ip] if t - simultaneous_node['time'] >= self.node_lifetime: old_node_ips.append(node_ip) for node_ip in old_node_ips: del self.factory.nodes['simultaneous'][node_ip] old_node_ips = [] for node_ip in list(self.factory.candidates): old_candidates = [] for candidate in self.factory.candidates[node_ip]: if node_ip not in self.factory.nodes['simultaneous' ] and t - candidate['time' ] >= self.challenge_timeout * 5: old_candidates.append(candidate) for candidate in old_candidates: self.factory.candidates[node_ip].remove(candidate) if not len(self.factory.candidates[node_ip] ) and node_ip not in self.factory.nodes['simultaneous']: old_node_ips.append(node_ip) for node_ip in old_node_ips: del self.factory.candidates[node_ip] except Exception as e: error = parse_exception(e) log_exception(error_log_path, error) print(self.log_entry('ERROR =', error))
connectionLost
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, reason):\n try:\n self.connected = False\n if debug:\n print(self.log_entry('CLOSED =', 'none'))\n t = time.time()\n if time.time() - self.factory.last_cleanup >= self.cleanup:\n self.factory.last_cleanup = t\n old_node_ips = []\n for node_ip in list(self.factory.nodes['passive']):\n passive_node = self.factory.nodes['passive'][node_ip]\n if t - passive_node['time'] >= self.node_lifetime:\n old_node_ips.append(node_ip)\n for node_ip in old_node_ips:\n del self.factory.nodes['passive'][node_ip]\n old_node_ips = []\n for node_ip in list(self.factory.nodes['simultaneous']):\n simultaneous_node = self.factory.nodes['simultaneous'][node_ip]\n if t - simultaneous_node['time'] >= self.node_lifetime:\n old_node_ips.append(node_ip)\n for node_ip in old_node_ips:\n del self.factory.nodes['simultaneous'][node_ip]\n old_node_ips = []\n for node_ip in list(self.factory.candidates):\n old_candidates = []\n for candidate in self.factory.candidates[node_ip]:\n if node_ip not in self.factory.nodes['simultaneous'\n ] and t - candidate['time'\n ] >= self.challenge_timeout * 5:\n old_candidates.append(candidate)\n for candidate in old_candidates:\n self.factory.candidates[node_ip].remove(candidate)\n if not len(self.factory.candidates[node_ip]\n ) and node_ip not in self.factory.nodes['simultaneous']:\n old_node_ips.append(node_ip)\n for node_ip in old_node_ips:\n del self.factory.candidates[node_ip]\n except Exception as e:\n error = parse_exception(e)\n log_exception(error_log_path, error)\n print(self.log_entry('ERROR =', error))\n", "role": "user" }, { "content": "connectionLost", "role": "assistant" } ]
def x(self): random.shuffle(self.server_list) myip = '' for server in self.server_list[:3]: myip = self.fetch(server) if myip != '': return myip else: continue return ''
get_external_ip
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n random.shuffle(self.server_list)\n myip = ''\n for server in self.server_list[:3]:\n myip = self.fetch(server)\n if myip != '':\n return myip\n else:\n continue\n return ''\n", "role": "user" }, { "content": "get_external_ip", "role": "assistant" } ]
def x(self, server): t = None socket_default_timeout = socket.getdefaulttimeout() opener = urllib.build_opener() opener.addheaders = [('User-agent', 'Mozilla/5.0 (X11; Linux x86_64; rv:24.0) Gecko/20100101 Firefox/24.0') ] try: t = Timer(self.timeout, self.handle_timeout, [self.url]) t.start() if version_info[0:2] == (2, 5): socket.setdefaulttimeout(self.timeout) self.url = opener.open(server) else: self.url = opener.open(server, timeout=self.timeout) content = self.url.read() if PY3K: try: content = content.decode('UTF-8') except UnicodeDecodeError: content = content.decode('ISO-8859-1') p = '(25[0-5]|2[0-4][0-9]|[01]?[0-9][0-9]?)\\.(' p += '25[0-5]|2[0-4][0-9]|[01]?[0-9][0-9]?)\\.(25[0-5]|2[0-4][0-9]|[' p += '01]?[0-9][0-9]?)\\.(25[0-5]|2[0-4][0-9]|[01]?[0-9][0-9]?)' m = re.search(p, content) myip = m.group(0) if len(myip) > 0: return myip else: return '' except Exception as e: print(e) return '' finally: if self.url is not None: self.url.close() self.url = None if t is not None: t.cancel() if socket.getdefaulttimeout() != socket_default_timeout: socket.setdefaulttimeout(socket_default_timeout)
fetch
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, server):\n t = None\n socket_default_timeout = socket.getdefaulttimeout()\n opener = urllib.build_opener()\n opener.addheaders = [('User-agent',\n 'Mozilla/5.0 (X11; Linux x86_64; rv:24.0) Gecko/20100101 Firefox/24.0')\n ]\n try:\n t = Timer(self.timeout, self.handle_timeout, [self.url])\n t.start()\n if version_info[0:2] == (2, 5):\n socket.setdefaulttimeout(self.timeout)\n self.url = opener.open(server)\n else:\n self.url = opener.open(server, timeout=self.timeout)\n content = self.url.read()\n if PY3K:\n try:\n content = content.decode('UTF-8')\n except UnicodeDecodeError:\n content = content.decode('ISO-8859-1')\n p = '(25[0-5]|2[0-4][0-9]|[01]?[0-9][0-9]?)\\\\.('\n p += '25[0-5]|2[0-4][0-9]|[01]?[0-9][0-9]?)\\\\.(25[0-5]|2[0-4][0-9]|['\n p += '01]?[0-9][0-9]?)\\\\.(25[0-5]|2[0-4][0-9]|[01]?[0-9][0-9]?)'\n m = re.search(p, content)\n myip = m.group(0)\n if len(myip) > 0:\n return myip\n else:\n return ''\n except Exception as e:\n print(e)\n return ''\n finally:\n if self.url is not None:\n self.url.close()\n self.url = None\n if t is not None:\n t.cancel()\n if socket.getdefaulttimeout() != socket_default_timeout:\n socket.setdefaulttimeout(socket_default_timeout)\n", "role": "user" }, { "content": "fetch", "role": "assistant" } ]
def x(self, their_unl, events, force_master, hairpin, nonce): our_unl = self.value.encode('ascii') their_unl = their_unl.encode('ascii') master = self.is_master(their_unl) """ Master defines who connects if either side can. It's used to eliminate having multiple connections with the same host. """ if force_master: master = 1 our_unl = self.deconstruct(our_unl) their_unl = self.deconstruct(their_unl) if our_unl is None: raise Exception('Unable to deconstruct our UNL.') if their_unl is None: raise Exception('Unable to deconstruct their UNL.') if our_unl['wan_ip'] == their_unl['wan_ip']: our_unl['wan_ip'] = our_unl['lan_ip'] their_unl['wan_ip'] = their_unl['lan_ip'] if hairpin: our_unl['node_type'] = 'passive' their_unl['node_type'] = 'passive' if nonce != '0' * 64: if sys.version_info >= (3, 0, 0): if type(nonce) == str: nonce.encode('ascii') elif type(nonce) == unicode: nonce = str(nonce) assert len(nonce) == 64 con_id = self.net.generate_con_id(nonce, our_unl['wan_ip'], their_unl['wan_ip']) else: con_id = None self.mutex.acquire() end_time = time.time() end_time += len(self.pending_unls) * 60 self.debug_print('Waiting for other unls to finish') while their_unl in self.pending_unls and time.time() < end_time: if events is None: self.mutex.release() return time.sleep(1) self.debug_print('Other unl finished') is_exception = 0 try: if their_unl['node_type'] == 'simultaneous' and our_unl['node_type' ] != 'passive': self.pending_sim_open.append(their_unl['value']) end_time = time.time() end_time += len(self.pending_unls) * 60 self.debug_print('wait for other hole punches to finish') while len(self.pending_sim_open) and time.time() < end_time: if self.pending_sim_open[0] == their_unl['value']: break time.sleep(1) self.debug_print('other hole punches finished') self.pending_unls.append(their_unl) self.mutex.release() con = self.get_connection(our_unl, their_unl, master, nonce, force_master, con_id) except Exception as e: is_exception = 1 print(e) print('EXCEPTION IN UNL.GET_CONNECTION') log_exception('error.log', parse_exception(e)) finally: if self.mutex.locked() and is_exception: self.mutex.release() if their_unl in self.pending_unls: self.pending_unls.remove(their_unl) if len(self.pending_sim_open): if self.pending_sim_open[0] == their_unl['value']: self.pending_sim_open = self.pending_sim_open[1:] if events is not None: if con is not None: if 'success' in events: events['success'](con) if con is None: if 'failure' in events: events['failure'](con)
connect_handler
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, their_unl, events, force_master, hairpin, nonce):\n our_unl = self.value.encode('ascii')\n their_unl = their_unl.encode('ascii')\n master = self.is_master(their_unl)\n \"\"\"\n Master defines who connects if either side can. It's used to\n eliminate having multiple connections with the same host.\n \"\"\"\n if force_master:\n master = 1\n our_unl = self.deconstruct(our_unl)\n their_unl = self.deconstruct(their_unl)\n if our_unl is None:\n raise Exception('Unable to deconstruct our UNL.')\n if their_unl is None:\n raise Exception('Unable to deconstruct their UNL.')\n if our_unl['wan_ip'] == their_unl['wan_ip']:\n our_unl['wan_ip'] = our_unl['lan_ip']\n their_unl['wan_ip'] = their_unl['lan_ip']\n if hairpin:\n our_unl['node_type'] = 'passive'\n their_unl['node_type'] = 'passive'\n if nonce != '0' * 64:\n if sys.version_info >= (3, 0, 0):\n if type(nonce) == str:\n nonce.encode('ascii')\n elif type(nonce) == unicode:\n nonce = str(nonce)\n assert len(nonce) == 64\n con_id = self.net.generate_con_id(nonce, our_unl['wan_ip'],\n their_unl['wan_ip'])\n else:\n con_id = None\n self.mutex.acquire()\n end_time = time.time()\n end_time += len(self.pending_unls) * 60\n self.debug_print('Waiting for other unls to finish')\n while their_unl in self.pending_unls and time.time() < end_time:\n if events is None:\n self.mutex.release()\n return\n time.sleep(1)\n self.debug_print('Other unl finished')\n is_exception = 0\n try:\n if their_unl['node_type'] == 'simultaneous' and our_unl['node_type'\n ] != 'passive':\n self.pending_sim_open.append(their_unl['value'])\n end_time = time.time()\n end_time += len(self.pending_unls) * 60\n self.debug_print('wait for other hole punches to finish')\n while len(self.pending_sim_open) and time.time() < end_time:\n if self.pending_sim_open[0] == their_unl['value']:\n break\n time.sleep(1)\n self.debug_print('other hole punches finished')\n self.pending_unls.append(their_unl)\n self.mutex.release()\n con = self.get_connection(our_unl, their_unl, master, nonce,\n force_master, con_id)\n except Exception as e:\n is_exception = 1\n print(e)\n print('EXCEPTION IN UNL.GET_CONNECTION')\n log_exception('error.log', parse_exception(e))\n finally:\n if self.mutex.locked() and is_exception:\n self.mutex.release()\n if their_unl in self.pending_unls:\n self.pending_unls.remove(their_unl)\n if len(self.pending_sim_open):\n if self.pending_sim_open[0] == their_unl['value']:\n self.pending_sim_open = self.pending_sim_open[1:]\n if events is not None:\n if con is not None:\n if 'success' in events:\n events['success'](con)\n if con is None:\n if 'failure' in events:\n events['failure'](con)\n", "role": "user" }, { "content": "connect_handler", "role": "assistant" } ]
def x(self, their_unl, events, force_master=1, hairpin=1, nonce='0' * 64): parms = their_unl, events, force_master, hairpin, nonce t = Thread(target=self.connect_handler, args=parms) t.start() self.unl_threads.append(t)
connect
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, their_unl, events, force_master=1, hairpin=1, nonce='0' * 64):\n parms = their_unl, events, force_master, hairpin, nonce\n t = Thread(target=self.connect_handler, args=parms)\n t.start()\n self.unl_threads.append(t)\n", "role": "user" }, { "content": "connect", "role": "assistant" } ]
def x(self): n = self.statx_n(self.data_points) """ Required to be able to compute the standard deviation. """ if n < 1: return Decimal('0') avg = self.statx_avg(self.data_points) sdev = self.statx_sdev(self.data_points) """ Incrementally remove aberration points. """ for k in range(0, self.clean_steps): """ Remove aberration points: keep only the sigma range around the average. """ min_val = avg - sdev max_val = avg + sdev cleaned_data_points = [] for i in range(0, n): v = self.data_points[i] if v < min_val or v > max_val: continue cleaned_data_points.append(v) self.data_points = cleaned_data_points[:] """ Recompute the new average using the "sound" points we kept. """ n = self.statx_n(self.data_points) """ Not enough data to compute standard deviation. """ if n < 2: break avg = self.statx_avg(self.data_points) sdev = self.statx_sdev(self.data_points) if sdev <= self.max_sdev or n < self.min_data: break """ If standard deviation is too large still, we cannot update our clock. Collect more points. If we don't have a minimum amount of data, don't attempt the update yet, continue collecting. """ if sdev > self.max_sdev or n < self.min_data: return Decimal('0') return avg
calculate_clock_skew
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n n = self.statx_n(self.data_points)\n \"\"\"\n Required to be able to compute the standard\n deviation.\n \"\"\"\n if n < 1:\n return Decimal('0')\n avg = self.statx_avg(self.data_points)\n sdev = self.statx_sdev(self.data_points)\n \"\"\"\n Incrementally remove aberration points.\n \"\"\"\n for k in range(0, self.clean_steps):\n \"\"\"\n Remove aberration points: keep only\n the sigma range around the average.\n \"\"\"\n min_val = avg - sdev\n max_val = avg + sdev\n cleaned_data_points = []\n for i in range(0, n):\n v = self.data_points[i]\n if v < min_val or v > max_val:\n continue\n cleaned_data_points.append(v)\n self.data_points = cleaned_data_points[:]\n \"\"\"\n Recompute the new average using the\n \"sound\" points we kept.\n \"\"\"\n n = self.statx_n(self.data_points)\n \"\"\"\n Not enough data to compute standard\n deviation.\n \"\"\"\n if n < 2:\n break\n avg = self.statx_avg(self.data_points)\n sdev = self.statx_sdev(self.data_points)\n if sdev <= self.max_sdev or n < self.min_data:\n break\n \"\"\"\n If standard deviation is too large still, we\n cannot update our clock. Collect more points.\n\n If we don't have a minimum amount of data,\n don't attempt the update yet, continue collecting.\n \"\"\"\n if sdev > self.max_sdev or n < self.min_data:\n return Decimal('0')\n return avg\n", "role": "user" }, { "content": "calculate_clock_skew", "role": "assistant" } ]
def x(self, mappings, node_ip, predictions, ntp): mappings = self.add_listen_sock(mappings) log.debug(mappings) self.simultaneous_cons = [] predictions = predictions.split(' ') self.simultaneous_fight(mappings, node_ip, predictions, ntp) if len(self.simultaneous_cons): """ There may be a problem here. I noticed that when these lines were removed during testing that connections tended to succeed more. There may be a lack of synchronization between the timing for connections to succeed so that a close on one side of the fight ends up ruining valid connections on this side. Will need to test more. Notes: the UNL synchronization code could actually fix this (potential) problem as a cool unintended side-effect. """ """ for i in range(1, len(self.simultaneous_cons)): try: print("Closing unneeded hole") #self.simultaneous_cons[i].s.close() except: pass """ try: return self.simultaneous_cons[0] except: log.debug('No holes found') for mapping in mappings: s = mapping['listen'] r, w, e = select.select([s], [], [], 0) for found_sock in r: if found_sock != s: continue log.debug('Accept logic works!') client, address = s.accept() con = Sock(blocking=0) con.set_sock(client) return con return None
attend_fight
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, mappings, node_ip, predictions, ntp):\n mappings = self.add_listen_sock(mappings)\n log.debug(mappings)\n self.simultaneous_cons = []\n predictions = predictions.split(' ')\n self.simultaneous_fight(mappings, node_ip, predictions, ntp)\n if len(self.simultaneous_cons):\n \"\"\"\n There may be a problem here. I noticed that when these lines\n were removed during testing that connections tended to\n succeed more. There may be a lack of synchronization between\n the timing for connections to succeed so that a close on\n one side of the fight ends up ruining valid connections on\n this side. Will need to test more.\n\n Notes: the UNL synchronization code could actually fix\n this (potential) problem as a cool unintended side-effect.\n \"\"\"\n \"\"\"\n for i in range(1, len(self.simultaneous_cons)):\n try:\n print(\"Closing unneeded hole\")\n #self.simultaneous_cons[i].s.close()\n except:\n pass\n \"\"\"\n try:\n return self.simultaneous_cons[0]\n except:\n log.debug('No holes found')\n for mapping in mappings:\n s = mapping['listen']\n r, w, e = select.select([s], [], [], 0)\n for found_sock in r:\n if found_sock != s:\n continue\n log.debug('Accept logic works!')\n client, address = s.accept()\n con = Sock(blocking=0)\n con.set_sock(client)\n return con\n return None\n", "role": "user" }, { "content": "attend_fight", "role": "assistant" } ]
def x(self): try: mappings = sequential_bind(self.mapping_no + 1, self.interface) con = self.server_connect(mappings[0]['sock']) except Exception as e: log.debug(e) log.debug('this err') return None mappings.remove(mappings[0]) msg = 'SOURCE TCP %s' % str(mappings[0]['source']) con.send_line(msg) reply = con.recv_line(timeout=2) remote_port = self.parse_remote_port(reply) if not remote_port: return None predictions = '' if self.nat_type != 'random': mappings = self.predict_mappings(mappings) for mapping in mappings: predictions += str(mapping['remote']) + ' ' predictions = predictions.rstrip() else: predictions = '1337' return [con, mappings, predictions]
sequential_connect
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n try:\n mappings = sequential_bind(self.mapping_no + 1, self.interface)\n con = self.server_connect(mappings[0]['sock'])\n except Exception as e:\n log.debug(e)\n log.debug('this err')\n return None\n mappings.remove(mappings[0])\n msg = 'SOURCE TCP %s' % str(mappings[0]['source'])\n con.send_line(msg)\n reply = con.recv_line(timeout=2)\n remote_port = self.parse_remote_port(reply)\n if not remote_port:\n return None\n predictions = ''\n if self.nat_type != 'random':\n mappings = self.predict_mappings(mappings)\n for mapping in mappings:\n predictions += str(mapping['remote']) + ' '\n predictions = predictions.rstrip()\n else:\n predictions = '1337'\n return [con, mappings, predictions]\n", "role": "user" }, { "content": "sequential_connect", "role": "assistant" } ]
def x(self): if self.server_con is not None: self.server_con.s.close() self.server_con = None self.mappings = None self.predictions = None parts = self.sequential_connect() if parts is None: return 0 con, mappings, predictions = parts con.blocking = 0 con.timeout = 0 con.s.settimeout(0) self.server_con = con self.mappings = mappings self.predictions = predictions msg = 'SIMULTANEOUS READY 0 0' ret = self.server_con.send_line(msg) if not ret: return 0 return 1
simultaneous_listen
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.server_con is not None:\n self.server_con.s.close()\n self.server_con = None\n self.mappings = None\n self.predictions = None\n parts = self.sequential_connect()\n if parts is None:\n return 0\n con, mappings, predictions = parts\n con.blocking = 0\n con.timeout = 0\n con.s.settimeout(0)\n self.server_con = con\n self.mappings = mappings\n self.predictions = predictions\n msg = 'SIMULTANEOUS READY 0 0'\n ret = self.server_con.send_line(msg)\n if not ret:\n return 0\n return 1\n", "role": "user" }, { "content": "simultaneous_listen", "role": "assistant" } ]
def x(self, mappings): if self.nat_type not in self.predictable_nats: msg = "Can't predict mappings for non-predictable NAT type." raise Exception(msg) for mapping in mappings: mapping['bound'] = mapping['sock'].getsockname()[1] if self.nat_type == 'preserving': mapping['remote'] = mapping['source'] if self.nat_type == 'delta': max_port = 65535 mapping['remote'] = int(mapping['source']) + self.delta if mapping['remote'] > max_port: mapping['remote'] -= max_port if mapping['remote'] < 0: mapping['remote'] = max_port - -mapping['remote'] if mapping['remote'] < 1 or mapping['remote'] > max_port: mapping['remote'] = 1 mapping['remote'] = str(mapping['remote']) return mappings
predict_mappings
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, mappings):\n if self.nat_type not in self.predictable_nats:\n msg = \"Can't predict mappings for non-predictable NAT type.\"\n raise Exception(msg)\n for mapping in mappings:\n mapping['bound'] = mapping['sock'].getsockname()[1]\n if self.nat_type == 'preserving':\n mapping['remote'] = mapping['source']\n if self.nat_type == 'delta':\n max_port = 65535\n mapping['remote'] = int(mapping['source']) + self.delta\n if mapping['remote'] > max_port:\n mapping['remote'] -= max_port\n if mapping['remote'] < 0:\n mapping['remote'] = max_port - -mapping['remote']\n if mapping['remote'] < 1 or mapping['remote'] > max_port:\n mapping['remote'] = 1\n mapping['remote'] = str(mapping['remote'])\n return mappings\n", "role": "user" }, { "content": "predict_mappings", "role": "assistant" } ]
def x(self, args, tries=1): if len(args) != 3: return 0 sock, node_ip, remote_port = args if sock is None or node_ip is None or remote_port is None: return 0 con = Sock(blocking=1, interface=self.interface) con.set_sock(sock) local = 0 if is_ip_private(node_ip): """ When simulating nodes on the same computer a delay needs to be set for the loop back interface to simulate the delays that occur over a WAN link. This requirement may also be needed for nodes on a LAN. sudo tc qdisc replace dev lo root handle 1:0 netem delay 0.5sec Speculation: The simulation problem may be to do with CPU cores. If the program is run on the same core then the connects will always be out of sync. If that's the case -- tries will need to be set to ~1000 which was what it was before. Perhaps a delay could be simulated by sleeping for random periods if its a local connection? That could help punch through at least once and then just set the tries to >= 1000. """ tries = 20 local = 1 source_port = sock.getsockname()[1] error = 0 log.debug('Throwing punch') for i in range(0, tries): try: con.connect(node_ip, remote_port) log.debug('Sim open success!') con.set_blocking(blocking=0, timeout=5) self.simultaneous_cons.append(con) return 1 except Exception as e: e = str(parse_exception(e)) log.debug(e) error = 1 continue if error: sock.close() return 0
throw_punch
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, args, tries=1):\n if len(args) != 3:\n return 0\n sock, node_ip, remote_port = args\n if sock is None or node_ip is None or remote_port is None:\n return 0\n con = Sock(blocking=1, interface=self.interface)\n con.set_sock(sock)\n local = 0\n if is_ip_private(node_ip):\n \"\"\"\n When simulating nodes on the same computer a delay needs to be set\n for the loop back interface to simulate the delays that occur over\n a WAN link. This requirement may also be needed for nodes on a LAN.\n\n sudo tc qdisc replace dev lo root handle 1:0 netem delay 0.5sec\n\n Speculation: The simulation problem may be to do with CPU cores.\n If the program is run on the same core then the connects will always\n be out of sync. If that's the case -- tries will need to be set to\n ~1000 which was what it was before. Perhaps a delay could be\n simulated by sleeping for random periods if its a local connection?\n That could help punch through at least once and then just set the\n tries to >= 1000.\n \"\"\"\n tries = 20\n local = 1\n source_port = sock.getsockname()[1]\n error = 0\n log.debug('Throwing punch')\n for i in range(0, tries):\n try:\n con.connect(node_ip, remote_port)\n log.debug('Sim open success!')\n con.set_blocking(blocking=0, timeout=5)\n self.simultaneous_cons.append(con)\n return 1\n except Exception as e:\n e = str(parse_exception(e))\n log.debug(e)\n error = 1\n continue\n if error:\n sock.close()\n return 0\n", "role": "user" }, { "content": "throw_punch", "role": "assistant" } ]
def x(self, my_mappings, node_ip, predictions, origin_ntp): p = request_priority_execution() log.debug('Getting NTP') if self.sys_clock is not None: our_ntp = self.sys_clock.time() else: our_ntp = get_ntp() log.debug('Our ntp = ' + str(our_ntp)) if our_ntp is None: return 0 current = float(our_ntp) future = float(origin_ntp) + float(self.ntp_delay) sleep_time = future - current log.debug('Waiting for fight') if sleep_time < 0: log.debug('We missed the meeting! It happened ' + str(-sleep_time) + 'seconds ago!') return 0 if sleep_time >= 300: log.debug('Future sleep time is too great!') return 0 busy_wait(sleep_time) release_priority_execution(p) log.debug('At fight') """ Time.sleep isn't guaranteed to sleep for the time specified which could cause synchronisation to be off between nodes and different OS' as per the discretion of the task scheduler. A busy wait is used to increase the accuracy of sleep. http://stackoverflow.com/questions/17499837/python-time-sleep-vs-busy-wait-accuracy http://stackoverflow.com/questions/1133857/how-accurate-is-pythons-time-sleep """ """ Making this algorithm "multi-threaded" has the potential to ruin predicted mappings for delta type NATs and NATs that have no care for source ports and assign incremental ports no matter what. """ threads = [] log.debug('Mapping len ' + str(len(my_mappings))) for mapping in my_mappings: prediction_len = len(predictions) if not prediction_len: break prediction = predictions[0] if self.nat_type == 'delta': self.throw_punch([mapping['sock'], node_ip, prediction]) else: args = [mapping['sock'], node_ip, prediction], 20 t = Thread(target=self.throw_punch, args=args) threads.append(t) t.start() predictions.remove(prediction) for t in threads: t.join() return 1
simultaneous_fight
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, my_mappings, node_ip, predictions, origin_ntp):\n p = request_priority_execution()\n log.debug('Getting NTP')\n if self.sys_clock is not None:\n our_ntp = self.sys_clock.time()\n else:\n our_ntp = get_ntp()\n log.debug('Our ntp = ' + str(our_ntp))\n if our_ntp is None:\n return 0\n current = float(our_ntp)\n future = float(origin_ntp) + float(self.ntp_delay)\n sleep_time = future - current\n log.debug('Waiting for fight')\n if sleep_time < 0:\n log.debug('We missed the meeting! It happened ' + str(-sleep_time) +\n 'seconds ago!')\n return 0\n if sleep_time >= 300:\n log.debug('Future sleep time is too great!')\n return 0\n busy_wait(sleep_time)\n release_priority_execution(p)\n log.debug('At fight')\n \"\"\"\n Time.sleep isn't guaranteed to sleep for the time specified\n which could cause synchronisation to be off between nodes\n and different OS' as per the discretion of the task scheduler.\n A busy wait is used to increase the accuracy of sleep.\n http://stackoverflow.com/questions/17499837/python-time-sleep-vs-busy-wait-accuracy\n http://stackoverflow.com/questions/1133857/how-accurate-is-pythons-time-sleep\n \"\"\"\n \"\"\"\n Making this algorithm \"multi-threaded\" has the potential to\n ruin predicted mappings for delta type NATs and NATs that\n have no care for source ports and assign incremental\n ports no matter what.\n \"\"\"\n threads = []\n log.debug('Mapping len ' + str(len(my_mappings)))\n for mapping in my_mappings:\n prediction_len = len(predictions)\n if not prediction_len:\n break\n prediction = predictions[0]\n if self.nat_type == 'delta':\n self.throw_punch([mapping['sock'], node_ip, prediction])\n else:\n args = [mapping['sock'], node_ip, prediction], 20\n t = Thread(target=self.throw_punch, args=args)\n threads.append(t)\n t.start()\n predictions.remove(prediction)\n for t in threads:\n t.join()\n return 1\n", "role": "user" }, { "content": "simultaneous_fight", "role": "assistant" } ]
def x(self, node_ip, node_port, proto): parts = self.sequential_connect() if parts is None: log.debug('Sequential connect failed') return None con, mappings, predictions = parts msg = 'CANDIDATE %s %s %s' % (node_ip, str(proto), predictions) con.send_line(msg) reply = con.recv_line(timeout=10) log.debug(reply) if 'PREDICTION SET' not in reply: log.debug('Prediction set failed') return None reply = con.recv_line(timeout=10) log.debug(reply) con.s.close() p = '^FIGHT ([0-9]+[.][0-9]+[.][0-9]+[.][0-9]+) ((?:[0-9]+\\s?)+)' p += ' (TCP|UDP) ([0-9]+(?:[.][0-9]+)?)$' parts = re.findall(p, reply) if not len(parts): log.debug('Invalid parts length') return None node_ip, predictions, proto, ntp = parts[0] log.debug('Received fight details') log.debug(str(parts[0])) log.debug('Attending fight now') return self.attend_fight(mappings, node_ip, predictions, ntp)
simultaneous_challenge
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, node_ip, node_port, proto):\n parts = self.sequential_connect()\n if parts is None:\n log.debug('Sequential connect failed')\n return None\n con, mappings, predictions = parts\n msg = 'CANDIDATE %s %s %s' % (node_ip, str(proto), predictions)\n con.send_line(msg)\n reply = con.recv_line(timeout=10)\n log.debug(reply)\n if 'PREDICTION SET' not in reply:\n log.debug('Prediction set failed')\n return None\n reply = con.recv_line(timeout=10)\n log.debug(reply)\n con.s.close()\n p = '^FIGHT ([0-9]+[.][0-9]+[.][0-9]+[.][0-9]+) ((?:[0-9]+\\\\s?)+)'\n p += ' (TCP|UDP) ([0-9]+(?:[.][0-9]+)?)$'\n parts = re.findall(p, reply)\n if not len(parts):\n log.debug('Invalid parts length')\n return None\n node_ip, predictions, proto, ntp = parts[0]\n log.debug('Received fight details')\n log.debug(str(parts[0]))\n log.debug('Attending fight now')\n return self.attend_fight(mappings, node_ip, predictions, ntp)\n", "role": "user" }, { "content": "simultaneous_challenge", "role": "assistant" } ]
def x(self, reply): remote_port = re.findall('^REMOTE (TCP|UDP) ([0-9]+)$', reply) if not len(remote_port): remote_port = 0 else: remote_port = int(remote_port[0][1]) if remote_port < 1 or remote_port > 65535: remote_port = 0 return remote_port
parse_remote_port
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, reply):\n remote_port = re.findall('^REMOTE (TCP|UDP) ([0-9]+)$', reply)\n if not len(remote_port):\n remote_port = 0\n else:\n remote_port = int(remote_port[0][1])\n if remote_port < 1 or remote_port > 65535:\n remote_port = 0\n return remote_port\n", "role": "user" }, { "content": "parse_remote_port", "role": "assistant" } ]
def x(self, return_instantly=1): if self.nat_type != 'unknown': return self.nat_type nat_type = 'random' if self.port_collisions * 5 > self.nat_tests: msg = 'Port collision number is too high compared to nat tests.' msg += ' Collisions must be in ratio 1 : 5 to avoid ambiguity' msg += ' in test results.' raise Exception(msg) """ Notes: This reuse test needs to ideally be performed against bootstrapping nodes on at least two different addresses and ports to each other because there are NAT types which allocate new mappings based on changes to these variables. """ def x(port=None, servers=None): servers = servers or self.rendezvous_servers con = None while con is None: try: sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM) sock.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1) port = port or get_unused_port(None) sock.bind(('', port)) source_port = sock.getsockname()[1] index = random.randrange(0, len(servers)) log.debug('Trying index: ' + str(index)) con = self.server_connect(sock, index, servers) except: time.sleep(1) sock.close() server = list(con.s.getpeername())[:] server = {'addr': server[0], 'port': server[1]} con.send_line('SOURCE TCP ' + str(source_port)) remote_port = con.recv_line(timeout=2) remote_port = self.parse_remote_port(remote_port) con.send_line('QUIT') return source_port, remote_port, server log.debug('Starting initial mappings for preserving + reuse tests') mappings = [] for i in range(0, self.nat_tests): src, remote, server = custom_server_con() mappings.append({'source': src, 'remote': int(remote), 'server': server}) log.debug(mappings) log.debug(len(mappings)) log.debug(self.nat_tests) log.debug('Finished mappings') preserving = 0 for mapping in mappings: if mapping['source'] == mapping['remote']: preserving += 1 if preserving >= self.nat_tests - self.port_collisions: nat_type = 'preserving' if return_instantly: return nat_type """ # Test reuse. log.debug("Testing reuse") reuse = 0 for mapping in mappings: addr = ("www.example.com", 80) servers = self.rendezvous_servers[:] servers.remove(mapping["server"]) log.debug("servers = " + str(servers)) src, remote, junk = custom_server_con(mapping["source"], servers) if remote == mapping["remote"]: reuse += 1 # Check reuse results. if reuse >= (self.nat_tests - self.port_collisions): nat_type = "reuse" if return_instantly: return nat_type # Load mappings for delta tests. mappings = sequential_bind(self.nat_tests, self.interface) for i in range(0, self.nat_tests): con = self.server_connect(mappings[i]["sock"]) con.send_line("SOURCE TCP " + str(mappings[i]["source"])) remote_port = self.parse_remote_port(con.recv_line(timeout=2)) mappings[i]["remote"] = int(remote_port) con.s.close() """ delta_ret = self.delta_test(mappings) if delta_ret['nat_type'] != 'random': self.delta = delta_ret['delta'] nat_type = 'delta' if return_instantly: return nat_type return nat_type
determine_nat
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, return_instantly=1):\n if self.nat_type != 'unknown':\n return self.nat_type\n nat_type = 'random'\n if self.port_collisions * 5 > self.nat_tests:\n msg = 'Port collision number is too high compared to nat tests.'\n msg += ' Collisions must be in ratio 1 : 5 to avoid ambiguity'\n msg += ' in test results.'\n raise Exception(msg)\n \"\"\"\n Notes: This reuse test needs to ideally be performed against\n bootstrapping nodes on at least two different addresses and\n ports to each other because there are NAT types which\n allocate new mappings based on changes to these variables.\n \"\"\"\n\n def x(port=None, servers=None):\n servers = servers or self.rendezvous_servers\n con = None\n while con is None:\n try:\n sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n sock.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1)\n port = port or get_unused_port(None)\n sock.bind(('', port))\n source_port = sock.getsockname()[1]\n index = random.randrange(0, len(servers))\n log.debug('Trying index: ' + str(index))\n con = self.server_connect(sock, index, servers)\n except:\n time.sleep(1)\n sock.close()\n server = list(con.s.getpeername())[:]\n server = {'addr': server[0], 'port': server[1]}\n con.send_line('SOURCE TCP ' + str(source_port))\n remote_port = con.recv_line(timeout=2)\n remote_port = self.parse_remote_port(remote_port)\n con.send_line('QUIT')\n return source_port, remote_port, server\n log.debug('Starting initial mappings for preserving + reuse tests')\n mappings = []\n for i in range(0, self.nat_tests):\n src, remote, server = custom_server_con()\n mappings.append({'source': src, 'remote': int(remote), 'server':\n server})\n log.debug(mappings)\n log.debug(len(mappings))\n log.debug(self.nat_tests)\n log.debug('Finished mappings')\n preserving = 0\n for mapping in mappings:\n if mapping['source'] == mapping['remote']:\n preserving += 1\n if preserving >= self.nat_tests - self.port_collisions:\n nat_type = 'preserving'\n if return_instantly:\n return nat_type\n \"\"\"\n # Test reuse.\n log.debug(\"Testing reuse\")\n reuse = 0\n for mapping in mappings:\n addr = (\"www.example.com\", 80)\n servers = self.rendezvous_servers[:]\n servers.remove(mapping[\"server\"])\n log.debug(\"servers = \" + str(servers))\n src, remote, junk = custom_server_con(mapping[\"source\"], servers)\n if remote == mapping[\"remote\"]:\n reuse += 1\n\n\n # Check reuse results.\n if reuse >= (self.nat_tests - self.port_collisions):\n nat_type = \"reuse\"\n if return_instantly:\n return nat_type\n\n # Load mappings for delta tests.\n mappings = sequential_bind(self.nat_tests, self.interface)\n for i in range(0, self.nat_tests):\n con = self.server_connect(mappings[i][\"sock\"])\n con.send_line(\"SOURCE TCP \" + str(mappings[i][\"source\"]))\n remote_port = self.parse_remote_port(con.recv_line(timeout=2))\n mappings[i][\"remote\"] = int(remote_port)\n con.s.close()\n \"\"\"\n delta_ret = self.delta_test(mappings)\n if delta_ret['nat_type'] != 'random':\n self.delta = delta_ret['delta']\n nat_type = 'delta'\n if return_instantly:\n return nat_type\n return nat_type\n", "role": "user" }, { "content": "determine_nat", "role": "assistant" } ]
def x(port=None): if port is None or port < 1024 or port > 65535: port = random.randint(1024, 65535) assert 1024 <= port <= 65535 while True: s = socket.socket(socket.AF_INET, socket.SOCK_STREAM) try: s.bind(('', port)) except socket.error as e: if e.errno in (98, 10048): return get_unused_port(None) raise e s.close() return port
get_unused_port
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(port=None):\n if port is None or port < 1024 or port > 65535:\n port = random.randint(1024, 65535)\n assert 1024 <= port <= 65535\n while True:\n s = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n try:\n s.bind(('', port))\n except socket.error as e:\n if e.errno in (98, 10048):\n return get_unused_port(None)\n raise e\n s.close()\n return port\n", "role": "user" }, { "content": "get_unused_port", "role": "assistant" } ]
def x(interface='default'): if sys.version_info < (3, 0, 0): if type(interface) == str: interface = unicode(interface) elif type(interface) == bytes: interface = interface.decode('utf-8') default_gateway = get_default_gateway(interface) gateways = netifaces.gateways() wan_id = None if netifaces.AF_INET in gateways: gw_list = gateways[netifaces.AF_INET] for gw_info in gw_list: if gw_info[0] == default_gateway: wan_id = gw_info[1] break interfaces = netifaces.interfaces() if wan_id in interfaces: families = netifaces.ifaddresses(wan_id) if netifaces.AF_INET in families: if_info_list = families[netifaces.AF_INET] for if_info in if_info_list: if 'addr' in if_info: return if_info['addr'] """ Execution may reach here if the host is using virtual interfaces on Linux and there are no gateways which suggests the host is a VPS or server. In this case """ if platform.system() == 'Linux': if ip is not None: return ip.routes['8.8.8.8']['prefsrc'] return None
get_lan_ip
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(interface='default'):\n if sys.version_info < (3, 0, 0):\n if type(interface) == str:\n interface = unicode(interface)\n elif type(interface) == bytes:\n interface = interface.decode('utf-8')\n default_gateway = get_default_gateway(interface)\n gateways = netifaces.gateways()\n wan_id = None\n if netifaces.AF_INET in gateways:\n gw_list = gateways[netifaces.AF_INET]\n for gw_info in gw_list:\n if gw_info[0] == default_gateway:\n wan_id = gw_info[1]\n break\n interfaces = netifaces.interfaces()\n if wan_id in interfaces:\n families = netifaces.ifaddresses(wan_id)\n if netifaces.AF_INET in families:\n if_info_list = families[netifaces.AF_INET]\n for if_info in if_info_list:\n if 'addr' in if_info:\n return if_info['addr']\n \"\"\"\n Execution may reach here if the host is using\n virtual interfaces on Linux and there are no gateways\n which suggests the host is a VPS or server. In this\n case\n \"\"\"\n if platform.system() == 'Linux':\n if ip is not None:\n return ip.routes['8.8.8.8']['prefsrc']\n return None\n", "role": "user" }, { "content": "get_lan_ip", "role": "assistant" } ]
def x(n=0): if n == 2: try: ip = myip() ip = extract_ip(ip) if is_ip_valid(ip): return ip except Exception as e: print(str(e)) return None from pyp2p.net import forwarding_servers for forwarding_server in forwarding_servers: url = 'http://' + forwarding_server['addr'] + ':' url += str(forwarding_server['port']) url += forwarding_server['url'] url += '?action=get_wan_ip' try: r = urlopen(url, timeout=5) response = r.read().decode('utf-8') response = extract_ip(response) if is_ip_valid(response): return response except Exception as e: print(str(e)) continue time.sleep(1) return get_wan_ip(n + 1)
get_wan_ip
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(n=0):\n if n == 2:\n try:\n ip = myip()\n ip = extract_ip(ip)\n if is_ip_valid(ip):\n return ip\n except Exception as e:\n print(str(e))\n return None\n from pyp2p.net import forwarding_servers\n for forwarding_server in forwarding_servers:\n url = 'http://' + forwarding_server['addr'] + ':'\n url += str(forwarding_server['port'])\n url += forwarding_server['url']\n url += '?action=get_wan_ip'\n try:\n r = urlopen(url, timeout=5)\n response = r.read().decode('utf-8')\n response = extract_ip(response)\n if is_ip_valid(response):\n return response\n except Exception as e:\n print(str(e))\n continue\n time.sleep(1)\n return get_wan_ip(n + 1)\n", "role": "user" }, { "content": "get_wan_ip", "role": "assistant" } ]
def x(): try: import netifaces return netifaces.gateways()['default'][netifaces.AF_INET][0] except ImportError: shell_command = 'netstat -rn' if os.name == 'posix': pattern = re.compile( '(?:default|0\\.0\\.0\\.0|::/0)\\s+([\\w\\.:]+)\\s+.*UG') elif os.name == 'nt': if platform.version().startswith('6.1'): pattern = re.compile('.*?0.0.0.0[ ]+0.0.0.0[ ]+(.*?)[ ]+?.*?\n' ) else: pattern = re.compile('.*?Default Gateway:[ ]+(.*?)\n') system_out = os.popen(shell_command, 'r').read() if not system_out: raise NATPMPNetworkError(NATPMP_GATEWAY_CANNOT_FIND, error_str( NATPMP_GATEWAY_CANNOT_FIND)) match = pattern.search(system_out) if not match: raise NATPMPNetworkError(NATPMP_GATEWAY_CANNOT_FIND, error_str( NATPMP_GATEWAY_CANNOT_FIND)) addr = match.groups()[0].strip() return addr
get_gateway_addr
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n try:\n import netifaces\n return netifaces.gateways()['default'][netifaces.AF_INET][0]\n except ImportError:\n shell_command = 'netstat -rn'\n if os.name == 'posix':\n pattern = re.compile(\n '(?:default|0\\\\.0\\\\.0\\\\.0|::/0)\\\\s+([\\\\w\\\\.:]+)\\\\s+.*UG')\n elif os.name == 'nt':\n if platform.version().startswith('6.1'):\n pattern = re.compile('.*?0.0.0.0[ ]+0.0.0.0[ ]+(.*?)[ ]+?.*?\\n'\n )\n else:\n pattern = re.compile('.*?Default Gateway:[ ]+(.*?)\\n')\n system_out = os.popen(shell_command, 'r').read()\n if not system_out:\n raise NATPMPNetworkError(NATPMP_GATEWAY_CANNOT_FIND, error_str(\n NATPMP_GATEWAY_CANNOT_FIND))\n match = pattern.search(system_out)\n if not match:\n raise NATPMPNetworkError(NATPMP_GATEWAY_CANNOT_FIND, error_str(\n NATPMP_GATEWAY_CANNOT_FIND))\n addr = match.groups()[0].strip()\n return addr\n", "role": "user" }, { "content": "get_gateway_addr", "role": "assistant" } ]
def x(gateway): if not gateway: raise NATPMPNetworkError(NATPMP_GATEWAY_NO_VALID_GATEWAY, error_str (NATPMP_GATEWAY_NO_VALID_GATEWAY)) response_socket = socket.socket(socket.AF_INET, socket.SOCK_DGRAM) response_socket.setblocking(0) response_socket.connect((gateway, NATPMP_PORT)) return response_socket
get_gateway_socket
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(gateway):\n if not gateway:\n raise NATPMPNetworkError(NATPMP_GATEWAY_NO_VALID_GATEWAY, error_str\n (NATPMP_GATEWAY_NO_VALID_GATEWAY))\n response_socket = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)\n response_socket.setblocking(0)\n response_socket.connect((gateway, NATPMP_PORT))\n return response_socket\n", "role": "user" }, { "content": "get_gateway_socket", "role": "assistant" } ]
def x(gateway_ip=None, retry=9): if gateway_ip is None: gateway_ip = get_gateway_addr() addr_request = PublicAddressRequest() addr_response = send_request_with_retry(gateway_ip, addr_request, response_data_class=PublicAddressResponse, retry=retry, response_size=12) if addr_response.result != 0: raise NATPMPResultError(addr_response.result, error_str( addr_response.result), addr_response) addr = addr_response.ip return addr
get_public_address
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(gateway_ip=None, retry=9):\n if gateway_ip is None:\n gateway_ip = get_gateway_addr()\n addr_request = PublicAddressRequest()\n addr_response = send_request_with_retry(gateway_ip, addr_request,\n response_data_class=PublicAddressResponse, retry=retry,\n response_size=12)\n if addr_response.result != 0:\n raise NATPMPResultError(addr_response.result, error_str(\n addr_response.result), addr_response)\n addr = addr_response.ip\n return addr\n", "role": "user" }, { "content": "get_public_address", "role": "assistant" } ]
def x(public_port, private_port, lifetime=3600, gateway_ip=None, retry=9, use_exception=True): return map_port(NATPMP_PROTOCOL_TCP, public_port, private_port, lifetime, gateway_ip=gateway_ip, retry=retry, use_exception= use_exception)
map_tcp_port
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(public_port, private_port, lifetime=3600, gateway_ip=None, retry=9,\n use_exception=True):\n return map_port(NATPMP_PROTOCOL_TCP, public_port, private_port,\n lifetime, gateway_ip=gateway_ip, retry=retry, use_exception=\n use_exception)\n", "role": "user" }, { "content": "map_tcp_port", "role": "assistant" } ]
def x(public_port, private_port, lifetime=3600, gateway_ip=None, retry=9, use_exception=True): return map_port(NATPMP_PROTOCOL_UDP, public_port, private_port, lifetime, gateway_ip=gateway_ip, retry=retry, use_exception= use_exception)
map_udp_port
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(public_port, private_port, lifetime=3600, gateway_ip=None, retry=9,\n use_exception=True):\n return map_port(NATPMP_PROTOCOL_UDP, public_port, private_port,\n lifetime, gateway_ip=gateway_ip, retry=retry, use_exception=\n use_exception)\n", "role": "user" }, { "content": "map_udp_port", "role": "assistant" } ]
def x(protocol, public_port, private_port, lifetime=3600, gateway_ip=None, retry=9, use_exception=True): if protocol not in [NATPMP_PROTOCOL_UDP, NATPMP_PROTOCOL_TCP]: raise ValueError( 'Must be either NATPMP_PROTOCOL_UDP or NATPMP_PROTOCOL_TCP') if gateway_ip is None: gateway_ip = get_gateway_addr() response = None port_mapping_request = PortMapRequest(protocol, private_port, public_port, lifetime) port_mapping_response = send_request_with_retry(gateway_ip, port_mapping_request, response_data_class=PortMapResponse, retry=retry) if port_mapping_response.result != 0 and use_exception: raise NATPMPResultError(port_mapping_response.result, error_str( port_mapping_response.result), port_mapping_response) return port_mapping_response
map_port
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(protocol, public_port, private_port, lifetime=3600, gateway_ip=None,\n retry=9, use_exception=True):\n if protocol not in [NATPMP_PROTOCOL_UDP, NATPMP_PROTOCOL_TCP]:\n raise ValueError(\n 'Must be either NATPMP_PROTOCOL_UDP or NATPMP_PROTOCOL_TCP')\n if gateway_ip is None:\n gateway_ip = get_gateway_addr()\n response = None\n port_mapping_request = PortMapRequest(protocol, private_port,\n public_port, lifetime)\n port_mapping_response = send_request_with_retry(gateway_ip,\n port_mapping_request, response_data_class=PortMapResponse, retry=retry)\n if port_mapping_response.result != 0 and use_exception:\n raise NATPMPResultError(port_mapping_response.result, error_str(\n port_mapping_response.result), port_mapping_response)\n return port_mapping_response\n", "role": "user" }, { "content": "map_port", "role": "assistant" } ]
def x(self, proto, src_port, dest_ip, dest_port=None): proto = proto.upper() valid_protos = ['TCP', 'UDP'] if proto not in valid_protos: raise Exception('Invalid protocol for forwarding.') valid_ports = range(1, 65535) if src_port not in valid_ports: raise Exception('Invalid port for forwarding.') if dest_port is None: dest_port = src_port if platform.system() == 'Windows': cmd = 'upnpc-static.exe -a %s %s %s %s' % (get_lan_ip(), str( src_port), str(dest_port), proto) out, err = subprocess.Popen(cmd, shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE).communicate() if 'is not recognized' in err: raise Exception('Missing upnpc-static.exe') return gateway_addr = self.find_gateway() if gateway_addr is None: raise Exception('Unable to find UPnP compatible gateway.') rhost = re.findall('([^/]+)', gateway_addr) res = urlopen(gateway_addr, timeout=self.timeout).read().decode('utf-8') res = res.replace('\r', '') res = res.replace('\n', '') res = res.replace('\t', '') pres = res.split('<serviceId>urn:upnp-org:serviceId:WANIPConn1</serviceId>' ) p2res = pres[1].split('</controlURL>') p3res = p2res[0].split('<controlURL>') ctrl = p3res[1] rip = res.split('<presentationURL>') rip1 = rip[1].split('</presentationURL>') router_ip = rip1[0] port_map_desc = 'PyP2P' msg = ( '<?xml version="1.0"?><s:Envelope xmlns:s="http://schemas.xmlsoap.org/soap/envelope/" s:encodingStyle="http://schemas.xmlsoap.org/soap/encoding/"><s:Body><u:AddPortMapping xmlns:u="urn:schemas-upnp-org:service:WANIPConnection:1"><NewRemoteHost></NewRemoteHost><NewExternalPort>' + str(src_port) + '</NewExternalPort><NewProtocol>' + str(proto) + '</NewProtocol><NewInternalPort>' + str(dest_port) + '</NewInternalPort><NewInternalClient>' + str(dest_ip) + '</NewInternalClient><NewEnabled>1</NewEnabled><NewPortMappingDescription>' + str(port_map_desc) + '</NewPortMappingDescription><NewLeaseDuration>0</NewLeaseDuration></u:AddPortMapping></s:Body></s:Envelope>' ) x = 'http://' + rhost[1] + '/' + ctrl if sys.version_info >= (3, 0, 0): msg = bytes(msg, 'utf-8') req = Request('http://' + rhost[1] + '/' + ctrl, msg) req.add_header('SOAPAction', '"urn:schemas-upnp-org:service:WANIPConnection:1#AddPortMapping"') req.add_header('Content-type', 'application/xml') res = urlopen(req, timeout=self.timeout)
forward_port
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, proto, src_port, dest_ip, dest_port=None):\n proto = proto.upper()\n valid_protos = ['TCP', 'UDP']\n if proto not in valid_protos:\n raise Exception('Invalid protocol for forwarding.')\n valid_ports = range(1, 65535)\n if src_port not in valid_ports:\n raise Exception('Invalid port for forwarding.')\n if dest_port is None:\n dest_port = src_port\n if platform.system() == 'Windows':\n cmd = 'upnpc-static.exe -a %s %s %s %s' % (get_lan_ip(), str(\n src_port), str(dest_port), proto)\n out, err = subprocess.Popen(cmd, shell=True, stdout=subprocess.PIPE,\n stderr=subprocess.PIPE).communicate()\n if 'is not recognized' in err:\n raise Exception('Missing upnpc-static.exe')\n return\n gateway_addr = self.find_gateway()\n if gateway_addr is None:\n raise Exception('Unable to find UPnP compatible gateway.')\n rhost = re.findall('([^/]+)', gateway_addr)\n res = urlopen(gateway_addr, timeout=self.timeout).read().decode('utf-8')\n res = res.replace('\\r', '')\n res = res.replace('\\n', '')\n res = res.replace('\\t', '')\n pres = res.split('<serviceId>urn:upnp-org:serviceId:WANIPConn1</serviceId>'\n )\n p2res = pres[1].split('</controlURL>')\n p3res = p2res[0].split('<controlURL>')\n ctrl = p3res[1]\n rip = res.split('<presentationURL>')\n rip1 = rip[1].split('</presentationURL>')\n router_ip = rip1[0]\n port_map_desc = 'PyP2P'\n msg = (\n '<?xml version=\"1.0\"?><s:Envelope xmlns:s=\"http://schemas.xmlsoap.org/soap/envelope/\" s:encodingStyle=\"http://schemas.xmlsoap.org/soap/encoding/\"><s:Body><u:AddPortMapping xmlns:u=\"urn:schemas-upnp-org:service:WANIPConnection:1\"><NewRemoteHost></NewRemoteHost><NewExternalPort>'\n + str(src_port) + '</NewExternalPort><NewProtocol>' + str(proto) +\n '</NewProtocol><NewInternalPort>' + str(dest_port) +\n '</NewInternalPort><NewInternalClient>' + str(dest_ip) +\n '</NewInternalClient><NewEnabled>1</NewEnabled><NewPortMappingDescription>'\n + str(port_map_desc) +\n '</NewPortMappingDescription><NewLeaseDuration>0</NewLeaseDuration></u:AddPortMapping></s:Body></s:Envelope>'\n )\n x = 'http://' + rhost[1] + '/' + ctrl\n if sys.version_info >= (3, 0, 0):\n msg = bytes(msg, 'utf-8')\n req = Request('http://' + rhost[1] + '/' + ctrl, msg)\n req.add_header('SOAPAction',\n '\"urn:schemas-upnp-org:service:WANIPConnection:1#AddPortMapping\"')\n req.add_header('Content-type', 'application/xml')\n res = urlopen(req, timeout=self.timeout)\n", "role": "user" }, { "content": "forward_port", "role": "assistant" } ]
def x(self): self.stop() self.initialize() self.handle = self.loop.call_at(self.get_next(), self.call_next)
start
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n self.stop()\n self.initialize()\n self.handle = self.loop.call_at(self.get_next(), self.call_next)\n", "role": "user" }, { "content": "start", "role": "assistant" } ]
def x(self): if self.handle is not None: self.handle.cancel() self.handle = self.future = self.croniter = None
stop
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.handle is not None:\n self.handle.cancel()\n self.handle = self.future = self.croniter = None\n", "role": "user" }, { "content": "stop", "role": "assistant" } ]
def x(self, *args): self.initialize() self.future = asyncio.Future(loop=self.loop) self.handle = self.loop.call_at(self.get_next(), self.call_func, *args) return self.future
next
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, *args):\n self.initialize()\n self.future = asyncio.Future(loop=self.loop)\n self.handle = self.loop.call_at(self.get_next(), self.call_func, *args)\n return self.future\n", "role": "user" }, { "content": "next", "role": "assistant" } ]
def x(self): if self.croniter is None: self.time = time.time() self.datetime = datetime.now(self.tz) self.loop_time = self.loop.time() self.croniter = croniter(self.spec, start_time=self.datetime)
initialize
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.croniter is None:\n self.time = time.time()\n self.datetime = datetime.now(self.tz)\n self.loop_time = self.loop.time()\n self.croniter = croniter(self.spec, start_time=self.datetime)\n", "role": "user" }, { "content": "initialize", "role": "assistant" } ]
def x(self): return self.loop_time + (self.croniter.get_next(float) - self.time)
get_next
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return self.loop_time + (self.croniter.get_next(float) - self.time)\n", "role": "user" }, { "content": "get_next", "role": "assistant" } ]
def x(self): if self.handle is not None: self.handle.cancel() next_time = self.get_next() self.handle = self.loop.call_at(next_time, self.call_next) self.call_func()
call_next
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.handle is not None:\n self.handle.cancel()\n next_time = self.get_next()\n self.handle = self.loop.call_at(next_time, self.call_next)\n self.call_func()\n", "role": "user" }, { "content": "call_next", "role": "assistant" } ]
def x(self, *args, **kwargs): asyncio.gather(self.cron(*args, **kwargs), loop=self.loop, return_exceptions=True).add_done_callback(self.set_result)
call_func
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, *args, **kwargs):\n asyncio.gather(self.cron(*args, **kwargs), loop=self.loop,\n return_exceptions=True).add_done_callback(self.set_result)\n", "role": "user" }, { "content": "call_func", "role": "assistant" } ]
def x(self, result): result = result.result()[0] if self.future is not None: if isinstance(result, Exception): self.future.set_exception(result) else: self.future.set_result(result) self.future = None elif isinstance(result, Exception): raise result
set_result
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, result):\n result = result.result()[0]\n if self.future is not None:\n if isinstance(result, Exception):\n self.future.set_exception(result)\n else:\n self.future.set_result(result)\n self.future = None\n elif isinstance(result, Exception):\n raise result\n", "role": "user" }, { "content": "set_result", "role": "assistant" } ]
def x(sval): match = re.match('\\s*' + TIMEFORMAT + '\\s*$', sval, re.I) if not match or not match.group(0).strip(): return mdict = match.groupdict() return sum(MULTIPLIERS[k] * cast(v) for k, v in mdict.items() if v is not None)
timeparse
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(sval):\n match = re.match('\\\\s*' + TIMEFORMAT + '\\\\s*$', sval, re.I)\n if not match or not match.group(0).strip():\n return\n mdict = match.groupdict()\n return sum(MULTIPLIERS[k] * cast(v) for k, v in mdict.items() if v is not\n None)\n", "role": "user" }, { "content": "timeparse", "role": "assistant" } ]
def x(text, callback=None, small_first_last=True): lines = re.split('[\r\n]+', text) processed = [] for line in lines: all_caps = line.upper() == line words = re.split('[\t ]', line) tc_line = [] for word in words: if callback: new_word = callback(word, all_caps=all_caps) if new_word: tc_line.append(_mark_immutable(new_word)) continue if all_caps: if UC_INITIALS.match(word): tc_line.append(word) continue if APOS_SECOND.match(word): if len(word[0]) == 1 and word[0] not in 'aeiouAEIOU': word = word[0].lower() + word[1] + word[2].upper() + word[ 3:] else: word = word[0].upper() + word[1] + word[2].upper() + word[ 3:] tc_line.append(word) continue match = MAC_MC.match(word) if match: tc_line.append('%s%s' % (match.group(1).capitalize(), titlecase(match.group(2), callback, small_first_last))) continue if INLINE_PERIOD.search(word ) or not all_caps and UC_ELSEWHERE.match(word): tc_line.append(word) continue if SMALL_WORDS.match(word): tc_line.append(word.lower()) continue if '/' in word and '//' not in word: slashed = map(lambda t: titlecase(t, callback, False), word .split('/')) tc_line.append('/'.join(slashed)) continue if '-' in word: hyphenated = map(lambda t: titlecase(t, callback, small_first_last), word.split('-')) tc_line.append('-'.join(hyphenated)) continue if all_caps: word = word.lower() tc_line.append(CAPFIRST.sub(lambda m: m.group(0).upper(), word)) if small_first_last and tc_line: if not isinstance(tc_line[0], Immutable): tc_line[0] = SMALL_FIRST.sub(lambda m: '%s%s' % (m.group(1), m.group(2).capitalize()), tc_line[0]) if not isinstance(tc_line[-1], Immutable): tc_line[-1] = SMALL_LAST.sub(lambda m: m.group(0). capitalize(), tc_line[-1]) result = ' '.join(tc_line) result = SUBPHRASE.sub(lambda m: '%s%s' % (m.group(1), m.group(2). capitalize()), result) processed.append(result) return '\n'.join(processed)
titlecase
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(text, callback=None, small_first_last=True):\n lines = re.split('[\\r\\n]+', text)\n processed = []\n for line in lines:\n all_caps = line.upper() == line\n words = re.split('[\\t ]', line)\n tc_line = []\n for word in words:\n if callback:\n new_word = callback(word, all_caps=all_caps)\n if new_word:\n tc_line.append(_mark_immutable(new_word))\n continue\n if all_caps:\n if UC_INITIALS.match(word):\n tc_line.append(word)\n continue\n if APOS_SECOND.match(word):\n if len(word[0]) == 1 and word[0] not in 'aeiouAEIOU':\n word = word[0].lower() + word[1] + word[2].upper() + word[\n 3:]\n else:\n word = word[0].upper() + word[1] + word[2].upper() + word[\n 3:]\n tc_line.append(word)\n continue\n match = MAC_MC.match(word)\n if match:\n tc_line.append('%s%s' % (match.group(1).capitalize(),\n titlecase(match.group(2), callback, small_first_last)))\n continue\n if INLINE_PERIOD.search(word\n ) or not all_caps and UC_ELSEWHERE.match(word):\n tc_line.append(word)\n continue\n if SMALL_WORDS.match(word):\n tc_line.append(word.lower())\n continue\n if '/' in word and '//' not in word:\n slashed = map(lambda t: titlecase(t, callback, False), word\n .split('/'))\n tc_line.append('/'.join(slashed))\n continue\n if '-' in word:\n hyphenated = map(lambda t: titlecase(t, callback,\n small_first_last), word.split('-'))\n tc_line.append('-'.join(hyphenated))\n continue\n if all_caps:\n word = word.lower()\n tc_line.append(CAPFIRST.sub(lambda m: m.group(0).upper(), word))\n if small_first_last and tc_line:\n if not isinstance(tc_line[0], Immutable):\n tc_line[0] = SMALL_FIRST.sub(lambda m: '%s%s' % (m.group(1),\n m.group(2).capitalize()), tc_line[0])\n if not isinstance(tc_line[-1], Immutable):\n tc_line[-1] = SMALL_LAST.sub(lambda m: m.group(0).\n capitalize(), tc_line[-1])\n result = ' '.join(tc_line)\n result = SUBPHRASE.sub(lambda m: '%s%s' % (m.group(1), m.group(2).\n capitalize()), result)\n processed.append(result)\n return '\\n'.join(processed)\n", "role": "user" }, { "content": "titlecase", "role": "assistant" } ]
def x(): parser = argparse.ArgumentParser() in_group = parser.add_mutually_exclusive_group() in_group.add_argument('string', nargs='*', default=[], help= 'String to titlecase') in_group.add_argument('-f', '--input-file', help= 'File to read from to titlecase') parser.add_argument('-o', '--output-file', help= 'File to write titlecased output to)') args = parser.parse_args() if args.input_file is not None: if args.input_file == '-': ifile = sys.stdin else: ifile = open(args.input_file) else: ifile = sys.stdin if args.output_file is not None: if args.output_file == '-': ofile = sys.stdout else: ofile = open(args.output_file, 'w') else: ofile = sys.stdout if len(args.string) > 0: in_string = ' '.join(args.string) else: with ifile: in_string = ifile.read() with ofile: ofile.write(titlecase(in_string))
cmd
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n parser = argparse.ArgumentParser()\n in_group = parser.add_mutually_exclusive_group()\n in_group.add_argument('string', nargs='*', default=[], help=\n 'String to titlecase')\n in_group.add_argument('-f', '--input-file', help=\n 'File to read from to titlecase')\n parser.add_argument('-o', '--output-file', help=\n 'File to write titlecased output to)')\n args = parser.parse_args()\n if args.input_file is not None:\n if args.input_file == '-':\n ifile = sys.stdin\n else:\n ifile = open(args.input_file)\n else:\n ifile = sys.stdin\n if args.output_file is not None:\n if args.output_file == '-':\n ofile = sys.stdout\n else:\n ofile = open(args.output_file, 'w')\n else:\n ofile = sys.stdout\n if len(args.string) > 0:\n in_string = ' '.join(args.string)\n else:\n with ifile:\n in_string = ifile.read()\n with ofile:\n ofile.write(titlecase(in_string))\n", "role": "user" }, { "content": "cmd", "role": "assistant" } ]
def x(cls, parser: OptionManager) ->None: parser.add_option('--eradicate-aggressive', default=False, help= 'Enables aggressive mode for eradicate; this may result in false positives' , action='store_true', type=None)
add_options
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, parser: OptionManager) ->None:\n parser.add_option('--eradicate-aggressive', default=False, help=\n 'Enables aggressive mode for eradicate; this may result in false positives'\n , action='store_true', type=None)\n", "role": "user" }, { "content": "add_options", "role": "assistant" } ]
def x(self) ->Generator[Tuple[int, int, str, type], None, None]: if self.filename != STDIN: buffer = StringIO() options = _Options(aggressive=self.options.eradicate_aggressive) fix_file(self.filename, options, buffer) traceback = buffer.getvalue() if traceback: yield 1, 0, self._error(traceback), type(self)
run
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self) ->Generator[Tuple[int, int, str, type], None, None]:\n if self.filename != STDIN:\n buffer = StringIO()\n options = _Options(aggressive=self.options.eradicate_aggressive)\n fix_file(self.filename, options, buffer)\n traceback = buffer.getvalue()\n if traceback:\n yield 1, 0, self._error(traceback), type(self)\n", "role": "user" }, { "content": "run", "role": "assistant" } ]
def x(g): yielded = set() for value in g: if value not in yielded: yield value yielded.add(value)
unique
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(g):\n yielded = set()\n for value in g:\n if value not in yielded:\n yield value\n yielded.add(value)\n", "role": "user" }, { "content": "unique", "role": "assistant" } ]
def x(t, name): for type_ in t.mro(): try: return vars(type_)[name] except KeyError: pass raise AttributeError(name)
static_get_type_attr
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(t, name):\n for type_ in t.mro():\n try:\n return vars(type_)[name]\n except KeyError:\n pass\n raise AttributeError(name)\n", "role": "user" }, { "content": "static_get_type_attr", "role": "assistant" } ]
def x(typename, conflicts): message = ('\nclass {C} received conflicting default implementations:'. format(C=typename)) for attrname, interfaces in conflicts.items(): message += dedent( """ The following interfaces provided default implementations for {attr!r}: {interfaces}""" ).format(attr=attrname, interfaces=bulleted_list(sorted(map( getname, interfaces)))) return InvalidImplementation(message)
conflicting_defaults
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(typename, conflicts):\n message = ('\\nclass {C} received conflicting default implementations:'.\n format(C=typename))\n for attrname, interfaces in conflicts.items():\n message += dedent(\n \"\"\"\n\n The following interfaces provided default implementations for {attr!r}:\n {interfaces}\"\"\"\n ).format(attr=attrname, interfaces=bulleted_list(sorted(map(\n getname, interfaces))))\n return InvalidImplementation(message)\n", "role": "user" }, { "content": "conflicting_defaults", "role": "assistant" } ]
def x(self, type_): missing = [] mistyped = {} mismatched = {} for name, iface_sig in self._signatures.items(): try: f = static_get_type_attr(type_, name) except AttributeError: missing.append(name) continue impl_sig = TypedSignature(f) if not issubclass(impl_sig.type, iface_sig.type): mistyped[name] = impl_sig.type if not compatible(impl_sig.signature, iface_sig.signature): mismatched[name] = impl_sig return missing, mistyped, mismatched
diff_signatures
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, type_):\n missing = []\n mistyped = {}\n mismatched = {}\n for name, iface_sig in self._signatures.items():\n try:\n f = static_get_type_attr(type_, name)\n except AttributeError:\n missing.append(name)\n continue\n impl_sig = TypedSignature(f)\n if not issubclass(impl_sig.type, iface_sig.type):\n mistyped[name] = impl_sig.type\n if not compatible(impl_sig.signature, iface_sig.signature):\n mismatched[name] = impl_sig\n return missing, mistyped, mismatched\n", "role": "user" }, { "content": "diff_signatures", "role": "assistant" } ]
def x(self, type_): raw_missing, mistyped, mismatched = self._diff_signatures(type_) missing = [] defaults_to_use = {} for name in raw_missing: try: defaults_to_use[name] = self._defaults[name].implementation except KeyError: missing.append(name) if not any((missing, mistyped, mismatched)): return defaults_to_use raise self._invalid_implementation(type_, missing, mistyped, mismatched)
verify
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, type_):\n raw_missing, mistyped, mismatched = self._diff_signatures(type_)\n missing = []\n defaults_to_use = {}\n for name in raw_missing:\n try:\n defaults_to_use[name] = self._defaults[name].implementation\n except KeyError:\n missing.append(name)\n if not any((missing, mistyped, mismatched)):\n return defaults_to_use\n raise self._invalid_implementation(type_, missing, mistyped, mismatched)\n", "role": "user" }, { "content": "verify", "role": "assistant" } ]
def x(self, t, missing, mistyped, mismatched): assert missing or mistyped or mismatched, "Implementation wasn't invalid." message = '\nclass {C} failed to implement interface {I}:'.format(C= getname(t), I=getname(self)) if missing: message += dedent( """ The following methods of {I} were not implemented: {missing_methods}""" ).format(I=getname(self), missing_methods=self. _format_missing_methods(missing)) if mistyped: message += dedent( """ The following methods of {I} were implemented with incorrect types: {mismatched_types}""" ).format(I=getname(self), mismatched_types=self. _format_mismatched_types(mistyped)) if mismatched: message += dedent( """ The following methods of {I} were implemented with invalid signatures: {mismatched_methods}""" ).format(I=getname(self), mismatched_methods=self. _format_mismatched_methods(mismatched)) return InvalidImplementation(message)
invalid_implementation
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, t, missing, mistyped, mismatched):\n assert missing or mistyped or mismatched, \"Implementation wasn't invalid.\"\n message = '\\nclass {C} failed to implement interface {I}:'.format(C=\n getname(t), I=getname(self))\n if missing:\n message += dedent(\n \"\"\"\n\n The following methods of {I} were not implemented:\n {missing_methods}\"\"\"\n ).format(I=getname(self), missing_methods=self.\n _format_missing_methods(missing))\n if mistyped:\n message += dedent(\n \"\"\"\n\n The following methods of {I} were implemented with incorrect types:\n {mismatched_types}\"\"\"\n ).format(I=getname(self), mismatched_types=self.\n _format_mismatched_types(mistyped))\n if mismatched:\n message += dedent(\n \"\"\"\n\n The following methods of {I} were implemented with invalid signatures:\n {mismatched_methods}\"\"\"\n ).format(I=getname(self), mismatched_methods=self.\n _format_mismatched_methods(mismatched))\n return InvalidImplementation(message)\n", "role": "user" }, { "content": "invalid_implementation", "role": "assistant" } ]
def x(cls, existing_class, subset=None, name=None): if name is None: name = existing_class.__name__ + 'Interface' if subset is None: subset = set(dir(existing_class)) - TRIVIAL_CLASS_ATTRIBUTES return InterfaceMeta(name, (Interface,), {name: static_get_type_attr( existing_class, name) for name in subset})
from_class
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, existing_class, subset=None, name=None):\n if name is None:\n name = existing_class.__name__ + 'Interface'\n if subset is None:\n subset = set(dir(existing_class)) - TRIVIAL_CLASS_ATTRIBUTES\n return InterfaceMeta(name, (Interface,), {name: static_get_type_attr(\n existing_class, name) for name in subset})\n", "role": "user" }, { "content": "from_class", "role": "assistant" } ]
def x(impl_sig, iface_sig): return all([positionals_compatible(takewhile(is_positional, impl_sig. parameters.values()), takewhile(is_positional, iface_sig.parameters .values())), keywords_compatible(valfilter(complement(is_positional ), impl_sig.parameters), valfilter(complement(is_positional), iface_sig.parameters))])
compatible
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(impl_sig, iface_sig):\n return all([positionals_compatible(takewhile(is_positional, impl_sig.\n parameters.values()), takewhile(is_positional, iface_sig.parameters\n .values())), keywords_compatible(valfilter(complement(is_positional\n ), impl_sig.parameters), valfilter(complement(is_positional),\n iface_sig.parameters))])\n", "role": "user" }, { "content": "compatible", "role": "assistant" } ]
def x(*args, **kwargs): func = kwargs['func'] del kwargs['func'] return aggregate_np(*args, func=lambda x: func(x), **kwargs)
aggregate_group_loop
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(*args, **kwargs):\n func = kwargs['func']\n del kwargs['func']\n return aggregate_np(*args, func=lambda x: func(x), **kwargs)\n", "role": "user" }, { "content": "aggregate_group_loop", "role": "assistant" } ]
def x(group_idx): cmp_pos = 0 steps = 1 if len(group_idx) < 1: return 0 for i in range(len(group_idx)): if group_idx[cmp_pos] != group_idx[i]: cmp_pos = i steps += 1 return steps
step_count
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(group_idx):\n cmp_pos = 0\n steps = 1\n if len(group_idx) < 1:\n return 0\n for i in range(len(group_idx)):\n if group_idx[cmp_pos] != group_idx[i]:\n cmp_pos = i\n steps += 1\n return steps\n", "role": "user" }, { "content": "step_count", "role": "assistant" } ]
def x(group_idx): ilen = step_count(group_idx) + 1 indices = np.empty(ilen, np.int64) indices[0] = 0 indices[-1] = group_idx.size cmp_pos = 0 ri = 1 for i in range(len(group_idx)): if group_idx[cmp_pos] != group_idx[i]: cmp_pos = i indices[ri] = i ri += 1 return indices
step_indices
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(group_idx):\n ilen = step_count(group_idx) + 1\n indices = np.empty(ilen, np.int64)\n indices[0] = 0\n indices[-1] = group_idx.size\n cmp_pos = 0\n ri = 1\n for i in range(len(group_idx)):\n if group_idx[cmp_pos] != group_idx[i]:\n cmp_pos = i\n indices[ri] = i\n ri += 1\n return indices\n", "role": "user" }, { "content": "step_indices", "role": "assistant" } ]
def x(cls, nans=False, reverse=False, scalar=False): _valgetter = cls._valgetter_scalar if scalar else cls._valgetter valgetter = nb.njit(_valgetter) outersetter = nb.njit(cls._outersetter) _cls_inner = nb.njit(cls._inner) if nans: def x(ri, val, ret, counter, mean): if not np.isnan(val): _cls_inner(ri, val, ret, counter, mean) inner = nb.njit(_inner) else: inner = _cls_inner def x(group_idx, a, ret, counter, mean, outer, fill_value, ddof): size = len(ret) rng = range(len(group_idx) - 1, -1, -1) if reverse else range(len( group_idx)) for i in rng: ri = group_idx[i] if ri < 0: raise ValueError('negative indices not supported') if ri >= size: raise ValueError( 'one or more indices in group_idx are too large') val = valgetter(a, i) inner(ri, val, ret, counter, mean) outersetter(outer, i, ret[ri]) return nb.njit(_loop, nogil=True)
callable
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, nans=False, reverse=False, scalar=False):\n _valgetter = cls._valgetter_scalar if scalar else cls._valgetter\n valgetter = nb.njit(_valgetter)\n outersetter = nb.njit(cls._outersetter)\n _cls_inner = nb.njit(cls._inner)\n if nans:\n\n def x(ri, val, ret, counter, mean):\n if not np.isnan(val):\n _cls_inner(ri, val, ret, counter, mean)\n inner = nb.njit(_inner)\n else:\n inner = _cls_inner\n\n def x(group_idx, a, ret, counter, mean, outer, fill_value, ddof):\n size = len(ret)\n rng = range(len(group_idx) - 1, -1, -1) if reverse else range(len(\n group_idx))\n for i in rng:\n ri = group_idx[i]\n if ri < 0:\n raise ValueError('negative indices not supported')\n if ri >= size:\n raise ValueError(\n 'one or more indices in group_idx are too large')\n val = valgetter(a, i)\n inner(ri, val, ret, counter, mean)\n outersetter(outer, i, ret[ri])\n return nb.njit(_loop, nogil=True)\n", "role": "user" }, { "content": "callable", "role": "assistant" } ]
def x(self, nans=False): jitfunc = nb.njit(self.func, nogil=True) def x(sortidx, group_idx, a, ret): size = len(ret) group_idx_srt = group_idx[sortidx] a_srt = a[sortidx] indices = step_indices(group_idx_srt) for i in range(len(indices) - 1): start_idx, stop_idx = indices[i], indices[i + 1] ri = group_idx_srt[start_idx] if ri < 0: raise ValueError('negative indices not supported') if ri >= size: raise ValueError( 'one or more indices in group_idx are too large') ret[ri] = jitfunc(a_srt[start_idx:stop_idx]) return nb.njit(_loop, nogil=True)
callable
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, nans=False):\n jitfunc = nb.njit(self.func, nogil=True)\n\n def x(sortidx, group_idx, a, ret):\n size = len(ret)\n group_idx_srt = group_idx[sortidx]\n a_srt = a[sortidx]\n indices = step_indices(group_idx_srt)\n for i in range(len(indices) - 1):\n start_idx, stop_idx = indices[i], indices[i + 1]\n ri = group_idx_srt[start_idx]\n if ri < 0:\n raise ValueError('negative indices not supported')\n if ri >= size:\n raise ValueError(\n 'one or more indices in group_idx are too large')\n ret[ri] = jitfunc(a_srt[start_idx:stop_idx])\n return nb.njit(_loop, nogil=True)\n", "role": "user" }, { "content": "callable", "role": "assistant" } ]
def x(*extra): alias = dict((k, k) for k in funcs_common) alias.update(_alias_str) alias.update((fn, fn) for fn in _alias_builtin.values()) alias.update(_alias_builtin) for d in extra: alias.update(d) alias.update((k, k) for k in set(alias.values())) for key in set(alias.values()): if key not in funcs_no_separate_nan: key = 'nan' + key alias[key] = key return alias
get_aliasing
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(*extra):\n alias = dict((k, k) for k in funcs_common)\n alias.update(_alias_str)\n alias.update((fn, fn) for fn in _alias_builtin.values())\n alias.update(_alias_builtin)\n for d in extra:\n alias.update(d)\n alias.update((k, k) for k in set(alias.values()))\n for key in set(alias.values()):\n if key not in funcs_no_separate_nan:\n key = 'nan' + key\n alias[key] = key\n return alias\n", "role": "user" }, { "content": "get_aliasing", "role": "assistant" } ]
def x(func, aliasing, implementations): try: func_str = aliasing[func] except KeyError: if callable(func): return func else: if func_str in implementations: return func_str if func_str.startswith('nan') and func_str[3: ] in funcs_no_separate_nan: raise ValueError('%s does not have a nan-version'.format( func_str[3:])) else: raise NotImplementedError('No such function available') raise ValueError( 'func %s is neither a valid function string nor a callable object'. format(func))
get_func
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(func, aliasing, implementations):\n try:\n func_str = aliasing[func]\n except KeyError:\n if callable(func):\n return func\n else:\n if func_str in implementations:\n return func_str\n if func_str.startswith('nan') and func_str[3:\n ] in funcs_no_separate_nan:\n raise ValueError('%s does not have a nan-version'.format(\n func_str[3:]))\n else:\n raise NotImplementedError('No such function available')\n raise ValueError(\n 'func %s is neither a valid function string nor a callable object'.\n format(func))\n", "role": "user" }, { "content": "get_func", "role": "assistant" } ]
def x(x, dtype=np.bool_): def x(x, dtype): try: converted = dtype.type(x) except (ValueError, OverflowError): return False return converted == x or np.isnan(x) def x(x, dtype, dtype_dict, default=None): while True: try: dtype = np.dtype(dtype_dict[dtype.name]) if check_type(x, dtype): return np.dtype(dtype) except KeyError: if default is not None: return np.dtype(default) raise ValueError('Can not determine dtype of %r' % x) dtype = np.dtype(dtype) if check_type(x, dtype): return dtype if np.issubdtype(dtype, np.inexact): return type_loop(x, dtype, _next_float_dtype) else: return type_loop(x, dtype, _next_int_dtype, default=np.float32)
minimum_dtype
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(x, dtype=np.bool_):\n\n def x(x, dtype):\n try:\n converted = dtype.type(x)\n except (ValueError, OverflowError):\n return False\n return converted == x or np.isnan(x)\n\n def x(x, dtype, dtype_dict, default=None):\n while True:\n try:\n dtype = np.dtype(dtype_dict[dtype.name])\n if check_type(x, dtype):\n return np.dtype(dtype)\n except KeyError:\n if default is not None:\n return np.dtype(default)\n raise ValueError('Can not determine dtype of %r' % x)\n dtype = np.dtype(dtype)\n if check_type(x, dtype):\n return dtype\n if np.issubdtype(dtype, np.inexact):\n return type_loop(x, dtype, _next_float_dtype)\n else:\n return type_loop(x, dtype, _next_int_dtype, default=np.float32)\n", "role": "user" }, { "content": "minimum_dtype", "role": "assistant" } ]
def x(group_idx, a, size=None, order='C', axis=None, ravel_group_idx=True, check_bounds=True): if not isinstance(a, (int, float, complex)): a = np.asanyarray(a) group_idx = np.asanyarray(group_idx) if not np.issubdtype(group_idx.dtype, np.integer): raise TypeError('group_idx must be of integer type') if check_bounds and np.any(group_idx < 0): raise ValueError('negative indices not supported') ndim_idx = np.ndim(group_idx) ndim_a = np.ndim(a) if axis is None: if ndim_a > 1: raise ValueError( 'a must be scalar or 1 dimensional, use .ravel to flatten. Alternatively specify axis.' ) elif axis >= ndim_a or axis < -ndim_a: raise ValueError('axis arg too large for np.ndim(a)') else: axis = axis if axis >= 0 else ndim_a + axis if ndim_idx > 1: raise NotImplementedError( 'only 1d indexing currentlysupported with axis arg.') elif a.shape[axis] != len(group_idx): raise ValueError("a.shape[axis] doesn't match length of group_idx." ) elif size is not None and not np.isscalar(size): raise NotImplementedError( 'when using axis arg, size must beNone or scalar.') else: size_in = np.max(group_idx) + 1 if size is None else size group_idx_in = group_idx group_idx = [] size = [] for ii, s in enumerate(a.shape): ii_idx = group_idx_in if ii == axis else np.arange(s) ii_shape = [1] * ndim_a ii_shape[ii] = s group_idx.append(ii_idx.reshape(ii_shape)) size.append(size_in if ii == axis else s) group_idx = np.ravel_multi_index(group_idx, size, order=order, mode='raise') flat_size = np.prod(size) ndim_idx = ndim_a return group_idx.ravel(), a.ravel(), flat_size, ndim_idx, size if ndim_idx == 1: if size is None: size = np.max(group_idx) + 1 else: if not np.isscalar(size): raise ValueError('output size must be scalar or None') if check_bounds and np.any(group_idx > size - 1): raise ValueError( 'one or more indices are too large for size %d' % size) flat_size = size else: if size is None: size = np.max(group_idx, axis=1) + 1 elif np.isscalar(size): raise ValueError('output size must be of length %d' % len( group_idx)) elif len(size) != len(group_idx): raise ValueError( '%d sizes given, but %d output dimensions specified in index' % (len(size), len(group_idx))) if ravel_group_idx: group_idx = np.ravel_multi_index(group_idx, size, order=order, mode='raise') flat_size = np.prod(size) if not (np.ndim(a) == 0 or len(a) == group_idx.size): raise ValueError( 'group_idx and a must be of the same length, or a can be scalar') return group_idx, a, flat_size, ndim_idx, size
input_validation
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(group_idx, a, size=None, order='C', axis=None, ravel_group_idx=True,\n check_bounds=True):\n if not isinstance(a, (int, float, complex)):\n a = np.asanyarray(a)\n group_idx = np.asanyarray(group_idx)\n if not np.issubdtype(group_idx.dtype, np.integer):\n raise TypeError('group_idx must be of integer type')\n if check_bounds and np.any(group_idx < 0):\n raise ValueError('negative indices not supported')\n ndim_idx = np.ndim(group_idx)\n ndim_a = np.ndim(a)\n if axis is None:\n if ndim_a > 1:\n raise ValueError(\n 'a must be scalar or 1 dimensional, use .ravel to flatten. Alternatively specify axis.'\n )\n elif axis >= ndim_a or axis < -ndim_a:\n raise ValueError('axis arg too large for np.ndim(a)')\n else:\n axis = axis if axis >= 0 else ndim_a + axis\n if ndim_idx > 1:\n raise NotImplementedError(\n 'only 1d indexing currentlysupported with axis arg.')\n elif a.shape[axis] != len(group_idx):\n raise ValueError(\"a.shape[axis] doesn't match length of group_idx.\"\n )\n elif size is not None and not np.isscalar(size):\n raise NotImplementedError(\n 'when using axis arg, size must beNone or scalar.')\n else:\n size_in = np.max(group_idx) + 1 if size is None else size\n group_idx_in = group_idx\n group_idx = []\n size = []\n for ii, s in enumerate(a.shape):\n ii_idx = group_idx_in if ii == axis else np.arange(s)\n ii_shape = [1] * ndim_a\n ii_shape[ii] = s\n group_idx.append(ii_idx.reshape(ii_shape))\n size.append(size_in if ii == axis else s)\n group_idx = np.ravel_multi_index(group_idx, size, order=order,\n mode='raise')\n flat_size = np.prod(size)\n ndim_idx = ndim_a\n return group_idx.ravel(), a.ravel(), flat_size, ndim_idx, size\n if ndim_idx == 1:\n if size is None:\n size = np.max(group_idx) + 1\n else:\n if not np.isscalar(size):\n raise ValueError('output size must be scalar or None')\n if check_bounds and np.any(group_idx > size - 1):\n raise ValueError(\n 'one or more indices are too large for size %d' % size)\n flat_size = size\n else:\n if size is None:\n size = np.max(group_idx, axis=1) + 1\n elif np.isscalar(size):\n raise ValueError('output size must be of length %d' % len(\n group_idx))\n elif len(size) != len(group_idx):\n raise ValueError(\n '%d sizes given, but %d output dimensions specified in index' %\n (len(size), len(group_idx)))\n if ravel_group_idx:\n group_idx = np.ravel_multi_index(group_idx, size, order=order,\n mode='raise')\n flat_size = np.prod(size)\n if not (np.ndim(a) == 0 or len(a) == group_idx.size):\n raise ValueError(\n 'group_idx and a must be of the same length, or a can be scalar')\n return group_idx, a, flat_size, ndim_idx, size\n", "role": "user" }, { "content": "input_validation", "role": "assistant" } ]
def x(n): if n.ndim != 1: raise ValueError('n is supposed to be 1d array.') n_mask = n.astype(bool) n_cumsum = np.cumsum(n) ret = np.ones(n_cumsum[-1] + 1, dtype=int) ret[n_cumsum[n_mask]] -= n[n_mask] ret[0] -= 1 return np.cumsum(ret)[:-1]
multi_arange
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(n):\n if n.ndim != 1:\n raise ValueError('n is supposed to be 1d array.')\n n_mask = n.astype(bool)\n n_cumsum = np.cumsum(n)\n ret = np.ones(n_cumsum[-1] + 1, dtype=int)\n ret[n_cumsum[n_mask]] -= n[n_mask]\n ret[0] -= 1\n return np.cumsum(ret)[:-1]\n", "role": "user" }, { "content": "multi_arange", "role": "assistant" } ]
def x(X): if X.ndim != 1: raise ValueError('this is for 1d masks only.') is_start = np.empty(len(X), dtype=bool) is_start[0] = X[0] if X.dtype.kind == 'b': is_start[1:] = ~X[:-1] & X[1:] M = X else: M = X.astype(bool) is_start[1:] = X[:-1] != X[1:] is_start[~M] = False L = np.cumsum(is_start) L[~M] = 0 return L
label_contiguous_1d
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(X):\n if X.ndim != 1:\n raise ValueError('this is for 1d masks only.')\n is_start = np.empty(len(X), dtype=bool)\n is_start[0] = X[0]\n if X.dtype.kind == 'b':\n is_start[1:] = ~X[:-1] & X[1:]\n M = X\n else:\n M = X.astype(bool)\n is_start[1:] = X[:-1] != X[1:]\n is_start[~M] = False\n L = np.cumsum(is_start)\n L[~M] = 0\n return L\n", "role": "user" }, { "content": "label_contiguous_1d", "role": "assistant" } ]
def x(group_idx): keep_group = np.zeros(np.max(group_idx) + 1, dtype=bool) keep_group[0] = True keep_group[group_idx] = True return relabel_groups_masked(group_idx, keep_group)
relabel_groups_unique
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(group_idx):\n keep_group = np.zeros(np.max(group_idx) + 1, dtype=bool)\n keep_group[0] = True\n keep_group[group_idx] = True\n return relabel_groups_masked(group_idx, keep_group)\n", "role": "user" }, { "content": "relabel_groups_unique", "role": "assistant" } ]
def x(group_idx, keep_group): keep_group = keep_group.astype(bool, copy=not keep_group[0]) if not keep_group[0]: keep_group[0] = True relabel = np.zeros(keep_group.size, dtype=group_idx.dtype) relabel[keep_group] = np.arange(np.count_nonzero(keep_group)) return relabel[group_idx]
relabel_groups_masked
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(group_idx, keep_group):\n keep_group = keep_group.astype(bool, copy=not keep_group[0])\n if not keep_group[0]:\n keep_group[0] = True\n relabel = np.zeros(keep_group.size, dtype=group_idx.dtype)\n relabel[keep_group] = np.arange(np.count_nonzero(keep_group))\n return relabel[group_idx]\n", "role": "user" }, { "content": "relabel_groups_masked", "role": "assistant" } ]
def x(group_idx, a, size, fill_value, dtype=None): if fill_value is not None and not (np.isscalar(fill_value) or len( fill_value) == 0): raise ValueError( 'fill_value must be None, a scalar or an empty sequence') order_group_idx = np.argsort(group_idx, kind='mergesort') counts = np.bincount(group_idx, minlength=size) ret = np.split(a[order_group_idx], np.cumsum(counts)[:-1]) ret = np.asanyarray(ret) if fill_value is None or np.isscalar(fill_value): _fill_untouched(group_idx, ret, fill_value) return ret
array
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(group_idx, a, size, fill_value, dtype=None):\n if fill_value is not None and not (np.isscalar(fill_value) or len(\n fill_value) == 0):\n raise ValueError(\n 'fill_value must be None, a scalar or an empty sequence')\n order_group_idx = np.argsort(group_idx, kind='mergesort')\n counts = np.bincount(group_idx, minlength=size)\n ret = np.split(a[order_group_idx], np.cumsum(counts)[:-1])\n ret = np.asanyarray(ret)\n if fill_value is None or np.isscalar(fill_value):\n _fill_untouched(group_idx, ret, fill_value)\n return ret\n", "role": "user" }, { "content": "array", "role": "assistant" } ]
def x(group_idx, a, size, fill_value, dtype=None, func=lambda g: g, **kwargs): groups = _array(group_idx, a, size, ()) ret = np.full(size, fill_value, dtype=dtype or np.float64) for i, grp in enumerate(groups): if np.ndim(grp) == 1 and len(grp) > 0: ret[i] = func(grp) return ret
generic_callable
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(group_idx, a, size, fill_value, dtype=None, func=lambda g: g, **kwargs):\n groups = _array(group_idx, a, size, ())\n ret = np.full(size, fill_value, dtype=dtype or np.float64)\n for i, grp in enumerate(groups):\n if np.ndim(grp) == 1 and len(grp) > 0:\n ret[i] = func(grp)\n return ret\n", "role": "user" }, { "content": "generic_callable", "role": "assistant" } ]
def x(group_idx, a, size, fill_value=None, dtype=None): sortidx = np.argsort(group_idx, kind='mergesort') invsortidx = np.argsort(sortidx, kind='mergesort') group_idx_srt = group_idx[sortidx] a_srt = a[sortidx] a_srt_cumsum = np.cumsum(a_srt, dtype=dtype) increasing = np.arange(len(a), dtype=int) group_starts = _min(group_idx_srt, increasing, size, fill_value=0)[ group_idx_srt] a_srt_cumsum += -a_srt_cumsum[group_starts] + a_srt[group_starts] return a_srt_cumsum[invsortidx]
cumsum
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(group_idx, a, size, fill_value=None, dtype=None):\n sortidx = np.argsort(group_idx, kind='mergesort')\n invsortidx = np.argsort(sortidx, kind='mergesort')\n group_idx_srt = group_idx[sortidx]\n a_srt = a[sortidx]\n a_srt_cumsum = np.cumsum(a_srt, dtype=dtype)\n increasing = np.arange(len(a), dtype=int)\n group_starts = _min(group_idx_srt, increasing, size, fill_value=0)[\n group_idx_srt]\n a_srt_cumsum += -a_srt_cumsum[group_starts] + a_srt[group_starts]\n return a_srt_cumsum[invsortidx]\n", "role": "user" }, { "content": "cumsum", "role": "assistant" } ]
def x(idx, ret, fill_value): untouched = np.ones_like(ret, dtype=bool) untouched[idx] = False ret[untouched] = fill_value
fill_untouched
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(idx, ret, fill_value):\n untouched = np.ones_like(ret, dtype=bool)\n untouched[idx] = False\n ret[untouched] = fill_value\n", "role": "user" }, { "content": "fill_untouched", "role": "assistant" } ]
def x(*args, **kwargs): extrafuncs = {'allnan': allnan, 'anynan': anynan, 'first': itemgetter(0 ), 'last': itemgetter(-1), 'nanfirst': nanfirst, 'nanlast': nanlast} func = kwargs.pop('func') func = extrafuncs.get(func, func) if isinstance(func, str): raise NotImplementedError( 'Grouploop needs to be called with a function') return aggregate_numpy.aggregate(*args, func=lambda x: func(x), **kwargs)
aggregate_grouploop
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(*args, **kwargs):\n extrafuncs = {'allnan': allnan, 'anynan': anynan, 'first': itemgetter(0\n ), 'last': itemgetter(-1), 'nanfirst': nanfirst, 'nanlast': nanlast}\n func = kwargs.pop('func')\n func = extrafuncs.get(func, func)\n if isinstance(func, str):\n raise NotImplementedError(\n 'Grouploop needs to be called with a function')\n return aggregate_numpy.aggregate(*args, func=lambda x: func(x), **kwargs)\n", "role": "user" }, { "content": "aggregate_grouploop", "role": "assistant" } ]
def x(group_idx, a, size, fill_value, dtype=None): dtype = minimum_dtype_scalar(fill_value, dtype, a) ret = np.full(size, fill_value, dtype=dtype) if fill_value != 1: ret[group_idx] = 1 np.multiply.at(ret, group_idx, a) return ret
prod
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(group_idx, a, size, fill_value, dtype=None):\n dtype = minimum_dtype_scalar(fill_value, dtype, a)\n ret = np.full(size, fill_value, dtype=dtype)\n if fill_value != 1:\n ret[group_idx] = 1\n np.multiply.at(ret, group_idx, a)\n return ret\n", "role": "user" }, { "content": "prod", "role": "assistant" } ]
def x(funcname, reverse=False, nans=False, scalar=False): varnames = ['group_idx', 'a', 'ret', 'counter'] codebase = c_base_reverse if reverse else c_base iteration = c_iter_scalar[funcname] if scalar else c_iter[funcname] if scalar: varnames.remove('a') return codebase % dict(init=c_init(varnames), iter=iteration, finish= c_finish.get(funcname, ''), ri_redir=c_ri_redir if nans else c_ri)
c_func
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(funcname, reverse=False, nans=False, scalar=False):\n varnames = ['group_idx', 'a', 'ret', 'counter']\n codebase = c_base_reverse if reverse else c_base\n iteration = c_iter_scalar[funcname] if scalar else c_iter[funcname]\n if scalar:\n varnames.remove('a')\n return codebase % dict(init=c_init(varnames), iter=iteration, finish=\n c_finish.get(funcname, ''), ri_redir=c_ri_redir if nans else c_ri)\n", "role": "user" }, { "content": "c_func", "role": "assistant" } ]
def x(group_idx): ilen = step_count(group_idx) + 1 indices = np.empty(ilen, int) indices[0] = 0 indices[-1] = group_idx.size inline(c_step_indices, ['group_idx', 'indices'], define_macros=c_macros, extra_compile_args=c_args) return indices
step_indices
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(group_idx):\n ilen = step_count(group_idx) + 1\n indices = np.empty(ilen, int)\n indices[0] = 0\n indices[-1] = group_idx.size\n inline(c_step_indices, ['group_idx', 'indices'], define_macros=c_macros,\n extra_compile_args=c_args)\n return indices\n", "role": "user" }, { "content": "step_indices", "role": "assistant" } ]
def x(self, size): s = 1 / self.density return np.random.choice([-np.sqrt(s / self.k), 0, np.sqrt(s / self.k)], size=size, p=[1 / (2 * s), 1 - 1 / s, 1 / (2 * s)])
create_proj_mat
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, size):\n s = 1 / self.density\n return np.random.choice([-np.sqrt(s / self.k), 0, np.sqrt(s / self.k)],\n size=size, p=[1 / (2 * s), 1 - 1 / s, 1 / (2 * s)])\n", "role": "user" }, { "content": "create_proj_mat", "role": "assistant" } ]
def x(data_home, size): all_genres = ['Action', 'Adventure', 'Animation', "Children's", 'Comedy', 'Crime', 'Documentary', 'Drama', 'Fantasy', 'Film-Noir', 'Horror', 'Musical', 'Mystery', 'Romance', 'Sci-Fi', 'Thriller', 'War', 'Western'] n_genre = len(all_genres) movies = {} if size == '100k': with open(os.path.join(data_home, 'u.item'), encoding='ISO-8859-1' ) as f: lines = list(map(lambda l: l.rstrip().split('|'), f.readlines())) for line in lines: movie_vec = np.zeros(n_genre) for i, flg_chr in enumerate(line[-n_genre:]): if flg_chr == '1': movie_vec[i] = 1.0 movie_id = int(line[0]) movies[movie_id] = movie_vec elif size == '1m': with open(os.path.join(data_home, 'movies.dat'), encoding='ISO-8859-1' ) as f: lines = list(map(lambda l: l.rstrip().split('::'), f.readlines())) for item_id_str, title, genres in lines: movie_vec = np.zeros(n_genre) for genre in genres.split('|'): i = all_genres.index(genre) movie_vec[i] = 1.0 item_id = int(item_id_str) movies[item_id] = movie_vec return movies
load_movies
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(data_home, size):\n all_genres = ['Action', 'Adventure', 'Animation', \"Children's\",\n 'Comedy', 'Crime', 'Documentary', 'Drama', 'Fantasy', 'Film-Noir',\n 'Horror', 'Musical', 'Mystery', 'Romance', 'Sci-Fi', 'Thriller',\n 'War', 'Western']\n n_genre = len(all_genres)\n movies = {}\n if size == '100k':\n with open(os.path.join(data_home, 'u.item'), encoding='ISO-8859-1'\n ) as f:\n lines = list(map(lambda l: l.rstrip().split('|'), f.readlines()))\n for line in lines:\n movie_vec = np.zeros(n_genre)\n for i, flg_chr in enumerate(line[-n_genre:]):\n if flg_chr == '1':\n movie_vec[i] = 1.0\n movie_id = int(line[0])\n movies[movie_id] = movie_vec\n elif size == '1m':\n with open(os.path.join(data_home, 'movies.dat'), encoding='ISO-8859-1'\n ) as f:\n lines = list(map(lambda l: l.rstrip().split('::'), f.readlines()))\n for item_id_str, title, genres in lines:\n movie_vec = np.zeros(n_genre)\n for genre in genres.split('|'):\n i = all_genres.index(genre)\n movie_vec[i] = 1.0\n item_id = int(item_id_str)\n movies[item_id] = movie_vec\n return movies\n", "role": "user" }, { "content": "load_movies", "role": "assistant" } ]