source_code
stringlengths
45
7.24k
function_name
stringlengths
1
111
messages
list
def x(self, kwargs): criteria = {} for key in self._search_by: if getattr(self, key, None): criteria[key] = getattr(self, key) elif key in kwargs and kwargs.get(key): criteria[key] = kwargs.get(key) return criteria
build_search_values
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, kwargs):\n criteria = {}\n for key in self._search_by:\n if getattr(self, key, None):\n criteria[key] = getattr(self, key)\n elif key in kwargs and kwargs.get(key):\n criteria[key] = kwargs.get(key)\n return criteria\n", "role": "user" }, { "content": "build_search_values", "role": "assistant" } ]
def x(self, ipv4addr): for addr in self.ipv4addrs: if isinstance(addr, dict) and addr['ipv4addr' ] == ipv4addr or isinstance(addr, HostIPv4 ) and addr.ipv4addr == ipv4addr: raise ValueError('Already exists') self.ipv4addrs.append({'ipv4addr': ipv4addr})
add_ipv4addr
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, ipv4addr):\n for addr in self.ipv4addrs:\n if isinstance(addr, dict) and addr['ipv4addr'\n ] == ipv4addr or isinstance(addr, HostIPv4\n ) and addr.ipv4addr == ipv4addr:\n raise ValueError('Already exists')\n self.ipv4addrs.append({'ipv4addr': ipv4addr})\n", "role": "user" }, { "content": "add_ipv4addr", "role": "assistant" } ]
def x(self, ipv4addr): for addr in self.ipv4addrs: if isinstance(addr, dict) and addr['ipv4addr' ] == ipv4addr or isinstance(addr, HostIPv4 ) and addr.ipv4addr == ipv4addr: self.ipv4addrs.remove(addr) break
remove_ipv4addr
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, ipv4addr):\n for addr in self.ipv4addrs:\n if isinstance(addr, dict) and addr['ipv4addr'\n ] == ipv4addr or isinstance(addr, HostIPv4\n ) and addr.ipv4addr == ipv4addr:\n self.ipv4addrs.remove(addr)\n break\n", "role": "user" }, { "content": "remove_ipv4addr", "role": "assistant" } ]
def x(self, ipv6addr): for addr in self.ipv6addrs: if isinstance(addr, dict) and addr['ipv6addr' ] == ipv6addr or isinstance(addr, HostIPv4 ) and addr.ipv6addr == ipv6addr: raise ValueError('Already exists') self.ipv6addrs.append({'ipv6addr': ipv6addr})
add_ipv6addr
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, ipv6addr):\n for addr in self.ipv6addrs:\n if isinstance(addr, dict) and addr['ipv6addr'\n ] == ipv6addr or isinstance(addr, HostIPv4\n ) and addr.ipv6addr == ipv6addr:\n raise ValueError('Already exists')\n self.ipv6addrs.append({'ipv6addr': ipv6addr})\n", "role": "user" }, { "content": "add_ipv6addr", "role": "assistant" } ]
def x(self, ipv6addr): for addr in self.ipv6addrs: if isinstance(addr, dict) and addr['ipv6addr' ] == ipv6addr or isinstance(addr, HostIPv4 ) and addr.ipv6addr == ipv6addr: self.ipv6addrs.remove(addr) break
remove_ipv6addr
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, ipv6addr):\n for addr in self.ipv6addrs:\n if isinstance(addr, dict) and addr['ipv6addr'\n ] == ipv6addr or isinstance(addr, HostIPv4\n ) and addr.ipv6addr == ipv6addr:\n self.ipv6addrs.remove(addr)\n break\n", "role": "user" }, { "content": "remove_ipv6addr", "role": "assistant" } ]
def x(self, uri, version=None, isolation_level='READ COMMITTED'): isolation_level = isolation_level or 'READ COMMITTED' kwargs = dict(isolation_level=isolation_level) types = {list: pg.ARRAY, tuple: pg.ARRAY, set: pg.ARRAY, dict: JSONDict, datetime: UTCEpoch} self.type_map.update(types) bs = self.config['batch_size'] self.config['batch_size'] = 5000 if bs == 999 else bs self._lock_required = False dsn = self.config['db_schema'] self.config['db_schema'] = dsn or 'public' return uri, kwargs
sqla_postgresql
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, uri, version=None, isolation_level='READ COMMITTED'):\n isolation_level = isolation_level or 'READ COMMITTED'\n kwargs = dict(isolation_level=isolation_level)\n types = {list: pg.ARRAY, tuple: pg.ARRAY, set: pg.ARRAY, dict: JSONDict,\n datetime: UTCEpoch}\n self.type_map.update(types)\n bs = self.config['batch_size']\n self.config['batch_size'] = 5000 if bs == 999 else bs\n self._lock_required = False\n dsn = self.config['db_schema']\n self.config['db_schema'] = dsn or 'public'\n return uri, kwargs\n", "role": "user" }, { "content": "sqla_postgresql", "role": "assistant" } ]
def x(self, objects, **kwargs): return autoschema(objects=objects, exclude_keys=self.RESTRICTED_KEYS, **kwargs)
autoschema
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, objects, **kwargs):\n return autoschema(objects=objects, exclude_keys=self.RESTRICTED_KEYS,\n **kwargs)\n", "role": "user" }, { "content": "autoschema", "role": "assistant" } ]
def x(self, query=None, date=None, table=None): table = table or self.config.get('table') sql_count = select([func.count()]) query = self._parse_query(table=table, query=query, date=date, fields= 'id', alias='anon_x') if query is not None: query = sql_count.select_from(query) else: table = self.get_table(table) query = sql_count query = query.select_from(table) return self.session_auto.execute(query).scalar()
count
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, query=None, date=None, table=None):\n table = table or self.config.get('table')\n sql_count = select([func.count()])\n query = self._parse_query(table=table, query=query, date=date, fields=\n 'id', alias='anon_x')\n if query is not None:\n query = sql_count.select_from(query)\n else:\n table = self.get_table(table)\n query = sql_count\n query = query.select_from(table)\n return self.session_auto.execute(query).scalar()\n", "role": "user" }, { "content": "count", "role": "assistant" } ]
def x(self, field, oids, date=None, level=None, table=None): table = self.get_table(table) fringe = str2list(oids) checked = set(fringe) loop_k = 0 while len(fringe) > 0: if level and loop_k == abs(level): break query = '_oid in %s' % list(fringe) docs = self.find(table=table, query=query, fields=[field], date= date, raw=True) fringe = {oid for doc in docs for oid in doc[field] or [] if oid not in checked} checked |= fringe loop_k += 1 return sorted(checked)
deptree
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, field, oids, date=None, level=None, table=None):\n table = self.get_table(table)\n fringe = str2list(oids)\n checked = set(fringe)\n loop_k = 0\n while len(fringe) > 0:\n if level and loop_k == abs(level):\n break\n query = '_oid in %s' % list(fringe)\n docs = self.find(table=table, query=query, fields=[field], date=\n date, raw=True)\n fringe = {oid for doc in docs for oid in doc[field] or [] if oid not in\n checked}\n checked |= fringe\n loop_k += 1\n return sorted(checked)\n", "role": "user" }, { "content": "deptree", "role": "assistant" } ]
def x(self, field, table=None): field = field if is_array(field) else [field] table = self.get_table(table, except_=False) if table is None: last = None else: is_defined(field, 'field must be defined!') last = self.find(table=table, fields=field, scalar=True, sort=field, limit=1, descending=True, date='~', default_fields=False) logger.debug('last %s.%s: %s' % (table, list2str(field), last)) return last
get_last_field
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, field, table=None):\n field = field if is_array(field) else [field]\n table = self.get_table(table, except_=False)\n if table is None:\n last = None\n else:\n is_defined(field, 'field must be defined!')\n last = self.find(table=table, fields=field, scalar=True, sort=field,\n limit=1, descending=True, date='~', default_fields=False)\n logger.debug('last %s.%s: %s' % (table, list2str(field), last))\n return last\n", "role": "user" }, { "content": "get_last_field", "role": "assistant" } ]
def x(self, fields, name=None, table=None, **kwargs): table = self.get_table(table) name = self._index_default_name(fields, name) fields = parse.parse_fields(fields) fields = self.columns(table, fields, reflect=True) session = self.session_new() index = Index(name, *fields) logger.info('Writing new index %s: %s' % (name, fields)) result = index.create(self.engine) session.commit() return result
index
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, fields, name=None, table=None, **kwargs):\n table = self.get_table(table)\n name = self._index_default_name(fields, name)\n fields = parse.parse_fields(fields)\n fields = self.columns(table, fields, reflect=True)\n session = self.session_new()\n index = Index(name, *fields)\n logger.info('Writing new index %s: %s' % (name, fields))\n result = index.create(self.engine)\n session.commit()\n return result\n", "role": "user" }, { "content": "index", "role": "assistant" } ]
def x(self): logger.info('Listing indexes') _ix = {} _i = self.inspector for tbl in _i.get_table_names(): _ix.setdefault(tbl, []) for ix in _i.get_indexes(tbl): _ix[tbl].append(ix) return _ix
index_list
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n logger.info('Listing indexes')\n _ix = {}\n _i = self.inspector\n for tbl in _i.get_table_names():\n _ix.setdefault(tbl, [])\n for ix in _i.get_indexes(tbl):\n _ix[tbl].append(ix)\n return _ix\n", "role": "user" }, { "content": "index_list", "role": "assistant" } ]
def x(self, startswith=None): logger.info('Listing cubes starting with "%s")' % startswith) startswith = unicode(startswith or '') tables = sorted(name for name in self.db_tables if name.startswith( startswith)) return tables
ls
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, startswith=None):\n logger.info('Listing cubes starting with \"%s\")' % startswith)\n startswith = unicode(startswith or '')\n tables = sorted(name for name in self.db_tables if name.startswith(\n startswith))\n return tables\n", "role": "user" }, { "content": "ls", "role": "assistant" } ]
def x(self, with_user, roles=None, table=None): table = self.get_table(table) is_true(table is not None, 'invalid table: %s' % table) with_user = validate_username(with_user) roles = roles or ['SELECT'] roles = validate_roles(roles, self.VALID_SHARE_ROLES) roles = list2str(roles) logger.info('Sharing cube %s with %s (%s)' % (table, with_user, roles)) sql = 'GRANT %s ON %s TO %s' % (roles, table, with_user) return self.session_auto.execute(sql)
share
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, with_user, roles=None, table=None):\n table = self.get_table(table)\n is_true(table is not None, 'invalid table: %s' % table)\n with_user = validate_username(with_user)\n roles = roles or ['SELECT']\n roles = validate_roles(roles, self.VALID_SHARE_ROLES)\n roles = list2str(roles)\n logger.info('Sharing cube %s with %s (%s)' % (table, with_user, roles))\n sql = 'GRANT %s ON %s TO %s' % (roles, table, with_user)\n return self.session_auto.execute(sql)\n", "role": "user" }, { "content": "share", "role": "assistant" } ]
def x(seqs, add_seq_names=False, out_filename=None, input_handler=None, params={}, WorkingDir=tempfile.gettempdir(), SuppressStderr=None, SuppressStdout=None): if out_filename: params['-out'] = out_filename ih = input_handler or guess_input_handler(seqs, add_seq_names) muscle_app = Muscle(params=params, InputHandler=ih, WorkingDir= WorkingDir, SuppressStderr=SuppressStderr, SuppressStdout= SuppressStdout) return muscle_app(seqs)
muscle_seqs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(seqs, add_seq_names=False, out_filename=None, input_handler=None,\n params={}, WorkingDir=tempfile.gettempdir(), SuppressStderr=None,\n SuppressStdout=None):\n if out_filename:\n params['-out'] = out_filename\n ih = input_handler or guess_input_handler(seqs, add_seq_names)\n muscle_app = Muscle(params=params, InputHandler=ih, WorkingDir=\n WorkingDir, SuppressStderr=SuppressStderr, SuppressStdout=\n SuppressStdout)\n return muscle_app(seqs)\n", "role": "user" }, { "content": "muscle_seqs", "role": "assistant" } ]
def x(seqs, input_handler=None, tree_type='neighborjoining', params={}, add_seq_names=True, WorkingDir=tempfile.gettempdir(), SuppressStderr= None, SuppressStdout=None, max_hours=5.0, constructor=PhyloNode, clean_up=True): params['-maxhours'] = max_hours if tree_type: params['-cluster2'] = tree_type params['-tree2'] = get_tmp_filename(WorkingDir) params['-out'] = get_tmp_filename(WorkingDir) muscle_res = muscle_seqs(seqs, input_handler=input_handler, params= params, add_seq_names=add_seq_names, WorkingDir=WorkingDir, SuppressStderr=SuppressStderr, SuppressStdout=SuppressStdout) tree = DndParser(muscle_res['Tree2Out'], constructor=constructor) aln = [line for line in muscle_res['MuscleOut']] if clean_up: muscle_res.cleanUp() return tree, aln
aln_tree_seqs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(seqs, input_handler=None, tree_type='neighborjoining', params={},\n add_seq_names=True, WorkingDir=tempfile.gettempdir(), SuppressStderr=\n None, SuppressStdout=None, max_hours=5.0, constructor=PhyloNode,\n clean_up=True):\n params['-maxhours'] = max_hours\n if tree_type:\n params['-cluster2'] = tree_type\n params['-tree2'] = get_tmp_filename(WorkingDir)\n params['-out'] = get_tmp_filename(WorkingDir)\n muscle_res = muscle_seqs(seqs, input_handler=input_handler, params=\n params, add_seq_names=add_seq_names, WorkingDir=WorkingDir,\n SuppressStderr=SuppressStderr, SuppressStdout=SuppressStdout)\n tree = DndParser(muscle_res['Tree2Out'], constructor=constructor)\n aln = [line for line in muscle_res['MuscleOut']]\n if clean_up:\n muscle_res.cleanUp()\n return tree, aln\n", "role": "user" }, { "content": "aln_tree_seqs", "role": "assistant" } ]
def x(seqs, moltype=DNA, params=None): if not params: params = {} seq_collection = SequenceCollection(seqs, MolType=moltype) int_map, int_keys = seq_collection.getIntMap() int_map = SequenceCollection(int_map, MolType=moltype) params.update({'-out': get_tmp_filename()}) app = Muscle(InputHandler='_input_as_multiline_string', params=params, WorkingDir=tempfile.gettempdir()) res = app(int_map.toFasta()) alignment = dict(parse_fasta(res['MuscleOut'])) new_alignment = {} for k, v in alignment.items(): new_alignment[int_keys[k]] = v new_alignment = Alignment(new_alignment, MolType=moltype) res.cleanUp() del (seq_collection, int_map, int_keys, app, res, alignment, params) return new_alignment
align_unaligned_seqs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(seqs, moltype=DNA, params=None):\n if not params:\n params = {}\n seq_collection = SequenceCollection(seqs, MolType=moltype)\n int_map, int_keys = seq_collection.getIntMap()\n int_map = SequenceCollection(int_map, MolType=moltype)\n params.update({'-out': get_tmp_filename()})\n app = Muscle(InputHandler='_input_as_multiline_string', params=params,\n WorkingDir=tempfile.gettempdir())\n res = app(int_map.toFasta())\n alignment = dict(parse_fasta(res['MuscleOut']))\n new_alignment = {}\n for k, v in alignment.items():\n new_alignment[int_keys[k]] = v\n new_alignment = Alignment(new_alignment, MolType=moltype)\n res.cleanUp()\n del (seq_collection, int_map, int_keys, app, res, alignment, params)\n return new_alignment\n", "role": "user" }, { "content": "align_unaligned_seqs", "role": "assistant" } ]
def x(seqs, moltype, best_tree=False, params=None): aln = align_unaligned_seqs(seqs, moltype=moltype, params=params) tree = build_tree_from_alignment(aln, moltype, best_tree, params) return {'Align': aln, 'Tree': tree}
align_and_build_tree
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(seqs, moltype, best_tree=False, params=None):\n aln = align_unaligned_seqs(seqs, moltype=moltype, params=params)\n tree = build_tree_from_alignment(aln, moltype, best_tree, params)\n return {'Align': aln, 'Tree': tree}\n", "role": "user" }, { "content": "align_and_build_tree", "role": "assistant" } ]
def x(aln, moltype=DNA, best_tree=False, params=None): app = Muscle(InputHandler='_input_as_multiline_string', params=params, WorkingDir=tempfile.gettempdir()) app.Parameters['-clusteronly'].on() app.Parameters['-tree1'].on(get_tmp_filename(app.WorkingDir)) app.Parameters['-seqtype'].on(moltype.label) seq_collection = SequenceCollection(aln, MolType=moltype) int_map, int_keys = seq_collection.getIntMap() int_map = SequenceCollection(int_map, MolType=moltype) result = app(int_map.toFasta()) tree = DndParser(result['Tree1Out'].read(), constructor=PhyloNode) for tip in tree.tips(): tip.Name = int_keys[tip.Name] result.cleanUp() del (seq_collection, app, result) return tree
build_tree_from_alignment
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(aln, moltype=DNA, best_tree=False, params=None):\n app = Muscle(InputHandler='_input_as_multiline_string', params=params,\n WorkingDir=tempfile.gettempdir())\n app.Parameters['-clusteronly'].on()\n app.Parameters['-tree1'].on(get_tmp_filename(app.WorkingDir))\n app.Parameters['-seqtype'].on(moltype.label)\n seq_collection = SequenceCollection(aln, MolType=moltype)\n int_map, int_keys = seq_collection.getIntMap()\n int_map = SequenceCollection(int_map, MolType=moltype)\n result = app(int_map.toFasta())\n tree = DndParser(result['Tree1Out'].read(), constructor=PhyloNode)\n for tip in tree.tips():\n tip.Name = int_keys[tip.Name]\n result.cleanUp()\n del (seq_collection, app, result)\n return tree\n", "role": "user" }, { "content": "build_tree_from_alignment", "role": "assistant" } ]
def x(seqs, aln, params=None): if not params: params = {} seqs_collection = SequenceCollection(seqs) seqs_int_map, seqs_int_keys = seqs_collection.getIntMap(prefix='seq_') seqs_int_map = SequenceCollection(seqs_int_map) aln_collection = SequenceCollection(aln) aln_int_map, aln_int_keys = aln_collection.getIntMap(prefix='aln_') aln_int_map = SequenceCollection(aln_int_map) params.update({'-out': get_tmp_filename(), '-profile': True}) seqs_filename = get_tmp_filename() seqs_out = open(seqs_filename, 'w') seqs_out.write(seqs_int_map.toFasta()) seqs_out.close() aln_filename = get_tmp_filename() aln_out = open(aln_filename, 'w') aln_out.write(aln_int_map.toFasta()) aln_out.close() app = Muscle(InputHandler='_input_as_multifile', params=params, WorkingDir=tempfile.gettempdir()) res = app((aln_filename, seqs_filename)) alignment = dict(parse_fasta(res['MuscleOut'])) new_alignment = {} for k, v in alignment.items(): if k in seqs_int_keys: new_alignment[seqs_int_keys[k]] = v else: new_alignment[aln_int_keys[k]] = v new_alignment = Alignment(new_alignment) res.cleanUp() del (seqs_collection, seqs_int_map, seqs_int_keys) del (aln_collection, aln_int_map, aln_int_keys) del (app, res, alignment, params) remove(seqs_filename) remove(aln_filename) return new_alignment
add_seqs_to_alignment
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(seqs, aln, params=None):\n if not params:\n params = {}\n seqs_collection = SequenceCollection(seqs)\n seqs_int_map, seqs_int_keys = seqs_collection.getIntMap(prefix='seq_')\n seqs_int_map = SequenceCollection(seqs_int_map)\n aln_collection = SequenceCollection(aln)\n aln_int_map, aln_int_keys = aln_collection.getIntMap(prefix='aln_')\n aln_int_map = SequenceCollection(aln_int_map)\n params.update({'-out': get_tmp_filename(), '-profile': True})\n seqs_filename = get_tmp_filename()\n seqs_out = open(seqs_filename, 'w')\n seqs_out.write(seqs_int_map.toFasta())\n seqs_out.close()\n aln_filename = get_tmp_filename()\n aln_out = open(aln_filename, 'w')\n aln_out.write(aln_int_map.toFasta())\n aln_out.close()\n app = Muscle(InputHandler='_input_as_multifile', params=params,\n WorkingDir=tempfile.gettempdir())\n res = app((aln_filename, seqs_filename))\n alignment = dict(parse_fasta(res['MuscleOut']))\n new_alignment = {}\n for k, v in alignment.items():\n if k in seqs_int_keys:\n new_alignment[seqs_int_keys[k]] = v\n else:\n new_alignment[aln_int_keys[k]] = v\n new_alignment = Alignment(new_alignment)\n res.cleanUp()\n del (seqs_collection, seqs_int_map, seqs_int_keys)\n del (aln_collection, aln_int_map, aln_int_keys)\n del (app, res, alignment, params)\n remove(seqs_filename)\n remove(aln_filename)\n return new_alignment\n", "role": "user" }, { "content": "add_seqs_to_alignment", "role": "assistant" } ]
def x(aln1, aln2, params=None): if not params: params = {} aln1_collection = SequenceCollection(aln1) aln1_int_map, aln1_int_keys = aln1_collection.getIntMap(prefix='aln1_') aln1_int_map = SequenceCollection(aln1_int_map) aln2_collection = SequenceCollection(aln2) aln2_int_map, aln2_int_keys = aln2_collection.getIntMap(prefix='aln2_') aln2_int_map = SequenceCollection(aln2_int_map) params.update({'-out': get_tmp_filename(), '-profile': True}) aln1_filename = get_tmp_filename() aln1_out = open(aln1_filename, 'w') aln1_out.write(aln1_int_map.toFasta()) aln1_out.close() aln2_filename = get_tmp_filename() aln2_out = open(aln2_filename, 'w') aln2_out.write(aln2_int_map.toFasta()) aln2_out.close() app = Muscle(InputHandler='_input_as_multifile', params=params, WorkingDir=tempfile.gettempdir()) res = app((aln1_filename, aln2_filename)) alignment = dict(parse_fasta(res['MuscleOut'])) new_alignment = {} for k, v in alignment.items(): if k in aln1_int_keys: new_alignment[aln1_int_keys[k]] = v else: new_alignment[aln2_int_keys[k]] = v new_alignment = Alignment(new_alignment) res.cleanUp() del (aln1_collection, aln1_int_map, aln1_int_keys) del (aln2_collection, aln2_int_map, aln2_int_keys) del (app, res, alignment, params) remove(aln1_filename) remove(aln2_filename) return new_alignment
align_two_alignments
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(aln1, aln2, params=None):\n if not params:\n params = {}\n aln1_collection = SequenceCollection(aln1)\n aln1_int_map, aln1_int_keys = aln1_collection.getIntMap(prefix='aln1_')\n aln1_int_map = SequenceCollection(aln1_int_map)\n aln2_collection = SequenceCollection(aln2)\n aln2_int_map, aln2_int_keys = aln2_collection.getIntMap(prefix='aln2_')\n aln2_int_map = SequenceCollection(aln2_int_map)\n params.update({'-out': get_tmp_filename(), '-profile': True})\n aln1_filename = get_tmp_filename()\n aln1_out = open(aln1_filename, 'w')\n aln1_out.write(aln1_int_map.toFasta())\n aln1_out.close()\n aln2_filename = get_tmp_filename()\n aln2_out = open(aln2_filename, 'w')\n aln2_out.write(aln2_int_map.toFasta())\n aln2_out.close()\n app = Muscle(InputHandler='_input_as_multifile', params=params,\n WorkingDir=tempfile.gettempdir())\n res = app((aln1_filename, aln2_filename))\n alignment = dict(parse_fasta(res['MuscleOut']))\n new_alignment = {}\n for k, v in alignment.items():\n if k in aln1_int_keys:\n new_alignment[aln1_int_keys[k]] = v\n else:\n new_alignment[aln2_int_keys[k]] = v\n new_alignment = Alignment(new_alignment)\n res.cleanUp()\n del (aln1_collection, aln1_int_map, aln1_int_keys)\n del (aln2_collection, aln2_int_map, aln2_int_keys)\n del (app, res, alignment, params)\n remove(aln1_filename)\n remove(aln2_filename)\n return new_alignment\n", "role": "user" }, { "content": "align_two_alignments", "role": "assistant" } ]
def x(): features = random.sample(string.ascii_letters, NUM_SAMPLES) num_capitalized = len([letter for letter in features if letter in string.ascii_uppercase]) num_lowercase = len([letter for letter in features if letter in string. ascii_lowercase]) if num_capitalized > num_lowercase: label = 1 else: label = -1 return label, features
get_example
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n features = random.sample(string.ascii_letters, NUM_SAMPLES)\n num_capitalized = len([letter for letter in features if letter in\n string.ascii_uppercase])\n num_lowercase = len([letter for letter in features if letter in string.\n ascii_lowercase])\n if num_capitalized > num_lowercase:\n label = 1\n else:\n label = -1\n return label, features\n", "role": "user" }, { "content": "get_example", "role": "assistant" } ]
def x(redis, path): file_hash = get_cache_buster(redis, path) return '%s/%s?v=%s' % (oz.settings['static_host'], path, file_hash)
static_url
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(redis, path):\n file_hash = get_cache_buster(redis, path)\n return '%s/%s?v=%s' % (oz.settings['static_host'], path, file_hash)\n", "role": "user" }, { "content": "static_url", "role": "assistant" } ]
def x(redis, path): return escape.to_unicode(redis.hget('cache-buster:{}:v3'.format(oz. settings['s3_bucket']), path))
get_cache_buster
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(redis, path):\n return escape.to_unicode(redis.hget('cache-buster:{}:v3'.format(oz.\n settings['s3_bucket']), path))\n", "role": "user" }, { "content": "get_cache_buster", "role": "assistant" } ]
def x(redis, path, hash): redis.hset('cache-buster:{}:v3'.format(oz.settings['s3_bucket']), path, hash)
set_cache_buster
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(redis, path, hash):\n redis.hset('cache-buster:{}:v3'.format(oz.settings['s3_bucket']), path,\n hash)\n", "role": "user" }, { "content": "set_cache_buster", "role": "assistant" } ]
def x(s3_bucket=None, validate=False): global S3Connection if S3Connection != None: settings = oz.settings s3_bucket = s3_bucket or settings['s3_bucket'] opts = {} if settings['s3_host']: opts['host'] = settings['s3_host'] if settings['aws_access_key'] and settings['aws_secret_key']: opts['aws_access_key_id'] = settings['aws_access_key'] opts['aws_secret_access_key'] = settings['aws_secret_key'] return S3Connection(**opts).get_bucket(s3_bucket, validate=validate) else: raise Exception( 'S3 not supported in this environment as boto is not installed')
get_bucket
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(s3_bucket=None, validate=False):\n global S3Connection\n if S3Connection != None:\n settings = oz.settings\n s3_bucket = s3_bucket or settings['s3_bucket']\n opts = {}\n if settings['s3_host']:\n opts['host'] = settings['s3_host']\n if settings['aws_access_key'] and settings['aws_secret_key']:\n opts['aws_access_key_id'] = settings['aws_access_key']\n opts['aws_secret_access_key'] = settings['aws_secret_key']\n return S3Connection(**opts).get_bucket(s3_bucket, validate=validate)\n else:\n raise Exception(\n 'S3 not supported in this environment as boto is not installed')\n", "role": "user" }, { "content": "get_bucket", "role": "assistant" } ]
def x(path, s3_bucket=None): bucket_name = s3_bucket or oz.settings['s3_bucket'] if bucket_name: bucket = get_bucket(bucket_name) key = bucket.get_key(path) if not key: key = bucket.new_key(path) return S3File(key) else: return LocalFile(oz.settings['static_path'], path)
get_file
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(path, s3_bucket=None):\n bucket_name = s3_bucket or oz.settings['s3_bucket']\n if bucket_name:\n bucket = get_bucket(bucket_name)\n key = bucket.get_key(path)\n if not key:\n key = bucket.new_key(path)\n return S3File(key)\n else:\n return LocalFile(oz.settings['static_path'], path)\n", "role": "user" }, { "content": "get_file", "role": "assistant" } ]
def x(self, new_path, replace=False): new_full_path = os.path.join(self.static_path, new_path) if replace or not os.path.exists(new_full_path): shutil.copy2(self.full_path, new_full_path) return True return False
copy
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, new_path, replace=False):\n new_full_path = os.path.join(self.static_path, new_path)\n if replace or not os.path.exists(new_full_path):\n shutil.copy2(self.full_path, new_full_path)\n return True\n return False\n", "role": "user" }, { "content": "copy", "role": "assistant" } ]
def x(self, new_path, replace=False): if replace or not get_file(new_path).exists(): self.key.copy(self.key.bucket, new_path) return True return False
copy
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, new_path, replace=False):\n if replace or not get_file(new_path).exists():\n self.key.copy(self.key.bucket, new_path)\n return True\n return False\n", "role": "user" }, { "content": "copy", "role": "assistant" } ]
def x(): global _cached_connection settings = oz.settings if settings['redis_cache_connections'] and _cached_connection != None: return _cached_connection else: conn = redis.StrictRedis(host=settings['redis_host'], port=settings ['redis_port'], db=settings['redis_db'], password=settings[ 'redis_password'], decode_responses=settings[ 'redis_decode_responses'], ssl=settings['redis_use_ssl'], ssl_keyfile=settings['redis_ssl_keyfile'], ssl_certfile= settings['redis_ssl_certfile'], ssl_cert_reqs=settings[ 'redis_ssl_cert_reqs'], ssl_ca_certs=settings['redis_ssl_ca_certs'] ) if settings['redis_cache_connections']: _cached_connection = conn return conn
create_connection
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n global _cached_connection\n settings = oz.settings\n if settings['redis_cache_connections'] and _cached_connection != None:\n return _cached_connection\n else:\n conn = redis.StrictRedis(host=settings['redis_host'], port=settings\n ['redis_port'], db=settings['redis_db'], password=settings[\n 'redis_password'], decode_responses=settings[\n 'redis_decode_responses'], ssl=settings['redis_use_ssl'],\n ssl_keyfile=settings['redis_ssl_keyfile'], ssl_certfile=\n settings['redis_ssl_certfile'], ssl_cert_reqs=settings[\n 'redis_ssl_cert_reqs'], ssl_ca_certs=settings['redis_ssl_ca_certs']\n )\n if settings['redis_cache_connections']:\n _cached_connection = conn\n return conn\n", "role": "user" }, { "content": "create_connection", "role": "assistant" } ]
def x(uc_lines, otu_id_field=1): hit_miss_index = 0 cluster_id_index = otu_id_field seq_label_index = 8 otus = {} unassigned_seqs = [] for line in uc_lines: if line.startswith('#') or len(line.strip()) == 0: continue curr_line = line.split('\t') if curr_line[hit_miss_index] == 'N': unassigned_seqs.append(curr_line[seq_label_index].split()[0]) if curr_line[hit_miss_index] == 'H': curr_seq_label = curr_line[seq_label_index].split()[0] curr_otu_id = curr_line[cluster_id_index].split()[0] try: otus[curr_otu_id].append(curr_seq_label) except KeyError: otus[curr_otu_id] = [curr_seq_label] return otus, unassigned_seqs
clusters_from_blast_uc_file
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(uc_lines, otu_id_field=1):\n hit_miss_index = 0\n cluster_id_index = otu_id_field\n seq_label_index = 8\n otus = {}\n unassigned_seqs = []\n for line in uc_lines:\n if line.startswith('#') or len(line.strip()) == 0:\n continue\n curr_line = line.split('\\t')\n if curr_line[hit_miss_index] == 'N':\n unassigned_seqs.append(curr_line[seq_label_index].split()[0])\n if curr_line[hit_miss_index] == 'H':\n curr_seq_label = curr_line[seq_label_index].split()[0]\n curr_otu_id = curr_line[cluster_id_index].split()[0]\n try:\n otus[curr_otu_id].append(curr_seq_label)\n except KeyError:\n otus[curr_otu_id] = [curr_seq_label]\n return otus, unassigned_seqs\n", "role": "user" }, { "content": "clusters_from_blast_uc_file", "role": "assistant" } ]
def x(fasta_filepath, output_filepath=None, log_name='sortlen.log', HALT_EXEC=False, save_intermediate_files=False, remove_usearch_logs= False, working_dir=None): if not output_filepath: _, output_filepath = mkstemp(prefix='usearch_fasta_sort', suffix= '.fasta') log_filepath = join(working_dir, log_name) params = {} app = Usearch(params, WorkingDir=working_dir, HALT_EXEC=HALT_EXEC) data = {'--mergesort': fasta_filepath, '--output': output_filepath} if not remove_usearch_logs: data['--log'] = log_filepath app_result = app(data) return app_result, output_filepath
usearch_fasta_sort_from_filepath
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(fasta_filepath, output_filepath=None, log_name='sortlen.log',\n HALT_EXEC=False, save_intermediate_files=False, remove_usearch_logs=\n False, working_dir=None):\n if not output_filepath:\n _, output_filepath = mkstemp(prefix='usearch_fasta_sort', suffix=\n '.fasta')\n log_filepath = join(working_dir, log_name)\n params = {}\n app = Usearch(params, WorkingDir=working_dir, HALT_EXEC=HALT_EXEC)\n data = {'--mergesort': fasta_filepath, '--output': output_filepath}\n if not remove_usearch_logs:\n data['--log'] = log_filepath\n app_result = app(data)\n return app_result, output_filepath\n", "role": "user" }, { "content": "usearch_fasta_sort_from_filepath", "role": "assistant" } ]
def x(fasta_filepath, output_filepath=None, sizein=True, sizeout=True, minsize=0, log_name='abundance_sort.log', usersort=False, HALT_EXEC= False, save_intermediate_files=False, remove_usearch_logs=False, working_dir=None): if not output_filepath: _, output_filepath = mkstemp(prefix='usearch_abundance_sorted', suffix='.fasta') log_filepath = join(working_dir, 'minsize_' + str(minsize) + '_' + log_name ) params = {} app = Usearch(params, WorkingDir=working_dir, HALT_EXEC=HALT_EXEC) if usersort: app.Parameters['--usersort'].on() if minsize: app.Parameters['--minsize'].on(minsize) if sizein: app.Parameters['--sizein'].on() if sizeout: app.Parameters['--sizeout'].on() data = {'--sortsize': fasta_filepath, '--output': output_filepath} if not remove_usearch_logs: data['--log'] = log_filepath try: app_result = app(data) except ApplicationError: raise ValueError('No data following filter steps, please check ' + 'parameter settings for usearch_qf.') return app_result, output_filepath
usearch_sort_by_abundance
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(fasta_filepath, output_filepath=None, sizein=True, sizeout=True,\n minsize=0, log_name='abundance_sort.log', usersort=False, HALT_EXEC=\n False, save_intermediate_files=False, remove_usearch_logs=False,\n working_dir=None):\n if not output_filepath:\n _, output_filepath = mkstemp(prefix='usearch_abundance_sorted',\n suffix='.fasta')\n log_filepath = join(working_dir, 'minsize_' + str(minsize) + '_' + log_name\n )\n params = {}\n app = Usearch(params, WorkingDir=working_dir, HALT_EXEC=HALT_EXEC)\n if usersort:\n app.Parameters['--usersort'].on()\n if minsize:\n app.Parameters['--minsize'].on(minsize)\n if sizein:\n app.Parameters['--sizein'].on()\n if sizeout:\n app.Parameters['--sizeout'].on()\n data = {'--sortsize': fasta_filepath, '--output': output_filepath}\n if not remove_usearch_logs:\n data['--log'] = log_filepath\n try:\n app_result = app(data)\n except ApplicationError:\n raise ValueError('No data following filter steps, please check ' +\n 'parameter settings for usearch_qf.')\n return app_result, output_filepath\n", "role": "user" }, { "content": "usearch_sort_by_abundance", "role": "assistant" } ]
def x(fasta_filepath, output_filepath=None, output_uc_filepath=None, percent_id_err=0.97, sizein=True, sizeout=True, w=64, slots=16769023, maxrejects=64, log_name='usearch_cluster_err_corrected.log', usersort= False, HALT_EXEC=False, save_intermediate_files=False, remove_usearch_logs=False, working_dir=None): if not output_filepath: _, output_filepath = mkstemp(prefix='usearch_cluster_err_corrected', suffix='.fasta') log_filepath = join(working_dir, log_name) params = {'--sizein': sizein, '--sizeout': sizeout, '--id': percent_id_err, '--w': w, '--slots': slots, '--maxrejects': maxrejects} app = Usearch(params, WorkingDir=working_dir, HALT_EXEC=HALT_EXEC) if usersort: app.Parameters['--usersort'].on() data = {'--cluster': fasta_filepath, '--consout': output_filepath} if not remove_usearch_logs: data['--log'] = log_filepath if output_uc_filepath: data['--uc'] = output_uc_filepath app_result = app(data) return app_result, output_filepath
usearch_cluster_error_correction
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(fasta_filepath, output_filepath=None, output_uc_filepath=None,\n percent_id_err=0.97, sizein=True, sizeout=True, w=64, slots=16769023,\n maxrejects=64, log_name='usearch_cluster_err_corrected.log', usersort=\n False, HALT_EXEC=False, save_intermediate_files=False,\n remove_usearch_logs=False, working_dir=None):\n if not output_filepath:\n _, output_filepath = mkstemp(prefix='usearch_cluster_err_corrected',\n suffix='.fasta')\n log_filepath = join(working_dir, log_name)\n params = {'--sizein': sizein, '--sizeout': sizeout, '--id':\n percent_id_err, '--w': w, '--slots': slots, '--maxrejects': maxrejects}\n app = Usearch(params, WorkingDir=working_dir, HALT_EXEC=HALT_EXEC)\n if usersort:\n app.Parameters['--usersort'].on()\n data = {'--cluster': fasta_filepath, '--consout': output_filepath}\n if not remove_usearch_logs:\n data['--log'] = log_filepath\n if output_uc_filepath:\n data['--uc'] = output_uc_filepath\n app_result = app(data)\n return app_result, output_filepath\n", "role": "user" }, { "content": "usearch_cluster_error_correction", "role": "assistant" } ]
def x(fasta_filepath, output_chimera_filepath=None, output_non_chimera_filepath=None, abundance_skew=2.0, log_name= 'uchime_de_novo_chimera_filtering.log', usersort=False, HALT_EXEC=False, save_intermediate_files=False, remove_usearch_logs=False, working_dir=None ): if not output_chimera_filepath: _, output_chimera_filepath = mkstemp(prefix='uchime_chimeras_', suffix='.fasta') if not output_non_chimera_filepath: _, output_non_chimera_filepath = mkstemp(prefix= 'uchime_non_chimeras_', suffix='.fasta') log_filepath = join(working_dir, log_name) params = {'--abskew': abundance_skew} app = Usearch(params, WorkingDir=working_dir, HALT_EXEC=HALT_EXEC) if usersort: app.Parameters['--usersort'].on() data = {'--uchime': fasta_filepath, '--chimeras': output_chimera_filepath, '--nonchimeras': output_non_chimera_filepath} if not remove_usearch_logs: data['--log'] = log_filepath app_result = app(data) if not save_intermediate_files: remove_files([output_chimera_filepath]) return app_result, output_non_chimera_filepath
usearch_chimera_filter_de_novo
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(fasta_filepath, output_chimera_filepath=None,\n output_non_chimera_filepath=None, abundance_skew=2.0, log_name=\n 'uchime_de_novo_chimera_filtering.log', usersort=False, HALT_EXEC=False,\n save_intermediate_files=False, remove_usearch_logs=False, working_dir=None\n ):\n if not output_chimera_filepath:\n _, output_chimera_filepath = mkstemp(prefix='uchime_chimeras_',\n suffix='.fasta')\n if not output_non_chimera_filepath:\n _, output_non_chimera_filepath = mkstemp(prefix=\n 'uchime_non_chimeras_', suffix='.fasta')\n log_filepath = join(working_dir, log_name)\n params = {'--abskew': abundance_skew}\n app = Usearch(params, WorkingDir=working_dir, HALT_EXEC=HALT_EXEC)\n if usersort:\n app.Parameters['--usersort'].on()\n data = {'--uchime': fasta_filepath, '--chimeras':\n output_chimera_filepath, '--nonchimeras': output_non_chimera_filepath}\n if not remove_usearch_logs:\n data['--log'] = log_filepath\n app_result = app(data)\n if not save_intermediate_files:\n remove_files([output_chimera_filepath])\n return app_result, output_non_chimera_filepath\n", "role": "user" }, { "content": "usearch_chimera_filter_de_novo", "role": "assistant" } ]
def x(fasta_filepath, output_filepath=None, percent_id=0.97, sizein=True, sizeout=True, w=64, slots=16769023, maxrejects=64, log_name= 'usearch_cluster_seqs.log', usersort=True, HALT_EXEC=False, save_intermediate_files=False, remove_usearch_logs=False, suppress_new_clusters=False, refseqs_fp=None, output_dir=None, working_dir=None, rev=False): if not output_filepath: _, output_filepath = mkstemp(prefix='usearch_cluster_ref_based', suffix='.uc') log_filepath = join(working_dir, log_name) uc_filepath = join(working_dir, 'clustered_seqs_post_chimera.uc') params = {'--sizein': sizein, '--sizeout': sizeout, '--id': percent_id, '--w': w, '--slots': slots, '--maxrejects': maxrejects} app = Usearch(params, WorkingDir=working_dir, HALT_EXEC=HALT_EXEC) if usersort: app.Parameters['--usersort'].on() if rev: app.Parameters['--rev'].on() data = {'--query': fasta_filepath, '--uc': uc_filepath, '--db': refseqs_fp} if not remove_usearch_logs: data['--log'] = log_filepath app_result = app(data) files_to_remove = [] if suppress_new_clusters: output_fna_filepath = join(output_dir, 'ref_clustered_seqs.fasta') output_filepath, labels_hits = get_fasta_from_uc_file(fasta_filepath, uc_filepath, hit_type='H', output_dir=output_dir, output_fna_filepath=output_fna_filepath) files_to_remove.append(uc_filepath) else: output_fna_clustered = join(output_dir, 'ref_clustered_seqs.fasta') output_filepath_ref_clusters, labels_hits = get_fasta_from_uc_file( fasta_filepath, uc_filepath, hit_type='H', output_dir= output_dir, output_fna_filepath=output_fna_clustered) output_fna_failures = join(output_dir, 'ref_clustered_seqs_failures.fasta') output_filepath_failures, labels_hits = get_fasta_from_uc_file( fasta_filepath, uc_filepath, hit_type='N', output_dir= output_dir, output_fna_filepath=output_fna_failures) app_result, output_filepath_clustered_failures = usearch_cluster_seqs( output_fna_failures, output_filepath=join(output_dir, 'clustered_seqs_reference_failures.fasta'), percent_id= percent_id, sizein=sizein, sizeout=sizeout, w=w, slots=slots, maxrejects=maxrejects, save_intermediate_files= save_intermediate_files, remove_usearch_logs= remove_usearch_logs, working_dir=working_dir) output_filepath = concatenate_fastas(output_fna_clustered, output_fna_failures, output_concat_filepath=join(output_dir, 'concatenated_reference_denovo_clusters.fasta')) files_to_remove.append(output_fna_clustered) files_to_remove.append(output_fna_failures) files_to_remove.append(output_filepath_clustered_failures) if not save_intermediate_files: remove_files(files_to_remove) return app_result, output_filepath
usearch_cluster_seqs_ref
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(fasta_filepath, output_filepath=None, percent_id=0.97, sizein=True,\n sizeout=True, w=64, slots=16769023, maxrejects=64, log_name=\n 'usearch_cluster_seqs.log', usersort=True, HALT_EXEC=False,\n save_intermediate_files=False, remove_usearch_logs=False,\n suppress_new_clusters=False, refseqs_fp=None, output_dir=None,\n working_dir=None, rev=False):\n if not output_filepath:\n _, output_filepath = mkstemp(prefix='usearch_cluster_ref_based',\n suffix='.uc')\n log_filepath = join(working_dir, log_name)\n uc_filepath = join(working_dir, 'clustered_seqs_post_chimera.uc')\n params = {'--sizein': sizein, '--sizeout': sizeout, '--id': percent_id,\n '--w': w, '--slots': slots, '--maxrejects': maxrejects}\n app = Usearch(params, WorkingDir=working_dir, HALT_EXEC=HALT_EXEC)\n if usersort:\n app.Parameters['--usersort'].on()\n if rev:\n app.Parameters['--rev'].on()\n data = {'--query': fasta_filepath, '--uc': uc_filepath, '--db': refseqs_fp}\n if not remove_usearch_logs:\n data['--log'] = log_filepath\n app_result = app(data)\n files_to_remove = []\n if suppress_new_clusters:\n output_fna_filepath = join(output_dir, 'ref_clustered_seqs.fasta')\n output_filepath, labels_hits = get_fasta_from_uc_file(fasta_filepath,\n uc_filepath, hit_type='H', output_dir=output_dir,\n output_fna_filepath=output_fna_filepath)\n files_to_remove.append(uc_filepath)\n else:\n output_fna_clustered = join(output_dir, 'ref_clustered_seqs.fasta')\n output_filepath_ref_clusters, labels_hits = get_fasta_from_uc_file(\n fasta_filepath, uc_filepath, hit_type='H', output_dir=\n output_dir, output_fna_filepath=output_fna_clustered)\n output_fna_failures = join(output_dir,\n 'ref_clustered_seqs_failures.fasta')\n output_filepath_failures, labels_hits = get_fasta_from_uc_file(\n fasta_filepath, uc_filepath, hit_type='N', output_dir=\n output_dir, output_fna_filepath=output_fna_failures)\n app_result, output_filepath_clustered_failures = usearch_cluster_seqs(\n output_fna_failures, output_filepath=join(output_dir,\n 'clustered_seqs_reference_failures.fasta'), percent_id=\n percent_id, sizein=sizein, sizeout=sizeout, w=w, slots=slots,\n maxrejects=maxrejects, save_intermediate_files=\n save_intermediate_files, remove_usearch_logs=\n remove_usearch_logs, working_dir=working_dir)\n output_filepath = concatenate_fastas(output_fna_clustered,\n output_fna_failures, output_concat_filepath=join(output_dir,\n 'concatenated_reference_denovo_clusters.fasta'))\n files_to_remove.append(output_fna_clustered)\n files_to_remove.append(output_fna_failures)\n files_to_remove.append(output_filepath_clustered_failures)\n if not save_intermediate_files:\n remove_files(files_to_remove)\n return app_result, output_filepath\n", "role": "user" }, { "content": "usearch_cluster_seqs_ref", "role": "assistant" } ]
def x(output_fna_clustered, output_fna_failures, output_concat_filepath): output_fp = open(output_concat_filepath, 'w') for label, seq in parse_fasta(open(output_fna_clustered, 'U')): output_fp.write('>%s\n%s\n' % (label, seq)) for label, seq in parse_fasta(open(output_fna_failures, 'U')): output_fp.write('>%s\n%s\n' % (label, seq)) return output_concat_filepath
concatenate_fastas
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(output_fna_clustered, output_fna_failures, output_concat_filepath):\n output_fp = open(output_concat_filepath, 'w')\n for label, seq in parse_fasta(open(output_fna_clustered, 'U')):\n output_fp.write('>%s\\n%s\\n' % (label, seq))\n for label, seq in parse_fasta(open(output_fna_failures, 'U')):\n output_fp.write('>%s\\n%s\\n' % (label, seq))\n return output_concat_filepath\n", "role": "user" }, { "content": "concatenate_fastas", "role": "assistant" } ]
def x(fasta_filepath, output_filepath=None, label_prefix='', label_suffix= '', retain_label_as_comment=False, count_start=0): fasta_i = open(fasta_filepath, 'U') if not output_filepath: _, output_filepath = mkstemp(prefix='enumerated_seqs_', suffix='.fasta' ) fasta_o = open(output_filepath, 'w') for label, seq in parse_fasta(fasta_i): curr_label = '>' + label_prefix + str(count_start) + label_suffix if retain_label_as_comment: curr_label += '\t' + label fasta_o.write(curr_label.strip() + '\n') fasta_o.write(seq.strip() + '\n') count_start += 1 return output_filepath
enumerate_otus
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(fasta_filepath, output_filepath=None, label_prefix='', label_suffix=\n '', retain_label_as_comment=False, count_start=0):\n fasta_i = open(fasta_filepath, 'U')\n if not output_filepath:\n _, output_filepath = mkstemp(prefix='enumerated_seqs_', suffix='.fasta'\n )\n fasta_o = open(output_filepath, 'w')\n for label, seq in parse_fasta(fasta_i):\n curr_label = '>' + label_prefix + str(count_start) + label_suffix\n if retain_label_as_comment:\n curr_label += '\\t' + label\n fasta_o.write(curr_label.strip() + '\\n')\n fasta_o.write(seq.strip() + '\\n')\n count_start += 1\n return output_filepath\n", "role": "user" }, { "content": "enumerate_otus", "role": "assistant" } ]
def x(fasta_filepath, uc_filepath, hit_type='H', output_fna_filepath=None, label_prefix='', output_dir=None): hit_type_index = 0 seq_label_index = 8 target_label_index = 9 labels_hits = {} labels_to_keep = [] for line in open(uc_filepath, 'U'): if line.startswith('#') or len(line.strip()) == 0: continue curr_line = line.split('\t') if curr_line[0] == hit_type: labels_hits[curr_line[seq_label_index]] = curr_line[ target_label_index].strip() labels_to_keep.append(curr_line[seq_label_index]) labels_to_keep = set(labels_to_keep) out_fna = open(output_fna_filepath, 'w') for label, seq in parse_fasta(open(fasta_filepath, 'U')): if label in labels_to_keep: if hit_type == 'H': out_fna.write('>' + labels_hits[label] + '\n%s\n' % seq) if hit_type == 'N': out_fna.write('>' + label + '\n%s\n' % seq) return output_fna_filepath, labels_hits
get_fasta_from_uc_file
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(fasta_filepath, uc_filepath, hit_type='H', output_fna_filepath=None,\n label_prefix='', output_dir=None):\n hit_type_index = 0\n seq_label_index = 8\n target_label_index = 9\n labels_hits = {}\n labels_to_keep = []\n for line in open(uc_filepath, 'U'):\n if line.startswith('#') or len(line.strip()) == 0:\n continue\n curr_line = line.split('\\t')\n if curr_line[0] == hit_type:\n labels_hits[curr_line[seq_label_index]] = curr_line[\n target_label_index].strip()\n labels_to_keep.append(curr_line[seq_label_index])\n labels_to_keep = set(labels_to_keep)\n out_fna = open(output_fna_filepath, 'w')\n for label, seq in parse_fasta(open(fasta_filepath, 'U')):\n if label in labels_to_keep:\n if hit_type == 'H':\n out_fna.write('>' + labels_hits[label] + '\\n%s\\n' % seq)\n if hit_type == 'N':\n out_fna.write('>' + label + '\\n%s\\n' % seq)\n return output_fna_filepath, labels_hits\n", "role": "user" }, { "content": "get_fasta_from_uc_file", "role": "assistant" } ]
def x(output_fp_de_novo_nonchimeras, output_fp_ref_nonchimeras, output_combined_fp, chimeras_retention='union'): de_novo_non_chimeras = [] reference_non_chimeras = [] de_novo_nonchimeras_f = open(output_fp_de_novo_nonchimeras, 'U') reference_nonchimeras_f = open(output_fp_ref_nonchimeras, 'U') output_combined_f = open(output_combined_fp, 'w') for label, seq in parse_fasta(de_novo_nonchimeras_f): de_novo_non_chimeras.append(label) de_novo_nonchimeras_f.close() for label, seq in parse_fasta(reference_nonchimeras_f): reference_non_chimeras.append(label) reference_nonchimeras_f.close() de_novo_non_chimeras = set(de_novo_non_chimeras) reference_non_chimeras = set(reference_non_chimeras) if chimeras_retention == 'union': all_non_chimeras = de_novo_non_chimeras.union(reference_non_chimeras) elif chimeras_retention == 'intersection': all_non_chimeras = de_novo_non_chimeras.intersection( reference_non_chimeras) de_novo_nonchimeras_f = open(output_fp_de_novo_nonchimeras, 'U') reference_nonchimeras_f = open(output_fp_ref_nonchimeras, 'U') labels_written = [] for label, seq in parse_fasta(de_novo_nonchimeras_f): if label in all_non_chimeras: if label not in labels_written: output_combined_f.write('>%s\n%s\n' % (label, seq)) labels_written.append(label) de_novo_nonchimeras_f.close() for label, seq in parse_fasta(reference_nonchimeras_f): if label in all_non_chimeras: if label not in labels_written: output_combined_f.write('>%s\n%s\n' % (label, seq)) labels_written.append(label) reference_nonchimeras_f.close() output_combined_f.close() return output_combined_fp
get_retained_chimeras
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(output_fp_de_novo_nonchimeras, output_fp_ref_nonchimeras,\n output_combined_fp, chimeras_retention='union'):\n de_novo_non_chimeras = []\n reference_non_chimeras = []\n de_novo_nonchimeras_f = open(output_fp_de_novo_nonchimeras, 'U')\n reference_nonchimeras_f = open(output_fp_ref_nonchimeras, 'U')\n output_combined_f = open(output_combined_fp, 'w')\n for label, seq in parse_fasta(de_novo_nonchimeras_f):\n de_novo_non_chimeras.append(label)\n de_novo_nonchimeras_f.close()\n for label, seq in parse_fasta(reference_nonchimeras_f):\n reference_non_chimeras.append(label)\n reference_nonchimeras_f.close()\n de_novo_non_chimeras = set(de_novo_non_chimeras)\n reference_non_chimeras = set(reference_non_chimeras)\n if chimeras_retention == 'union':\n all_non_chimeras = de_novo_non_chimeras.union(reference_non_chimeras)\n elif chimeras_retention == 'intersection':\n all_non_chimeras = de_novo_non_chimeras.intersection(\n reference_non_chimeras)\n de_novo_nonchimeras_f = open(output_fp_de_novo_nonchimeras, 'U')\n reference_nonchimeras_f = open(output_fp_ref_nonchimeras, 'U')\n labels_written = []\n for label, seq in parse_fasta(de_novo_nonchimeras_f):\n if label in all_non_chimeras:\n if label not in labels_written:\n output_combined_f.write('>%s\\n%s\\n' % (label, seq))\n labels_written.append(label)\n de_novo_nonchimeras_f.close()\n for label, seq in parse_fasta(reference_nonchimeras_f):\n if label in all_non_chimeras:\n if label not in labels_written:\n output_combined_f.write('>%s\\n%s\\n' % (label, seq))\n labels_written.append(label)\n reference_nonchimeras_f.close()\n output_combined_f.close()\n return output_combined_fp\n", "role": "user" }, { "content": "get_retained_chimeras", "role": "assistant" } ]
def x(original_fasta, filtered_fasta, output_filepath=None, log_name= 'assign_reads_to_otus.log', perc_id_blast=0.97, global_alignment=True, HALT_EXEC=False, save_intermediate_files=False, remove_usearch_logs= False, working_dir=None): if not output_filepath: _, output_filepath = mkstemp(prefix='assign_reads_to_otus', suffix= '.uc') log_filepath = join(working_dir, log_name) params = {'--id': perc_id_blast, '--global': global_alignment} app = Usearch(params, WorkingDir=working_dir, HALT_EXEC=HALT_EXEC) data = {'--query': original_fasta, '--db': filtered_fasta, '--uc': output_filepath} if not remove_usearch_logs: data['--log'] = log_filepath app_result = app(data) return app_result, output_filepath
assign_reads_to_otus
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(original_fasta, filtered_fasta, output_filepath=None, log_name=\n 'assign_reads_to_otus.log', perc_id_blast=0.97, global_alignment=True,\n HALT_EXEC=False, save_intermediate_files=False, remove_usearch_logs=\n False, working_dir=None):\n if not output_filepath:\n _, output_filepath = mkstemp(prefix='assign_reads_to_otus', suffix=\n '.uc')\n log_filepath = join(working_dir, log_name)\n params = {'--id': perc_id_blast, '--global': global_alignment}\n app = Usearch(params, WorkingDir=working_dir, HALT_EXEC=HALT_EXEC)\n data = {'--query': original_fasta, '--db': filtered_fasta, '--uc':\n output_filepath}\n if not remove_usearch_logs:\n data['--log'] = log_filepath\n app_result = app(data)\n return app_result, output_filepath\n", "role": "user" }, { "content": "assign_reads_to_otus", "role": "assistant" } ]
def x(seq_path, output_dir='.', rev=False, minlen=64, remove_usearch_logs= False, HALT_EXEC=False, output_fna_filepath=None, output_uc_filepath= None, log_name='abundance_sorted.log', threads=1.0): if not output_fna_filepath: _, output_fna_filepath = mkstemp(prefix='abundance_sorted', suffix= '.fna') if not output_uc_filepath: _, output_uc_filepath = mkstemp(prefix='abundance_sorted', suffix='.uc' ) log_filepath = join(output_dir, log_name) params = {'--minseqlength': minlen, '--sizeout': True, '--derep_fulllength': seq_path, '--output': output_fna_filepath, '--uc': output_uc_filepath, '--threads': threads} if rev: params['--strand'] = 'both' if not remove_usearch_logs: params['--log'] = log_filepath app = Usearch61(params, WorkingDir=output_dir, HALT_EXEC=HALT_EXEC) app_result = app() return output_fna_filepath, output_uc_filepath, app_result
sort_by_abundance_usearch61
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(seq_path, output_dir='.', rev=False, minlen=64, remove_usearch_logs=\n False, HALT_EXEC=False, output_fna_filepath=None, output_uc_filepath=\n None, log_name='abundance_sorted.log', threads=1.0):\n if not output_fna_filepath:\n _, output_fna_filepath = mkstemp(prefix='abundance_sorted', suffix=\n '.fna')\n if not output_uc_filepath:\n _, output_uc_filepath = mkstemp(prefix='abundance_sorted', suffix='.uc'\n )\n log_filepath = join(output_dir, log_name)\n params = {'--minseqlength': minlen, '--sizeout': True,\n '--derep_fulllength': seq_path, '--output': output_fna_filepath,\n '--uc': output_uc_filepath, '--threads': threads}\n if rev:\n params['--strand'] = 'both'\n if not remove_usearch_logs:\n params['--log'] = log_filepath\n app = Usearch61(params, WorkingDir=output_dir, HALT_EXEC=HALT_EXEC)\n app_result = app()\n return output_fna_filepath, output_uc_filepath, app_result\n", "role": "user" }, { "content": "sort_by_abundance_usearch61", "role": "assistant" } ]
def x(seq_path, output_dir='.', minlen=64, remove_usearch_logs=False, HALT_EXEC=False, output_fna_filepath=None, log_name='length_sorted.log'): if not output_fna_filepath: _, output_fna_filepath = mkstemp(prefix='length_sorted', suffix='.fna') log_filepath = join(output_dir, log_name) params = {'--minseqlength': minlen, '--sortbylength': seq_path, '--output': output_fna_filepath} if not remove_usearch_logs: params['--log'] = log_filepath app = Usearch61(params, WorkingDir=output_dir, HALT_EXEC=HALT_EXEC) app_result = app() return output_fna_filepath, app_result
sort_by_length_usearch61
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(seq_path, output_dir='.', minlen=64, remove_usearch_logs=False,\n HALT_EXEC=False, output_fna_filepath=None, log_name='length_sorted.log'):\n if not output_fna_filepath:\n _, output_fna_filepath = mkstemp(prefix='length_sorted', suffix='.fna')\n log_filepath = join(output_dir, log_name)\n params = {'--minseqlength': minlen, '--sortbylength': seq_path,\n '--output': output_fna_filepath}\n if not remove_usearch_logs:\n params['--log'] = log_filepath\n app = Usearch61(params, WorkingDir=output_dir, HALT_EXEC=HALT_EXEC)\n app_result = app()\n return output_fna_filepath, app_result\n", "role": "user" }, { "content": "sort_by_length_usearch61", "role": "assistant" } ]
def x(intermediate_fasta, refseqs_fp, percent_id=0.97, rev=False, minlen=64, output_dir='.', remove_usearch_logs=False, wordlength=8, usearch61_maxrejects=32, usearch61_maxaccepts=1, HALT_EXEC=False, output_uc_filepath=None, log_filepath='ref_clustered.log', threads=1.0): log_filepath = join(output_dir, log_filepath) params = {'--usearch_global': intermediate_fasta, '--db': refseqs_fp, '--minseqlength': minlen, '--id': percent_id, '--uc': output_uc_filepath, '--wordlength': wordlength, '--maxrejects': usearch61_maxrejects, '--maxaccepts': usearch61_maxaccepts, '--threads': threads} if not remove_usearch_logs: params['--log'] = log_filepath if rev: params['--strand'] = 'both' else: params['--strand'] = 'plus' clusters_fp = output_uc_filepath app = Usearch61(params, WorkingDir=output_dir, HALT_EXEC=HALT_EXEC) app_result = app() return clusters_fp, app_result
usearch61_cluster_ref
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(intermediate_fasta, refseqs_fp, percent_id=0.97, rev=False, minlen=64,\n output_dir='.', remove_usearch_logs=False, wordlength=8,\n usearch61_maxrejects=32, usearch61_maxaccepts=1, HALT_EXEC=False,\n output_uc_filepath=None, log_filepath='ref_clustered.log', threads=1.0):\n log_filepath = join(output_dir, log_filepath)\n params = {'--usearch_global': intermediate_fasta, '--db': refseqs_fp,\n '--minseqlength': minlen, '--id': percent_id, '--uc':\n output_uc_filepath, '--wordlength': wordlength, '--maxrejects':\n usearch61_maxrejects, '--maxaccepts': usearch61_maxaccepts,\n '--threads': threads}\n if not remove_usearch_logs:\n params['--log'] = log_filepath\n if rev:\n params['--strand'] = 'both'\n else:\n params['--strand'] = 'plus'\n clusters_fp = output_uc_filepath\n app = Usearch61(params, WorkingDir=output_dir, HALT_EXEC=HALT_EXEC)\n app_result = app()\n return clusters_fp, app_result\n", "role": "user" }, { "content": "usearch61_cluster_ref", "role": "assistant" } ]
def x(intermediate_fasta, percent_id=0.97, minlen=64, output_dir='.', remove_usearch_logs=False, wordlength=8, usearch61_maxrejects=8, usearch61_maxaccepts=1, HALT_EXEC=False, output_uc_filepath=None, log_name='fast_clustered.log', threads=1.0): log_filepath = join(output_dir, log_name) params = {'--minseqlength': minlen, '--cluster_fast': intermediate_fasta, '--id': percent_id, '--uc': output_uc_filepath, '--wordlength': wordlength, '--maxrejects': usearch61_maxrejects, '--maxaccepts': usearch61_maxaccepts, '--usersort': True, '--threads': threads} if not remove_usearch_logs: params['--log'] = log_filepath clusters_fp = output_uc_filepath app = Usearch61(params, WorkingDir=output_dir, HALT_EXEC=HALT_EXEC) app_result = app() return clusters_fp, app_result
usearch61_fast_cluster
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(intermediate_fasta, percent_id=0.97, minlen=64, output_dir='.',\n remove_usearch_logs=False, wordlength=8, usearch61_maxrejects=8,\n usearch61_maxaccepts=1, HALT_EXEC=False, output_uc_filepath=None,\n log_name='fast_clustered.log', threads=1.0):\n log_filepath = join(output_dir, log_name)\n params = {'--minseqlength': minlen, '--cluster_fast':\n intermediate_fasta, '--id': percent_id, '--uc': output_uc_filepath,\n '--wordlength': wordlength, '--maxrejects': usearch61_maxrejects,\n '--maxaccepts': usearch61_maxaccepts, '--usersort': True,\n '--threads': threads}\n if not remove_usearch_logs:\n params['--log'] = log_filepath\n clusters_fp = output_uc_filepath\n app = Usearch61(params, WorkingDir=output_dir, HALT_EXEC=HALT_EXEC)\n app_result = app()\n return clusters_fp, app_result\n", "role": "user" }, { "content": "usearch61_fast_cluster", "role": "assistant" } ]
def x(intermediate_fasta, percent_id=0.97, minlen=64, rev=False, output_dir ='.', remove_usearch_logs=False, wordlength=8, usearch61_maxrejects=32, usearch61_maxaccepts=1, sizeorder=False, HALT_EXEC=False, output_uc_filepath=None, log_name='smallmem_clustered.log', sizeout= False, consout_filepath=None): log_filepath = join(output_dir, log_name) params = {'--minseqlength': minlen, '--cluster_smallmem': intermediate_fasta, '--id': percent_id, '--uc': output_uc_filepath, '--wordlength': wordlength, '--maxrejects': usearch61_maxrejects, '--maxaccepts': usearch61_maxaccepts, '--usersort': True} if sizeorder: params['--sizeorder'] = True if not remove_usearch_logs: params['--log'] = log_filepath if rev: params['--strand'] = 'both' else: params['--strand'] = 'plus' if sizeout: params['--sizeout'] = True if consout_filepath: params['--consout'] = consout_filepath clusters_fp = output_uc_filepath app = Usearch61(params, WorkingDir=output_dir, HALT_EXEC=HALT_EXEC) app_result = app() return clusters_fp, app_result
usearch61_smallmem_cluster
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(intermediate_fasta, percent_id=0.97, minlen=64, rev=False, output_dir\n ='.', remove_usearch_logs=False, wordlength=8, usearch61_maxrejects=32,\n usearch61_maxaccepts=1, sizeorder=False, HALT_EXEC=False,\n output_uc_filepath=None, log_name='smallmem_clustered.log', sizeout=\n False, consout_filepath=None):\n log_filepath = join(output_dir, log_name)\n params = {'--minseqlength': minlen, '--cluster_smallmem':\n intermediate_fasta, '--id': percent_id, '--uc': output_uc_filepath,\n '--wordlength': wordlength, '--maxrejects': usearch61_maxrejects,\n '--maxaccepts': usearch61_maxaccepts, '--usersort': True}\n if sizeorder:\n params['--sizeorder'] = True\n if not remove_usearch_logs:\n params['--log'] = log_filepath\n if rev:\n params['--strand'] = 'both'\n else:\n params['--strand'] = 'plus'\n if sizeout:\n params['--sizeout'] = True\n if consout_filepath:\n params['--consout'] = consout_filepath\n clusters_fp = output_uc_filepath\n app = Usearch61(params, WorkingDir=output_dir, HALT_EXEC=HALT_EXEC)\n app_result = app()\n return clusters_fp, app_result\n", "role": "user" }, { "content": "usearch61_smallmem_cluster", "role": "assistant" } ]
def x(abundance_fp, uchime_denovo_fp, minlen=64, output_dir='.', remove_usearch_logs=False, uchime_denovo_log_fp='uchime_denovo.log', usearch61_minh=0.28, usearch61_xn=8.0, usearch61_dn=1.4, usearch61_mindiffs=3, usearch61_mindiv=0.8, usearch61_abundance_skew= 2.0, HALT_EXEC=False): params = {'--minseqlength': minlen, '--uchime_denovo': abundance_fp, '--uchimeout': uchime_denovo_fp, '--minh': usearch61_minh, '--xn': usearch61_xn, '--dn': usearch61_dn, '--mindiffs': usearch61_mindiffs, '--mindiv': usearch61_mindiv, '--abskew': usearch61_abundance_skew} if not remove_usearch_logs: params['--log'] = uchime_denovo_log_fp app = Usearch61(params, WorkingDir=output_dir, HALT_EXEC=HALT_EXEC) app_result = app() return uchime_denovo_fp, app_result
usearch61_chimera_check_denovo
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(abundance_fp, uchime_denovo_fp, minlen=64, output_dir='.',\n remove_usearch_logs=False, uchime_denovo_log_fp='uchime_denovo.log',\n usearch61_minh=0.28, usearch61_xn=8.0, usearch61_dn=1.4,\n usearch61_mindiffs=3, usearch61_mindiv=0.8, usearch61_abundance_skew=\n 2.0, HALT_EXEC=False):\n params = {'--minseqlength': minlen, '--uchime_denovo': abundance_fp,\n '--uchimeout': uchime_denovo_fp, '--minh': usearch61_minh, '--xn':\n usearch61_xn, '--dn': usearch61_dn, '--mindiffs':\n usearch61_mindiffs, '--mindiv': usearch61_mindiv, '--abskew':\n usearch61_abundance_skew}\n if not remove_usearch_logs:\n params['--log'] = uchime_denovo_log_fp\n app = Usearch61(params, WorkingDir=output_dir, HALT_EXEC=HALT_EXEC)\n app_result = app()\n return uchime_denovo_fp, app_result\n", "role": "user" }, { "content": "usearch61_chimera_check_denovo", "role": "assistant" } ]
def x(abundance_fp, uchime_ref_fp, reference_seqs_fp, minlen=64, output_dir ='.', remove_usearch_logs=False, uchime_ref_log_fp='uchime_ref.log', usearch61_minh=0.28, usearch61_xn=8.0, usearch61_dn=1.4, usearch61_mindiffs=3, usearch61_mindiv=0.8, threads=1.0, HALT_EXEC=False): params = {'--minseqlength': minlen, '--uchime_ref': abundance_fp, '--uchimeout': uchime_ref_fp, '--db': reference_seqs_fp, '--minh': usearch61_minh, '--xn': usearch61_xn, '--dn': usearch61_dn, '--mindiffs': usearch61_mindiffs, '--mindiv': usearch61_mindiv, '--strand': 'plus', '--threads': threads} if not remove_usearch_logs: params['--log'] = uchime_ref_log_fp app = Usearch61(params, WorkingDir=output_dir, HALT_EXEC=HALT_EXEC) app_result = app() return uchime_ref_fp, app_result
usearch61_chimera_check_ref
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(abundance_fp, uchime_ref_fp, reference_seqs_fp, minlen=64, output_dir\n ='.', remove_usearch_logs=False, uchime_ref_log_fp='uchime_ref.log',\n usearch61_minh=0.28, usearch61_xn=8.0, usearch61_dn=1.4,\n usearch61_mindiffs=3, usearch61_mindiv=0.8, threads=1.0, HALT_EXEC=False):\n params = {'--minseqlength': minlen, '--uchime_ref': abundance_fp,\n '--uchimeout': uchime_ref_fp, '--db': reference_seqs_fp, '--minh':\n usearch61_minh, '--xn': usearch61_xn, '--dn': usearch61_dn,\n '--mindiffs': usearch61_mindiffs, '--mindiv': usearch61_mindiv,\n '--strand': 'plus', '--threads': threads}\n if not remove_usearch_logs:\n params['--log'] = uchime_ref_log_fp\n app = Usearch61(params, WorkingDir=output_dir, HALT_EXEC=HALT_EXEC)\n app_result = app()\n return uchime_ref_fp, app_result\n", "role": "user" }, { "content": "usearch61_chimera_check_ref", "role": "assistant" } ]
def x(dereplicated_uc_lines): dereplicated_clusters = {} seed_hit_ix = 0 seq_id_ix = 8 seed_id_ix = 9 for line in dereplicated_uc_lines: if line.startswith('#') or len(line.strip()) == 0: continue curr_line = line.strip().split('\t') if curr_line[seed_hit_ix] == 'S': dereplicated_clusters[curr_line[seq_id_ix]] = [] if curr_line[seed_hit_ix] == 'H': curr_seq_id = curr_line[seq_id_ix] dereplicated_clusters[curr_line[seed_id_ix]].append(curr_seq_id) return dereplicated_clusters
parse_dereplicated_uc
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(dereplicated_uc_lines):\n dereplicated_clusters = {}\n seed_hit_ix = 0\n seq_id_ix = 8\n seed_id_ix = 9\n for line in dereplicated_uc_lines:\n if line.startswith('#') or len(line.strip()) == 0:\n continue\n curr_line = line.strip().split('\\t')\n if curr_line[seed_hit_ix] == 'S':\n dereplicated_clusters[curr_line[seq_id_ix]] = []\n if curr_line[seed_hit_ix] == 'H':\n curr_seq_id = curr_line[seq_id_ix]\n dereplicated_clusters[curr_line[seed_id_ix]].append(curr_seq_id)\n return dereplicated_clusters\n", "role": "user" }, { "content": "parse_dereplicated_uc", "role": "assistant" } ]
def x(clustered_uc_lines, otu_prefix='denovo', ref_clustered=False): clusters = {} failures = [] seed_hit_ix = 0 otu_id_ix = 1 seq_id_ix = 8 ref_id_ix = 9 for line in clustered_uc_lines: if line.startswith('#') or len(line.strip()) == 0: continue curr_line = line.strip().split('\t') if curr_line[seed_hit_ix] == 'S': clusters[otu_prefix + curr_line[otu_id_ix]] = [curr_line[ seq_id_ix].split(';')[0].split()[0]] if curr_line[seed_hit_ix] == 'H': curr_id = curr_line[seq_id_ix].split(';')[0].split()[0] if ref_clustered: try: clusters[otu_prefix + curr_line[ref_id_ix]].append(curr_id) except KeyError: clusters[otu_prefix + curr_line[ref_id_ix]] = [curr_id] else: clusters[otu_prefix + curr_line[otu_id_ix]].append(curr_id) if curr_line[seed_hit_ix] == 'N': failures.append(curr_line[seq_id_ix].split(';')[0]) return clusters, failures
parse_usearch61_clusters
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(clustered_uc_lines, otu_prefix='denovo', ref_clustered=False):\n clusters = {}\n failures = []\n seed_hit_ix = 0\n otu_id_ix = 1\n seq_id_ix = 8\n ref_id_ix = 9\n for line in clustered_uc_lines:\n if line.startswith('#') or len(line.strip()) == 0:\n continue\n curr_line = line.strip().split('\\t')\n if curr_line[seed_hit_ix] == 'S':\n clusters[otu_prefix + curr_line[otu_id_ix]] = [curr_line[\n seq_id_ix].split(';')[0].split()[0]]\n if curr_line[seed_hit_ix] == 'H':\n curr_id = curr_line[seq_id_ix].split(';')[0].split()[0]\n if ref_clustered:\n try:\n clusters[otu_prefix + curr_line[ref_id_ix]].append(curr_id)\n except KeyError:\n clusters[otu_prefix + curr_line[ref_id_ix]] = [curr_id]\n else:\n clusters[otu_prefix + curr_line[otu_id_ix]].append(curr_id)\n if curr_line[seed_hit_ix] == 'N':\n failures.append(curr_line[seq_id_ix].split(';')[0])\n return clusters, failures\n", "role": "user" }, { "content": "parse_usearch61_clusters", "role": "assistant" } ]
def x(de_novo_clusters, dereplicated_clusters): clusters = {} for curr_denovo_key in de_novo_clusters.keys(): clusters[curr_denovo_key] = de_novo_clusters[curr_denovo_key] curr_clusters = [] for curr_denovo_id in de_novo_clusters[curr_denovo_key]: curr_clusters += dereplicated_clusters[curr_denovo_id] clusters[curr_denovo_key] += curr_clusters return clusters
merge_clusters_dereplicated_seqs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(de_novo_clusters, dereplicated_clusters):\n clusters = {}\n for curr_denovo_key in de_novo_clusters.keys():\n clusters[curr_denovo_key] = de_novo_clusters[curr_denovo_key]\n curr_clusters = []\n for curr_denovo_id in de_novo_clusters[curr_denovo_key]:\n curr_clusters += dereplicated_clusters[curr_denovo_id]\n clusters[curr_denovo_key] += curr_clusters\n return clusters\n", "role": "user" }, { "content": "merge_clusters_dereplicated_seqs", "role": "assistant" } ]
def x(failures, dereplicated_clusters): curr_failures = set(failures) dereplicated_ids = set(dereplicated_clusters) for curr_failure in curr_failures: if curr_failure in dereplicated_ids: failures += dereplicated_clusters[curr_failure] return failures
merge_failures_dereplicated_seqs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(failures, dereplicated_clusters):\n curr_failures = set(failures)\n dereplicated_ids = set(dereplicated_clusters)\n for curr_failure in curr_failures:\n if curr_failure in dereplicated_ids:\n failures += dereplicated_clusters[curr_failure]\n return failures\n", "role": "user" }, { "content": "merge_failures_dereplicated_seqs", "role": "assistant" } ]
def x(seq_path, failures, output_fasta_fp): parsed_out = open(output_fasta_fp, 'w') for label, seq in parse_fasta(open(seq_path), 'U'): curr_label = label.split()[0] if curr_label in failures: parsed_out.write('>%s\n%s\n' % (label, seq)) parsed_out.close() return output_fasta_fp
parse_usearch61_failures
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(seq_path, failures, output_fasta_fp):\n parsed_out = open(output_fasta_fp, 'w')\n for label, seq in parse_fasta(open(seq_path), 'U'):\n curr_label = label.split()[0]\n if curr_label in failures:\n parsed_out.write('>%s\\n%s\\n' % (label, seq))\n parsed_out.close()\n return output_fasta_fp\n", "role": "user" }, { "content": "parse_usearch61_failures", "role": "assistant" } ]
def x(self, status_code, **kwargs): if oz.settings.get('debug'): exception_type, exception_value, tback = sys.exc_info() is_breakpoint = isinstance(exception_value, oz.error_pages. DebugBreakException) frames = oz.error_pages.get_frames(tback, is_breakpoint) frames.reverse() if is_breakpoint: exception_type = 'Debug breakpoint' exception_value = '' self.render(oz.settings['error_pages_template'], exception_type= exception_type, exception_value=exception_value, frames=frames, request_input=self.request.body, request_cookies=self.cookies, request_headers=self.request.headers, request_path=self.request .uri, request_method=self.request.method, response_output=''. join(self._write_buffer), response_headers=self._headers, prettify_object=oz.error_pages.prettify_object) return oz.break_trigger
on_error_page_write_error
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, status_code, **kwargs):\n if oz.settings.get('debug'):\n exception_type, exception_value, tback = sys.exc_info()\n is_breakpoint = isinstance(exception_value, oz.error_pages.\n DebugBreakException)\n frames = oz.error_pages.get_frames(tback, is_breakpoint)\n frames.reverse()\n if is_breakpoint:\n exception_type = 'Debug breakpoint'\n exception_value = ''\n self.render(oz.settings['error_pages_template'], exception_type=\n exception_type, exception_value=exception_value, frames=frames,\n request_input=self.request.body, request_cookies=self.cookies,\n request_headers=self.request.headers, request_path=self.request\n .uri, request_method=self.request.method, response_output=''.\n join(self._write_buffer), response_headers=self._headers,\n prettify_object=oz.error_pages.prettify_object)\n return oz.break_trigger\n", "role": "user" }, { "content": "on_error_page_write_error", "role": "assistant" } ]
def x(self, name): value = self.get_cookie(name) if value != None: self.clear_cookie(name) return escape.url_unescape(value)
get_blink_cookie
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, name):\n value = self.get_cookie(name)\n if value != None:\n self.clear_cookie(name)\n return escape.url_unescape(value)\n", "role": "user" }, { "content": "get_blink_cookie", "role": "assistant" } ]
def x(self, message, type='info'): self.set_cookie('blink_message', escape.url_escape(message), httponly=True) self.set_cookie('blink_type', escape.url_escape(type), httponly=True)
set_blink
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, message, type='info'):\n self.set_cookie('blink_message', escape.url_escape(message), httponly=True)\n self.set_cookie('blink_type', escape.url_escape(type), httponly=True)\n", "role": "user" }, { "content": "set_blink", "role": "assistant" } ]
def x(lines): clusters = [] curr_cluster = [] for l in lines: if l.startswith('>Cluster'): if not curr_cluster: continue clusters.append(curr_cluster) curr_cluster = [] else: curr_cluster.append(clean_cluster_seq_id(l.split()[2])) if curr_cluster: clusters.append(curr_cluster) return clusters
parse_cdhit_clstr_file
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(lines):\n clusters = []\n curr_cluster = []\n for l in lines:\n if l.startswith('>Cluster'):\n if not curr_cluster:\n continue\n clusters.append(curr_cluster)\n curr_cluster = []\n else:\n curr_cluster.append(clean_cluster_seq_id(l.split()[2]))\n if curr_cluster:\n clusters.append(curr_cluster)\n return clusters\n", "role": "user" }, { "content": "parse_cdhit_clstr_file", "role": "assistant" } ]
def x(self, data): if data: self.Parameters['-i'].on(super(CD_HIT, self). _input_as_multiline_string(data)) return ''
input_as_multiline_string
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, data):\n if data:\n self.Parameters['-i'].on(super(CD_HIT, self).\n _input_as_multiline_string(data))\n return ''\n", "role": "user" }, { "content": "input_as_multiline_string", "role": "assistant" } ]
def x(self, data): if data: self.Parameters['-i'].on(super(CD_HIT, self)._input_as_lines(data)) return ''
input_as_lines
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, data):\n if data:\n self.Parameters['-i'].on(super(CD_HIT, self)._input_as_lines(data))\n return ''\n", "role": "user" }, { "content": "input_as_lines", "role": "assistant" } ]
def x(self, data): result = {} result['FASTA'] = ResultPath(Path=self._get_seqs_outfile()) result['CLSTR'] = ResultPath(Path=self._get_clstr_outfile()) return result
get_result_paths
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, data):\n result = {}\n result['FASTA'] = ResultPath(Path=self._get_seqs_outfile())\n result['CLSTR'] = ResultPath(Path=self._get_clstr_outfile())\n return result\n", "role": "user" }, { "content": "get_result_paths", "role": "assistant" } ]
def x(self, tokens): look_ahead = LookAhead(tokens) segments = Segment() while not look_ahead.empty(): if look_ahead.peek().type not in self.whitespace: paragraph = MatchableSegment(look_ahead.i) while not look_ahead.empty() and look_ahead.peek( ).type not in self.paragraph_end: if look_ahead.peek().type == 'tab_open': tab_depth = 1 sentence = MatchableSegment(look_ahead.i, [next( look_ahead)]) while not look_ahead.empty() and tab_depth > 0: tab_depth += look_ahead.peek().type == 'tab_open' tab_depth -= look_ahead.peek().type == 'tab_close' sentence.append(next(look_ahead)) paragraph.append(sentence) elif look_ahead.peek().type not in self.whitespace: sentence = MatchableSegment(look_ahead.i, [next( look_ahead)]) sub_depth = int(sentence[0].type in SUB_OPEN) while not look_ahead.empty(): sub_depth += look_ahead.peek().type in SUB_OPEN sub_depth -= look_ahead.peek().type in SUB_CLOSE sentence.append(next(look_ahead)) if sentence[-1 ].type in self.sentence_end and sub_depth <= 0: non_whitespace = sum(s.type not in self. whitespace for s in sentence) if non_whitespace >= self.min_sentence: break paragraph.append(sentence) else: whitespace = Segment(look_ahead.i, [next(look_ahead)]) paragraph.append(whitespace) segments.append(paragraph) else: whitespace = Segment(look_ahead.i, [next(look_ahead)]) segments.append(whitespace) return segments
segment
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, tokens):\n look_ahead = LookAhead(tokens)\n segments = Segment()\n while not look_ahead.empty():\n if look_ahead.peek().type not in self.whitespace:\n paragraph = MatchableSegment(look_ahead.i)\n while not look_ahead.empty() and look_ahead.peek(\n ).type not in self.paragraph_end:\n if look_ahead.peek().type == 'tab_open':\n tab_depth = 1\n sentence = MatchableSegment(look_ahead.i, [next(\n look_ahead)])\n while not look_ahead.empty() and tab_depth > 0:\n tab_depth += look_ahead.peek().type == 'tab_open'\n tab_depth -= look_ahead.peek().type == 'tab_close'\n sentence.append(next(look_ahead))\n paragraph.append(sentence)\n elif look_ahead.peek().type not in self.whitespace:\n sentence = MatchableSegment(look_ahead.i, [next(\n look_ahead)])\n sub_depth = int(sentence[0].type in SUB_OPEN)\n while not look_ahead.empty():\n sub_depth += look_ahead.peek().type in SUB_OPEN\n sub_depth -= look_ahead.peek().type in SUB_CLOSE\n sentence.append(next(look_ahead))\n if sentence[-1\n ].type in self.sentence_end and sub_depth <= 0:\n non_whitespace = sum(s.type not in self.\n whitespace for s in sentence)\n if non_whitespace >= self.min_sentence:\n break\n paragraph.append(sentence)\n else:\n whitespace = Segment(look_ahead.i, [next(look_ahead)])\n paragraph.append(whitespace)\n segments.append(paragraph)\n else:\n whitespace = Segment(look_ahead.i, [next(look_ahead)])\n segments.append(whitespace)\n return segments\n", "role": "user" }, { "content": "segment", "role": "assistant" } ]
def x(self, payload, token, expiration=None, priority=None, identifier=None): created_conn = False while not created_conn: if len(self.conns) == 0: self.conns.append(PushConnection(self, self.address, self. certfile, self.keyfile)) created_conn = True conn = random.choice(self.conns) try: conn.send(payload, token, expiration=expiration, priority= priority, identifier=identifier) return except: logger.info('Connection died: removing') self.conns.remove(conn) raise SendFailedException()
send
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, payload, token, expiration=None, priority=None, identifier=None):\n created_conn = False\n while not created_conn:\n if len(self.conns) == 0:\n self.conns.append(PushConnection(self, self.address, self.\n certfile, self.keyfile))\n created_conn = True\n conn = random.choice(self.conns)\n try:\n conn.send(payload, token, expiration=expiration, priority=\n priority, identifier=identifier)\n return\n except:\n logger.info('Connection died: removing')\n self.conns.remove(conn)\n raise SendFailedException()\n", "role": "user" }, { "content": "send", "role": "assistant" } ]
def x(self): if not self.fbaddress: raise Exception( 'Attempted to fetch feedback but no feedback_address supplied') fbconn = FeedbackConnection(self, self.fbaddress, self.certfile, self. keyfile) return fbconn.get_all()
get_all_feedback
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if not self.fbaddress:\n raise Exception(\n 'Attempted to fetch feedback but no feedback_address supplied')\n fbconn = FeedbackConnection(self, self.fbaddress, self.certfile, self.\n keyfile)\n return fbconn.get_all()\n", "role": "user" }, { "content": "get_all_feedback", "role": "assistant" } ]
def x(self, response): if response.startswith('\n'): response = response[1:] tree = etree.fromstring(response) return tree
parse_xml
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, response):\n if response.startswith('\\n'):\n response = response[1:]\n tree = etree.fromstring(response)\n return tree\n", "role": "user" }, { "content": "parse_xml", "role": "assistant" } ]
def x(self, error): status_code = error.response.status_code message = error.message if 403 == status_code: raise NewRelicInvalidApiKeyException(message) elif 404 == status_code: raise NewRelicUnknownApplicationException(message) elif 422 == status_code: raise NewRelicInvalidParameterException(message) else: raise NewRelicApiException(message)
handle_api_error
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, error):\n status_code = error.response.status_code\n message = error.message\n if 403 == status_code:\n raise NewRelicInvalidApiKeyException(message)\n elif 404 == status_code:\n raise NewRelicUnknownApplicationException(message)\n elif 422 == status_code:\n raise NewRelicInvalidParameterException(message)\n else:\n raise NewRelicApiException(message)\n", "role": "user" }, { "content": "handle_api_error", "role": "assistant" } ]
def x(self, api_call, window=60): current = datetime.datetime.now() try: previous = getattr(self, api_call.__name__ + '_window') previous.__str__ except AttributeError: now = datetime.datetime.now() outside_window = datetime.timedelta(seconds=window + 1) previous = now - outside_window if current - previous > datetime.timedelta(seconds=window): setattr(self, api_call.__name__ + '_window', current) else: timeout = window - (current - previous).seconds raise NewRelicApiRateLimitException(str(timeout))
api_rate_limit_exceeded
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, api_call, window=60):\n current = datetime.datetime.now()\n try:\n previous = getattr(self, api_call.__name__ + '_window')\n previous.__str__\n except AttributeError:\n now = datetime.datetime.now()\n outside_window = datetime.timedelta(seconds=window + 1)\n previous = now - outside_window\n if current - previous > datetime.timedelta(seconds=window):\n setattr(self, api_call.__name__ + '_window', current)\n else:\n timeout = window - (current - previous).seconds\n raise NewRelicApiRateLimitException(str(timeout))\n", "role": "user" }, { "content": "api_rate_limit_exceeded", "role": "assistant" } ]
def x(self): endpoint = 'https://rpm.newrelic.com' uri = '{endpoint}/accounts/{id}/applications.xml'.format(endpoint= endpoint, id=self.account_id) response = self._make_get_request(uri) applications = [] for application in response.findall('.//application'): application_properties = {} for field in application: application_properties[field.tag] = field.text applications.append(Application(application_properties)) return applications
view_applications
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n endpoint = 'https://rpm.newrelic.com'\n uri = '{endpoint}/accounts/{id}/applications.xml'.format(endpoint=\n endpoint, id=self.account_id)\n response = self._make_get_request(uri)\n applications = []\n for application in response.findall('.//application'):\n application_properties = {}\n for field in application:\n application_properties[field.tag] = field.text\n applications.append(Application(application_properties))\n return applications\n", "role": "user" }, { "content": "view_applications", "role": "assistant" } ]
def x(self, applications): endpoint = 'https://api.newrelic.com' uri = ('{endpoint}/api/v1/accounts/{account_id}/applications/delete.xml' .format(endpoint=endpoint, account_id=self.account_id)) payload = applications response = self._make_post_request(uri, payload) failed_deletions = {} for application in response.findall('.//application'): if not 'deleted' in application.findall('.//result')[0].text: failed_deletions['app_id'] = application.get('id') return failed_deletions
delete_applications
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, applications):\n endpoint = 'https://api.newrelic.com'\n uri = ('{endpoint}/api/v1/accounts/{account_id}/applications/delete.xml'\n .format(endpoint=endpoint, account_id=self.account_id))\n payload = applications\n response = self._make_post_request(uri, payload)\n failed_deletions = {}\n for application in response.findall('.//application'):\n if not 'deleted' in application.findall('.//result')[0].text:\n failed_deletions['app_id'] = application.get('id')\n return failed_deletions\n", "role": "user" }, { "content": "delete_applications", "role": "assistant" } ]
def x(self, application_id=None, application_name=None, description=None, revision=None, changelog=None, user=None): endpoint = 'https://rpm.newrelic.com' uri = '{endpoint}/deployments.xml'.format(endpoint=endpoint) deploy_event = {} if not application_id is None: deploy_event['deployment[application_id]'] = application_id elif not application_name is None: deploy_event['deployment[app_name]'] = application_name else: raise NewRelicInvalidParameterException( 'Must specify either application_id or application_name.') if not description is None: deploy_event['deployment[description]'] = description if not revision is None: deploy_event['deployment[revision]'] = revision if not changelog is None: deploy_event['deployment[changelog]'] = changelog if not user is None: deploy_event['deployment[user]'] = user response = self._make_post_request(uri, deploy_event) result = {} for value in response: result[value.tag] = value.text return result
notify_deployment
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, application_id=None, application_name=None, description=None,\n revision=None, changelog=None, user=None):\n endpoint = 'https://rpm.newrelic.com'\n uri = '{endpoint}/deployments.xml'.format(endpoint=endpoint)\n deploy_event = {}\n if not application_id is None:\n deploy_event['deployment[application_id]'] = application_id\n elif not application_name is None:\n deploy_event['deployment[app_name]'] = application_name\n else:\n raise NewRelicInvalidParameterException(\n 'Must specify either application_id or application_name.')\n if not description is None:\n deploy_event['deployment[description]'] = description\n if not revision is None:\n deploy_event['deployment[revision]'] = revision\n if not changelog is None:\n deploy_event['deployment[changelog]'] = changelog\n if not user is None:\n deploy_event['deployment[user]'] = user\n response = self._make_post_request(uri, deploy_event)\n result = {}\n for value in response:\n result[value.tag] = value.text\n return result\n", "role": "user" }, { "content": "notify_deployment", "role": "assistant" } ]
def x(self, agent_id, re=None, limit=5000): self._api_rate_limit_exceeded(self.get_metric_names) parameters = {'re': re, 'limit': limit} endpoint = 'https://api.newrelic.com' uri = '{endpoint}/api/v1/applications/{agent_id}/metrics.xml'.format( endpoint=endpoint, agent_id=agent_id) response = self._make_get_request(uri, parameters=parameters, timeout= max(self.timeout, 5.0)) metrics = {} for metric in response.findall('.//metric'): fields = [] for field in metric.findall('.//field'): fields.append(field.get('name')) metrics[metric.get('name')] = fields return metrics
get_metric_names
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, agent_id, re=None, limit=5000):\n self._api_rate_limit_exceeded(self.get_metric_names)\n parameters = {'re': re, 'limit': limit}\n endpoint = 'https://api.newrelic.com'\n uri = '{endpoint}/api/v1/applications/{agent_id}/metrics.xml'.format(\n endpoint=endpoint, agent_id=agent_id)\n response = self._make_get_request(uri, parameters=parameters, timeout=\n max(self.timeout, 5.0))\n metrics = {}\n for metric in response.findall('.//metric'):\n fields = []\n for field in metric.findall('.//field'):\n fields.append(field.get('name'))\n metrics[metric.get('name')] = fields\n return metrics\n", "role": "user" }, { "content": "get_metric_names", "role": "assistant" } ]
def x(self, applications, metrics, field, begin, end, summary=False): self._api_rate_limit_exceeded(self.get_metric_data) parameters = {} try: int(applications[0]) except ValueError: app_string = 'app' else: app_string = 'app_id' if len(applications) > 1: app_string = app_string + '[]' parameters[app_string] = applications parameters['metrics[]'] = metrics parameters['field'] = field parameters['begin'] = begin parameters['end'] = end parameters['summary'] = int(summary) endpoint = 'https://api.newrelic.com' uri = '{endpoint}/api/v1/accounts/{account_id}/metrics/data.xml'.format( endpoint=endpoint, account_id=self.account_id) response = self._make_get_request(uri, parameters=parameters, timeout= max(self.timeout, 5.0)) metrics = [] for metric in response.findall('.//metric'): metrics.append(Metric(metric)) return metrics
get_metric_data
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, applications, metrics, field, begin, end, summary=False):\n self._api_rate_limit_exceeded(self.get_metric_data)\n parameters = {}\n try:\n int(applications[0])\n except ValueError:\n app_string = 'app'\n else:\n app_string = 'app_id'\n if len(applications) > 1:\n app_string = app_string + '[]'\n parameters[app_string] = applications\n parameters['metrics[]'] = metrics\n parameters['field'] = field\n parameters['begin'] = begin\n parameters['end'] = end\n parameters['summary'] = int(summary)\n endpoint = 'https://api.newrelic.com'\n uri = '{endpoint}/api/v1/accounts/{account_id}/metrics/data.xml'.format(\n endpoint=endpoint, account_id=self.account_id)\n response = self._make_get_request(uri, parameters=parameters, timeout=\n max(self.timeout, 5.0))\n metrics = []\n for metric in response.findall('.//metric'):\n metrics.append(Metric(metric))\n return metrics\n", "role": "user" }, { "content": "get_metric_data", "role": "assistant" } ]
def x(self, application_id): endpoint = 'https://rpm.newrelic.com' remote_file = 'threshold_values.xml' uri = ('{endpoint}/accounts/{account_id}/applications/{app_id}/{xml}'. format(endpoint=endpoint, account_id=self.account_id, app_id= application_id, xml=remote_file)) response = self._make_get_request(uri) thresholds = [] for threshold_value in response.findall('.//threshold_value'): properties = {} for tag, text in threshold_value.items(): properties[tag] = text thresholds.append(Threshold(properties)) return thresholds
get_threshold_values
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, application_id):\n endpoint = 'https://rpm.newrelic.com'\n remote_file = 'threshold_values.xml'\n uri = ('{endpoint}/accounts/{account_id}/applications/{app_id}/{xml}'.\n format(endpoint=endpoint, account_id=self.account_id, app_id=\n application_id, xml=remote_file))\n response = self._make_get_request(uri)\n thresholds = []\n for threshold_value in response.findall('.//threshold_value'):\n properties = {}\n for tag, text in threshold_value.items():\n properties[tag] = text\n thresholds.append(Threshold(properties))\n return thresholds\n", "role": "user" }, { "content": "get_threshold_values", "role": "assistant" } ]
def x(self): endpoint = 'https://api.newrelic.com' uri = '{endpoint}/api/v1/accounts/{id}/servers.xml'.format(endpoint= endpoint, id=self.account_id) response = self._make_get_request(uri) servers = [] for server in response.findall('.//server'): server_properties = {} for field in server: server_properties[field.tag] = field.text servers.append(Server(server_properties)) return servers
view_servers
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n endpoint = 'https://api.newrelic.com'\n uri = '{endpoint}/api/v1/accounts/{id}/servers.xml'.format(endpoint=\n endpoint, id=self.account_id)\n response = self._make_get_request(uri)\n servers = []\n for server in response.findall('.//server'):\n server_properties = {}\n for field in server:\n server_properties[field.tag] = field.text\n servers.append(Server(server_properties))\n return servers\n", "role": "user" }, { "content": "view_servers", "role": "assistant" } ]
def x(self, server_id): endpoint = 'https://api.newrelic.com' uri = ('{endpoint}/api/v1/accounts/{account_id}/servers/{server_id}.xml' .format(endpoint=endpoint, account_id=self.account_id, server_id= server_id)) response = self._make_delete_request(uri) failed_deletions = [] for server in response.findall('.//server'): if not 'deleted' in server.findall('.//result')[0].text: failed_deletions.append({'server_id': server.get('id')}) return failed_deletions
delete_servers
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, server_id):\n endpoint = 'https://api.newrelic.com'\n uri = ('{endpoint}/api/v1/accounts/{account_id}/servers/{server_id}.xml'\n .format(endpoint=endpoint, account_id=self.account_id, server_id=\n server_id))\n response = self._make_delete_request(uri)\n failed_deletions = []\n for server in response.findall('.//server'):\n if not 'deleted' in server.findall('.//result')[0].text:\n failed_deletions.append({'server_id': server.get('id')})\n return failed_deletions\n", "role": "user" }, { "content": "delete_servers", "role": "assistant" } ]
def x(reads1_infile_path, reads2_infile_path, perc_max_diff=None, min_overlap=None, outfile_label='fastqjoin', params={}, working_dir= tempfile.gettempdir(), SuppressStderr=True, SuppressStdout=True, HALT_EXEC=False): abs_r1_path = os.path.abspath(reads1_infile_path) abs_r2_path = os.path.abspath(reads2_infile_path) infile_paths = [abs_r1_path, abs_r2_path] for p in infile_paths: if not os.path.exists(p): raise IOError('File not found at: %s' % p) fastq_join_app = FastqJoin(params=params, WorkingDir=working_dir, SuppressStderr=SuppressStderr, SuppressStdout=SuppressStdout, HALT_EXEC=HALT_EXEC) if perc_max_diff is not None: if isinstance(perc_max_diff, int) and 0 <= perc_max_diff <= 100: fastq_join_app.Parameters['-p'].on(perc_max_diff) else: raise ValueError('perc_max_diff must be int between 0-100!') if min_overlap is not None: if isinstance(min_overlap, int) and 0 < min_overlap: fastq_join_app.Parameters['-m'].on(min_overlap) else: raise ValueError('min_overlap must be an int >= 0!') if outfile_label is not None: if isinstance(outfile_label, str): fastq_join_app.Parameters['-o'].on(outfile_label + '.') else: raise ValueError('outfile_label must be a string!') else: pass result = fastq_join_app(infile_paths) path_dict = {} path_dict['Assembled'] = result['Assembled'].name path_dict['UnassembledReads1'] = result['UnassembledReads1'].name path_dict['UnassembledReads2'] = result['UnassembledReads2'].name for path in path_dict.values(): if not os.path.exists(path): raise IOError('Output file not found at: %s' % path) for key, file_path in path_dict.items(): new_file_path = file_path + '.fastq' shutil.move(file_path, new_file_path) path_dict[key] = new_file_path return path_dict
join_paired_end_reads_fastqjoin
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(reads1_infile_path, reads2_infile_path, perc_max_diff=None,\n min_overlap=None, outfile_label='fastqjoin', params={}, working_dir=\n tempfile.gettempdir(), SuppressStderr=True, SuppressStdout=True,\n HALT_EXEC=False):\n abs_r1_path = os.path.abspath(reads1_infile_path)\n abs_r2_path = os.path.abspath(reads2_infile_path)\n infile_paths = [abs_r1_path, abs_r2_path]\n for p in infile_paths:\n if not os.path.exists(p):\n raise IOError('File not found at: %s' % p)\n fastq_join_app = FastqJoin(params=params, WorkingDir=working_dir,\n SuppressStderr=SuppressStderr, SuppressStdout=SuppressStdout,\n HALT_EXEC=HALT_EXEC)\n if perc_max_diff is not None:\n if isinstance(perc_max_diff, int) and 0 <= perc_max_diff <= 100:\n fastq_join_app.Parameters['-p'].on(perc_max_diff)\n else:\n raise ValueError('perc_max_diff must be int between 0-100!')\n if min_overlap is not None:\n if isinstance(min_overlap, int) and 0 < min_overlap:\n fastq_join_app.Parameters['-m'].on(min_overlap)\n else:\n raise ValueError('min_overlap must be an int >= 0!')\n if outfile_label is not None:\n if isinstance(outfile_label, str):\n fastq_join_app.Parameters['-o'].on(outfile_label + '.')\n else:\n raise ValueError('outfile_label must be a string!')\n else:\n pass\n result = fastq_join_app(infile_paths)\n path_dict = {}\n path_dict['Assembled'] = result['Assembled'].name\n path_dict['UnassembledReads1'] = result['UnassembledReads1'].name\n path_dict['UnassembledReads2'] = result['UnassembledReads2'].name\n for path in path_dict.values():\n if not os.path.exists(path):\n raise IOError('Output file not found at: %s' % path)\n for key, file_path in path_dict.items():\n new_file_path = file_path + '.fastq'\n shutil.move(file_path, new_file_path)\n path_dict[key] = new_file_path\n return path_dict\n", "role": "user" }, { "content": "join_paired_end_reads_fastqjoin", "role": "assistant" } ]
def x(self): if self.Parameters['-o'].isOn(): output_path = self._absolute(str(self.Parameters['-o'].Value)) else: raise ValueError('No output path specified.') return output_path
get_output_path
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.Parameters['-o'].isOn():\n output_path = self._absolute(str(self.Parameters['-o'].Value))\n else:\n raise ValueError('No output path specified.')\n return output_path\n", "role": "user" }, { "content": "get_output_path", "role": "assistant" } ]
def x(self): if self.Parameters['-r'].isOn(): stitch_path = self._absolute(str(self.Parameters['-r'].Value)) return stitch_path elif self.Parameters['-r'].isOff(): return None
get_stitch_report_path
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.Parameters['-r'].isOn():\n stitch_path = self._absolute(str(self.Parameters['-r'].Value))\n return stitch_path\n elif self.Parameters['-r'].isOff():\n return None\n", "role": "user" }, { "content": "get_stitch_report_path", "role": "assistant" } ]
def x(self, data): output_path = self._get_output_path() result = {} result['Assembled'] = ResultPath(Path=output_path + 'join', IsWritten=True) result['UnassembledReads1'] = ResultPath(Path=output_path + 'un1', IsWritten=True) result['UnassembledReads2'] = ResultPath(Path=output_path + 'un2', IsWritten=True) stitch_path = self._get_stitch_report_path() if stitch_path: result['Report'] = ResultPath(Path=stitch_path, IsWritten=True) mate_path_string = output_path + 'join2' mate_unassembled_path_string = output_path + 'un3' if os.path.exists(mate_path_string) and os.path.exists( mate_unassembled_path_string): result['Mate'] = ResultPath(Path=mate_path_string, IsWritten=True) result['MateUnassembled'] = ResultPath(Path= mate_unassembled_path_string, IsWritten=True) else: pass return result
get_result_paths
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, data):\n output_path = self._get_output_path()\n result = {}\n result['Assembled'] = ResultPath(Path=output_path + 'join', IsWritten=True)\n result['UnassembledReads1'] = ResultPath(Path=output_path + 'un1',\n IsWritten=True)\n result['UnassembledReads2'] = ResultPath(Path=output_path + 'un2',\n IsWritten=True)\n stitch_path = self._get_stitch_report_path()\n if stitch_path:\n result['Report'] = ResultPath(Path=stitch_path, IsWritten=True)\n mate_path_string = output_path + 'join2'\n mate_unassembled_path_string = output_path + 'un3'\n if os.path.exists(mate_path_string) and os.path.exists(\n mate_unassembled_path_string):\n result['Mate'] = ResultPath(Path=mate_path_string, IsWritten=True)\n result['MateUnassembled'] = ResultPath(Path=\n mate_unassembled_path_string, IsWritten=True)\n else:\n pass\n return result\n", "role": "user" }, { "content": "get_result_paths", "role": "assistant" } ]
def x(fields, as_dict=False): _fields = {} if fields in ['~', None, False, True, {}, []]: _fields = {} elif isinstance(fields, dict): _fields.update({unicode(k).strip(): int(v) for k, v in fields. iteritems()}) elif isinstance(fields, basestring): _fields.update({unicode(s).strip(): (1) for s in fields.split(',')}) elif isinstance(fields, (list, tuple)): _fields.update({unicode(s).strip(): (1) for s in fields}) else: raise ValueError('invalid fields value') if as_dict: return _fields else: return sorted(_fields.keys())
parse_fields
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(fields, as_dict=False):\n _fields = {}\n if fields in ['~', None, False, True, {}, []]:\n _fields = {}\n elif isinstance(fields, dict):\n _fields.update({unicode(k).strip(): int(v) for k, v in fields.\n iteritems()})\n elif isinstance(fields, basestring):\n _fields.update({unicode(s).strip(): (1) for s in fields.split(',')})\n elif isinstance(fields, (list, tuple)):\n _fields.update({unicode(s).strip(): (1) for s in fields})\n else:\n raise ValueError('invalid fields value')\n if as_dict:\n return _fields\n else:\n return sorted(_fields.keys())\n", "role": "user" }, { "content": "parse_fields", "role": "assistant" } ]
def x(date, func='date'): if isinstance(date, basestring): date = date.strip() if not date: return '_end == None' if date == '~': return '' before = lambda d: '_start < %s("%s")' % (func, ts2dt(d) if d else None) after = lambda d: '(_end >= %s("%s") or _end == None)' % (func, ts2dt(d ) if d else None) split = date.split('~') split = [re.sub('\\+\\d\\d:\\d\\d', '', d.replace('T', ' ')) for d in split ] if len(split) == 1: return '%s and %s' % (before(split[0]), after(split[0])) elif split[0] in ['', None]: return before(split[1]) elif split[1] in ['', None]: return after(split[0]) else: return '%s and %s' % (before(split[1]), after(split[0]))
date_range
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(date, func='date'):\n if isinstance(date, basestring):\n date = date.strip()\n if not date:\n return '_end == None'\n if date == '~':\n return ''\n before = lambda d: '_start < %s(\"%s\")' % (func, ts2dt(d) if d else None)\n after = lambda d: '(_end >= %s(\"%s\") or _end == None)' % (func, ts2dt(d\n ) if d else None)\n split = date.split('~')\n split = [re.sub('\\\\+\\\\d\\\\d:\\\\d\\\\d', '', d.replace('T', ' ')) for d in split\n ]\n if len(split) == 1:\n return '%s and %s' % (before(split[0]), after(split[0]))\n elif split[0] in ['', None]:\n return before(split[1])\n elif split[1] in ['', None]:\n return after(split[0])\n else:\n return '%s and %s' % (before(split[1]), after(split[0]))\n", "role": "user" }, { "content": "date_range", "role": "assistant" } ]
def x(table, query=None, date=None, fields=None, distinct=False, limit=None, alias=None): date = date_range(date) limit = int(limit or -1) if query and date: query = '%s and %s' % (query, date) elif date: query = date elif query: pass else: query = None fields = parse_fields(fields=fields) or None fields = fields if fields else table.columns msg = 'parse(query=%s, fields=%s)' % (query, fields) logger.debug(msg) kwargs = {} if query: interpreter = MQLInterpreter(table) query = interpreter.parse(query) kwargs['whereclause'] = query if distinct: kwargs['distinct'] = distinct query = select(fields, from_obj=table, **kwargs) if limit >= 1: query = query.limit(limit) if alias: query = query.alias(alias) return query
parse
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(table, query=None, date=None, fields=None, distinct=False, limit=None,\n alias=None):\n date = date_range(date)\n limit = int(limit or -1)\n if query and date:\n query = '%s and %s' % (query, date)\n elif date:\n query = date\n elif query:\n pass\n else:\n query = None\n fields = parse_fields(fields=fields) or None\n fields = fields if fields else table.columns\n msg = 'parse(query=%s, fields=%s)' % (query, fields)\n logger.debug(msg)\n kwargs = {}\n if query:\n interpreter = MQLInterpreter(table)\n query = interpreter.parse(query)\n kwargs['whereclause'] = query\n if distinct:\n kwargs['distinct'] = distinct\n query = select(fields, from_obj=table, **kwargs)\n if limit >= 1:\n query = query.limit(limit)\n if alias:\n query = query.alias(alias)\n return query\n", "role": "user" }, { "content": "parse", "role": "assistant" } ]
def x(self, func, *func_args): worker = self.pick_sticky(0) args = (func,) + func_args coro = worker.enqueue(enums.Task.FUNC, args) asyncio.ensure_future(coro)
enqueue_sync
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, func, *func_args):\n worker = self.pick_sticky(0)\n args = (func,) + func_args\n coro = worker.enqueue(enums.Task.FUNC, args)\n asyncio.ensure_future(coro)\n", "role": "user" }, { "content": "enqueue_sync", "role": "assistant" } ]
async def async_enqueue_sync(self, func, *func_args): """ Enqueue an arbitrary synchronous function. """ worker = self.pick_sticky(0) args = (func,) + func_args await worker.enqueue(enums.Task.FUNC, args)
async_enqueue_sync
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def async_enqueue_sync(self, func, *func_args):\n \"\"\"\n Enqueue an arbitrary synchronous function.\n \"\"\"\n worker = self.pick_sticky(0)\n args = (func,) + func_args\n await worker.enqueue(enums.Task.FUNC, args)\n", "role": "user" }, { "content": "async_enqueue_sync", "role": "assistant" } ]
def x(self, resource): worker = self.pick_sticky(resource.url_string) coro = worker.enqueue(enums.Task.DOWNLOAD, (resource,)) asyncio.ensure_future(coro)
enqueue_download
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, resource):\n worker = self.pick_sticky(resource.url_string)\n coro = worker.enqueue(enums.Task.DOWNLOAD, (resource,))\n asyncio.ensure_future(coro)\n", "role": "user" }, { "content": "enqueue_download", "role": "assistant" } ]
async def async_enqueue_download(self, resource): """ Enqueue the download of the given foreign resource. """ worker = self.pick_sticky(resource.url_string) await worker.enqueue(enums.Task.DOWNLOAD, (resource,))
async_enqueue_download
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def async_enqueue_download(self, resource):\n \"\"\"\n Enqueue the download of the given foreign resource.\n \"\"\"\n worker = self.pick_sticky(resource.url_string)\n await worker.enqueue(enums.Task.DOWNLOAD, (resource,))\n", "role": "user" }, { "content": "async_enqueue_download", "role": "assistant" } ]
def x(self, converter, from_resource, to_resource): worker = self.pick_sticky(from_resource.url_string) args = converter, from_resource, to_resource coro = worker.enqueue(enums.Task.CONVERT, args) asyncio.ensure_future(coro)
enqueue_convert
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, converter, from_resource, to_resource):\n worker = self.pick_sticky(from_resource.url_string)\n args = converter, from_resource, to_resource\n coro = worker.enqueue(enums.Task.CONVERT, args)\n asyncio.ensure_future(coro)\n", "role": "user" }, { "content": "enqueue_convert", "role": "assistant" } ]
async def async_enqueue_convert(self, converter, from_, to): """ Enqueue use of the given converter to convert to given from and to resources. """ worker = self.pick_sticky(from_.url_string) args = converter, from_, to await worker.enqueue(enums.Task.CONVERT, args)
async_enqueue_convert
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def async_enqueue_convert(self, converter, from_, to):\n \"\"\"\n Enqueue use of the given converter to convert to given\n from and to resources.\n \"\"\"\n worker = self.pick_sticky(from_.url_string)\n args = converter, from_, to\n await worker.enqueue(enums.Task.CONVERT, args)\n", "role": "user" }, { "content": "async_enqueue_convert", "role": "assistant" } ]
async def async_enqueue_multiconvert(self, url_string, to_type): """ Enqueue a multi-step conversion process, from the given URL string (which is assumed to have been downloaded / resolved) """ worker = self.pick_sticky(url_string) args = url_string, to_type await worker.enqueue(enums.Task.MULTICONVERT, args)
async_enqueue_multiconvert
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def async_enqueue_multiconvert(self, url_string, to_type):\n \"\"\"\n Enqueue a multi-step conversion process, from the given URL string\n (which is assumed to have been downloaded / resolved)\n \"\"\"\n worker = self.pick_sticky(url_string)\n args = url_string, to_type\n await worker.enqueue(enums.Task.MULTICONVERT, args)\n", "role": "user" }, { "content": "async_enqueue_multiconvert", "role": "assistant" } ]
def x(self, **kwargs): if self._strain_out_of_plane is not None: return self._strain_out_of_plane / -2.0 * (self.unstrained.c11(** kwargs) / self.unstrained.c12(**kwargs)) else: return 1 - self.unstrained.a(**kwargs) / self.substrate.a(**kwargs)
strain_in_plane
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, **kwargs):\n if self._strain_out_of_plane is not None:\n return self._strain_out_of_plane / -2.0 * (self.unstrained.c11(**\n kwargs) / self.unstrained.c12(**kwargs))\n else:\n return 1 - self.unstrained.a(**kwargs) / self.substrate.a(**kwargs)\n", "role": "user" }, { "content": "strain_in_plane", "role": "assistant" } ]
def x(self, **kwargs): if self._strain_out_of_plane is not None: return self._strain_out_of_plane else: return -2 * self.unstrained.c12(**kwargs) / self.unstrained.c11(** kwargs) * self.strain_in_plane(**kwargs)
strain_out_of_plane
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, **kwargs):\n if self._strain_out_of_plane is not None:\n return self._strain_out_of_plane\n else:\n return -2 * self.unstrained.c12(**kwargs) / self.unstrained.c11(**\n kwargs) * self.strain_in_plane(**kwargs)\n", "role": "user" }, { "content": "strain_out_of_plane", "role": "assistant" } ]
def x(self, **kwargs): if self.substrate is not None: return self.substrate.a(**kwargs) else: return self.unstrained.a(**kwargs) / (1.0 - self.strain_in_plane(** kwargs))
substrate_a
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, **kwargs):\n if self.substrate is not None:\n return self.substrate.a(**kwargs)\n else:\n return self.unstrained.a(**kwargs) / (1.0 - self.strain_in_plane(**\n kwargs))\n", "role": "user" }, { "content": "substrate_a", "role": "assistant" } ]
def x(self, **kwargs): return self.unstrained.CBO(**kwargs) + self.CBO_strain_shift(**kwargs)
CBO
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, **kwargs):\n return self.unstrained.CBO(**kwargs) + self.CBO_strain_shift(**kwargs)\n", "role": "user" }, { "content": "CBO", "role": "assistant" } ]
def x(self, **kwargs): return self.unstrained.CBO_Gamma(**kwargs) + self.CBO_strain_shift(**kwargs )
CBO_Gamma
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, **kwargs):\n return self.unstrained.CBO_Gamma(**kwargs) + self.CBO_strain_shift(**kwargs\n )\n", "role": "user" }, { "content": "CBO_Gamma", "role": "assistant" } ]
def x(self, **kwargs): return self.unstrained.CBO_L(**kwargs) + self.CBO_strain_shift(**kwargs)
CBO_L
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, **kwargs):\n return self.unstrained.CBO_L(**kwargs) + self.CBO_strain_shift(**kwargs)\n", "role": "user" }, { "content": "CBO_L", "role": "assistant" } ]
def x(self, **kwargs): return self.unstrained.CBO_X(**kwargs) + self.CBO_strain_shift(**kwargs)
CBO_X
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, **kwargs):\n return self.unstrained.CBO_X(**kwargs) + self.CBO_strain_shift(**kwargs)\n", "role": "user" }, { "content": "CBO_X", "role": "assistant" } ]
def x(self, **kwargs): return self.unstrained.Eg(**kwargs) + self.Eg_strain_shift(**kwargs)
Eg
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, **kwargs):\n return self.unstrained.Eg(**kwargs) + self.Eg_strain_shift(**kwargs)\n", "role": "user" }, { "content": "Eg", "role": "assistant" } ]
def x(self): for subsegment_or_token in self: if isinstance(subsegment_or_token, Segment): subsegment = subsegment_or_token for token in subsegment.tokens(): yield token else: token = subsegment_or_token yield token
tokens
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n for subsegment_or_token in self:\n if isinstance(subsegment_or_token, Segment):\n subsegment = subsegment_or_token\n for token in subsegment.tokens():\n yield token\n else:\n token = subsegment_or_token\n yield token\n", "role": "user" }, { "content": "tokens", "role": "assistant" } ]
def x(self, status_code, **kwargs): return_error = {'code': self.get_status()} exc_info = kwargs.get('exc_info') if exc_info and isinstance(exc_info[1], oz.json_api.ApiError): return_error['error'] = exc_info[1].message else: return_error['error'] = API_ERROR_CODE_MAP.get(self.get_status(), 'Unknown error') if oz.settings.get('debug'): return_error['trace'] = ''.join(traceback.format_exception(*exc_info)) self.finish(return_error) return oz.break_trigger
api_on_write_error
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, status_code, **kwargs):\n return_error = {'code': self.get_status()}\n exc_info = kwargs.get('exc_info')\n if exc_info and isinstance(exc_info[1], oz.json_api.ApiError):\n return_error['error'] = exc_info[1].message\n else:\n return_error['error'] = API_ERROR_CODE_MAP.get(self.get_status(),\n 'Unknown error')\n if oz.settings.get('debug'):\n return_error['trace'] = ''.join(traceback.format_exception(*exc_info))\n self.finish(return_error)\n return oz.break_trigger\n", "role": "user" }, { "content": "api_on_write_error", "role": "assistant" } ]
def x(self, obj): callback = self.get_argument('callback', None) if oz.settings['allow_jsonp' ] else None if callback and not CALLBACK_VALIDATOR.match(callback): raise oz.json_api.ApiError( 'Invalid callback identifier - only functions with ASCII characters are allowed' ) json = escape.json_encode(obj) if callback: self.set_header('Content-Type', 'application/javascript; charset=UTF-8' ) self.finish('%s(%s)' % (callback, json)) else: self.set_header('Content-Type', 'application/json; charset=UTF-8') self.finish(json)
respond
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, obj):\n callback = self.get_argument('callback', None) if oz.settings['allow_jsonp'\n ] else None\n if callback and not CALLBACK_VALIDATOR.match(callback):\n raise oz.json_api.ApiError(\n 'Invalid callback identifier - only functions with ASCII characters are allowed'\n )\n json = escape.json_encode(obj)\n if callback:\n self.set_header('Content-Type', 'application/javascript; charset=UTF-8'\n )\n self.finish('%s(%s)' % (callback, json))\n else:\n self.set_header('Content-Type', 'application/json; charset=UTF-8')\n self.finish(json)\n", "role": "user" }, { "content": "respond", "role": "assistant" } ]
def x(self): if self._decoded_body == None: raw_content_type = self.request.headers.get('content-type') or '' content_type = raw_content_type.split(';')[0].strip().lower() if content_type == 'application/json': try: self._decoded_body = escape.json_decode(self.request.body) except: raise oz.json_api.ApiError('Bad JSON body') else: raise oz.json_api.ApiError('JSON body expected') return self._decoded_body
body
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self._decoded_body == None:\n raw_content_type = self.request.headers.get('content-type') or ''\n content_type = raw_content_type.split(';')[0].strip().lower()\n if content_type == 'application/json':\n try:\n self._decoded_body = escape.json_decode(self.request.body)\n except:\n raise oz.json_api.ApiError('Bad JSON body')\n else:\n raise oz.json_api.ApiError('JSON body expected')\n return self._decoded_body\n", "role": "user" }, { "content": "body", "role": "assistant" } ]