source_code stringlengths 45 7.24k | function_name stringlengths 1 111 | messages list |
|---|---|---|
def x(self, kwargs):
criteria = {}
for key in self._search_by:
if getattr(self, key, None):
criteria[key] = getattr(self, key)
elif key in kwargs and kwargs.get(key):
criteria[key] = kwargs.get(key)
return criteria
| build_search_values | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, kwargs):\n criteria = {}\n for key in self._search_by:\n if getattr(self, key, None):\n criteria[key] = getattr(self, key)\n elif key in kwargs and kwargs.get(key):\n criteria[key] = kwargs.get(key)\n return criteria\n",
"role": "user"
},
{
"content": "build_search_values",
"role": "assistant"
}
] |
def x(self, ipv4addr):
for addr in self.ipv4addrs:
if isinstance(addr, dict) and addr['ipv4addr'
] == ipv4addr or isinstance(addr, HostIPv4
) and addr.ipv4addr == ipv4addr:
raise ValueError('Already exists')
self.ipv4addrs.append({'ipv4addr': ipv4addr})
| add_ipv4addr | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, ipv4addr):\n for addr in self.ipv4addrs:\n if isinstance(addr, dict) and addr['ipv4addr'\n ] == ipv4addr or isinstance(addr, HostIPv4\n ) and addr.ipv4addr == ipv4addr:\n raise ValueError('Already exists')\n self.ipv4addrs.append({'ipv4addr': ipv4addr})\n",
"role": "user"
},
{
"content": "add_ipv4addr",
"role": "assistant"
}
] |
def x(self, ipv4addr):
for addr in self.ipv4addrs:
if isinstance(addr, dict) and addr['ipv4addr'
] == ipv4addr or isinstance(addr, HostIPv4
) and addr.ipv4addr == ipv4addr:
self.ipv4addrs.remove(addr)
break
| remove_ipv4addr | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, ipv4addr):\n for addr in self.ipv4addrs:\n if isinstance(addr, dict) and addr['ipv4addr'\n ] == ipv4addr or isinstance(addr, HostIPv4\n ) and addr.ipv4addr == ipv4addr:\n self.ipv4addrs.remove(addr)\n break\n",
"role": "user"
},
{
"content": "remove_ipv4addr",
"role": "assistant"
}
] |
def x(self, ipv6addr):
for addr in self.ipv6addrs:
if isinstance(addr, dict) and addr['ipv6addr'
] == ipv6addr or isinstance(addr, HostIPv4
) and addr.ipv6addr == ipv6addr:
raise ValueError('Already exists')
self.ipv6addrs.append({'ipv6addr': ipv6addr})
| add_ipv6addr | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, ipv6addr):\n for addr in self.ipv6addrs:\n if isinstance(addr, dict) and addr['ipv6addr'\n ] == ipv6addr or isinstance(addr, HostIPv4\n ) and addr.ipv6addr == ipv6addr:\n raise ValueError('Already exists')\n self.ipv6addrs.append({'ipv6addr': ipv6addr})\n",
"role": "user"
},
{
"content": "add_ipv6addr",
"role": "assistant"
}
] |
def x(self, ipv6addr):
for addr in self.ipv6addrs:
if isinstance(addr, dict) and addr['ipv6addr'
] == ipv6addr or isinstance(addr, HostIPv4
) and addr.ipv6addr == ipv6addr:
self.ipv6addrs.remove(addr)
break
| remove_ipv6addr | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, ipv6addr):\n for addr in self.ipv6addrs:\n if isinstance(addr, dict) and addr['ipv6addr'\n ] == ipv6addr or isinstance(addr, HostIPv4\n ) and addr.ipv6addr == ipv6addr:\n self.ipv6addrs.remove(addr)\n break\n",
"role": "user"
},
{
"content": "remove_ipv6addr",
"role": "assistant"
}
] |
def x(self, uri, version=None, isolation_level='READ COMMITTED'):
isolation_level = isolation_level or 'READ COMMITTED'
kwargs = dict(isolation_level=isolation_level)
types = {list: pg.ARRAY, tuple: pg.ARRAY, set: pg.ARRAY, dict: JSONDict,
datetime: UTCEpoch}
self.type_map.update(types)
bs = self.config['batch_size']
self.config['batch_size'] = 5000 if bs == 999 else bs
self._lock_required = False
dsn = self.config['db_schema']
self.config['db_schema'] = dsn or 'public'
return uri, kwargs
| sqla_postgresql | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, uri, version=None, isolation_level='READ COMMITTED'):\n isolation_level = isolation_level or 'READ COMMITTED'\n kwargs = dict(isolation_level=isolation_level)\n types = {list: pg.ARRAY, tuple: pg.ARRAY, set: pg.ARRAY, dict: JSONDict,\n datetime: UTCEpoch}\n self.type_map.update(types)\n bs = self.config['batch_size']\n self.config['batch_size'] = 5000 if bs == 999 else bs\n self._lock_required = False\n dsn = self.config['db_schema']\n self.config['db_schema'] = dsn or 'public'\n return uri, kwargs\n",
"role": "user"
},
{
"content": "sqla_postgresql",
"role": "assistant"
}
] |
def x(self, objects, **kwargs):
return autoschema(objects=objects, exclude_keys=self.RESTRICTED_KEYS,
**kwargs)
| autoschema | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, objects, **kwargs):\n return autoschema(objects=objects, exclude_keys=self.RESTRICTED_KEYS,\n **kwargs)\n",
"role": "user"
},
{
"content": "autoschema",
"role": "assistant"
}
] |
def x(self, query=None, date=None, table=None):
table = table or self.config.get('table')
sql_count = select([func.count()])
query = self._parse_query(table=table, query=query, date=date, fields=
'id', alias='anon_x')
if query is not None:
query = sql_count.select_from(query)
else:
table = self.get_table(table)
query = sql_count
query = query.select_from(table)
return self.session_auto.execute(query).scalar()
| count | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, query=None, date=None, table=None):\n table = table or self.config.get('table')\n sql_count = select([func.count()])\n query = self._parse_query(table=table, query=query, date=date, fields=\n 'id', alias='anon_x')\n if query is not None:\n query = sql_count.select_from(query)\n else:\n table = self.get_table(table)\n query = sql_count\n query = query.select_from(table)\n return self.session_auto.execute(query).scalar()\n",
"role": "user"
},
{
"content": "count",
"role": "assistant"
}
] |
def x(self, field, oids, date=None, level=None, table=None):
table = self.get_table(table)
fringe = str2list(oids)
checked = set(fringe)
loop_k = 0
while len(fringe) > 0:
if level and loop_k == abs(level):
break
query = '_oid in %s' % list(fringe)
docs = self.find(table=table, query=query, fields=[field], date=
date, raw=True)
fringe = {oid for doc in docs for oid in doc[field] or [] if oid not in
checked}
checked |= fringe
loop_k += 1
return sorted(checked)
| deptree | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, field, oids, date=None, level=None, table=None):\n table = self.get_table(table)\n fringe = str2list(oids)\n checked = set(fringe)\n loop_k = 0\n while len(fringe) > 0:\n if level and loop_k == abs(level):\n break\n query = '_oid in %s' % list(fringe)\n docs = self.find(table=table, query=query, fields=[field], date=\n date, raw=True)\n fringe = {oid for doc in docs for oid in doc[field] or [] if oid not in\n checked}\n checked |= fringe\n loop_k += 1\n return sorted(checked)\n",
"role": "user"
},
{
"content": "deptree",
"role": "assistant"
}
] |
def x(self, field, table=None):
field = field if is_array(field) else [field]
table = self.get_table(table, except_=False)
if table is None:
last = None
else:
is_defined(field, 'field must be defined!')
last = self.find(table=table, fields=field, scalar=True, sort=field,
limit=1, descending=True, date='~', default_fields=False)
logger.debug('last %s.%s: %s' % (table, list2str(field), last))
return last
| get_last_field | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, field, table=None):\n field = field if is_array(field) else [field]\n table = self.get_table(table, except_=False)\n if table is None:\n last = None\n else:\n is_defined(field, 'field must be defined!')\n last = self.find(table=table, fields=field, scalar=True, sort=field,\n limit=1, descending=True, date='~', default_fields=False)\n logger.debug('last %s.%s: %s' % (table, list2str(field), last))\n return last\n",
"role": "user"
},
{
"content": "get_last_field",
"role": "assistant"
}
] |
def x(self, fields, name=None, table=None, **kwargs):
table = self.get_table(table)
name = self._index_default_name(fields, name)
fields = parse.parse_fields(fields)
fields = self.columns(table, fields, reflect=True)
session = self.session_new()
index = Index(name, *fields)
logger.info('Writing new index %s: %s' % (name, fields))
result = index.create(self.engine)
session.commit()
return result
| index | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, fields, name=None, table=None, **kwargs):\n table = self.get_table(table)\n name = self._index_default_name(fields, name)\n fields = parse.parse_fields(fields)\n fields = self.columns(table, fields, reflect=True)\n session = self.session_new()\n index = Index(name, *fields)\n logger.info('Writing new index %s: %s' % (name, fields))\n result = index.create(self.engine)\n session.commit()\n return result\n",
"role": "user"
},
{
"content": "index",
"role": "assistant"
}
] |
def x(self):
logger.info('Listing indexes')
_ix = {}
_i = self.inspector
for tbl in _i.get_table_names():
_ix.setdefault(tbl, [])
for ix in _i.get_indexes(tbl):
_ix[tbl].append(ix)
return _ix
| index_list | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n logger.info('Listing indexes')\n _ix = {}\n _i = self.inspector\n for tbl in _i.get_table_names():\n _ix.setdefault(tbl, [])\n for ix in _i.get_indexes(tbl):\n _ix[tbl].append(ix)\n return _ix\n",
"role": "user"
},
{
"content": "index_list",
"role": "assistant"
}
] |
def x(self, startswith=None):
logger.info('Listing cubes starting with "%s")' % startswith)
startswith = unicode(startswith or '')
tables = sorted(name for name in self.db_tables if name.startswith(
startswith))
return tables
| ls | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, startswith=None):\n logger.info('Listing cubes starting with \"%s\")' % startswith)\n startswith = unicode(startswith or '')\n tables = sorted(name for name in self.db_tables if name.startswith(\n startswith))\n return tables\n",
"role": "user"
},
{
"content": "ls",
"role": "assistant"
}
] |
def x(self, with_user, roles=None, table=None):
table = self.get_table(table)
is_true(table is not None, 'invalid table: %s' % table)
with_user = validate_username(with_user)
roles = roles or ['SELECT']
roles = validate_roles(roles, self.VALID_SHARE_ROLES)
roles = list2str(roles)
logger.info('Sharing cube %s with %s (%s)' % (table, with_user, roles))
sql = 'GRANT %s ON %s TO %s' % (roles, table, with_user)
return self.session_auto.execute(sql)
| share | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, with_user, roles=None, table=None):\n table = self.get_table(table)\n is_true(table is not None, 'invalid table: %s' % table)\n with_user = validate_username(with_user)\n roles = roles or ['SELECT']\n roles = validate_roles(roles, self.VALID_SHARE_ROLES)\n roles = list2str(roles)\n logger.info('Sharing cube %s with %s (%s)' % (table, with_user, roles))\n sql = 'GRANT %s ON %s TO %s' % (roles, table, with_user)\n return self.session_auto.execute(sql)\n",
"role": "user"
},
{
"content": "share",
"role": "assistant"
}
] |
def x(seqs, add_seq_names=False, out_filename=None, input_handler=None,
params={}, WorkingDir=tempfile.gettempdir(), SuppressStderr=None,
SuppressStdout=None):
if out_filename:
params['-out'] = out_filename
ih = input_handler or guess_input_handler(seqs, add_seq_names)
muscle_app = Muscle(params=params, InputHandler=ih, WorkingDir=
WorkingDir, SuppressStderr=SuppressStderr, SuppressStdout=
SuppressStdout)
return muscle_app(seqs)
| muscle_seqs | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(seqs, add_seq_names=False, out_filename=None, input_handler=None,\n params={}, WorkingDir=tempfile.gettempdir(), SuppressStderr=None,\n SuppressStdout=None):\n if out_filename:\n params['-out'] = out_filename\n ih = input_handler or guess_input_handler(seqs, add_seq_names)\n muscle_app = Muscle(params=params, InputHandler=ih, WorkingDir=\n WorkingDir, SuppressStderr=SuppressStderr, SuppressStdout=\n SuppressStdout)\n return muscle_app(seqs)\n",
"role": "user"
},
{
"content": "muscle_seqs",
"role": "assistant"
}
] |
def x(seqs, input_handler=None, tree_type='neighborjoining', params={},
add_seq_names=True, WorkingDir=tempfile.gettempdir(), SuppressStderr=
None, SuppressStdout=None, max_hours=5.0, constructor=PhyloNode,
clean_up=True):
params['-maxhours'] = max_hours
if tree_type:
params['-cluster2'] = tree_type
params['-tree2'] = get_tmp_filename(WorkingDir)
params['-out'] = get_tmp_filename(WorkingDir)
muscle_res = muscle_seqs(seqs, input_handler=input_handler, params=
params, add_seq_names=add_seq_names, WorkingDir=WorkingDir,
SuppressStderr=SuppressStderr, SuppressStdout=SuppressStdout)
tree = DndParser(muscle_res['Tree2Out'], constructor=constructor)
aln = [line for line in muscle_res['MuscleOut']]
if clean_up:
muscle_res.cleanUp()
return tree, aln
| aln_tree_seqs | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(seqs, input_handler=None, tree_type='neighborjoining', params={},\n add_seq_names=True, WorkingDir=tempfile.gettempdir(), SuppressStderr=\n None, SuppressStdout=None, max_hours=5.0, constructor=PhyloNode,\n clean_up=True):\n params['-maxhours'] = max_hours\n if tree_type:\n params['-cluster2'] = tree_type\n params['-tree2'] = get_tmp_filename(WorkingDir)\n params['-out'] = get_tmp_filename(WorkingDir)\n muscle_res = muscle_seqs(seqs, input_handler=input_handler, params=\n params, add_seq_names=add_seq_names, WorkingDir=WorkingDir,\n SuppressStderr=SuppressStderr, SuppressStdout=SuppressStdout)\n tree = DndParser(muscle_res['Tree2Out'], constructor=constructor)\n aln = [line for line in muscle_res['MuscleOut']]\n if clean_up:\n muscle_res.cleanUp()\n return tree, aln\n",
"role": "user"
},
{
"content": "aln_tree_seqs",
"role": "assistant"
}
] |
def x(seqs, moltype=DNA, params=None):
if not params:
params = {}
seq_collection = SequenceCollection(seqs, MolType=moltype)
int_map, int_keys = seq_collection.getIntMap()
int_map = SequenceCollection(int_map, MolType=moltype)
params.update({'-out': get_tmp_filename()})
app = Muscle(InputHandler='_input_as_multiline_string', params=params,
WorkingDir=tempfile.gettempdir())
res = app(int_map.toFasta())
alignment = dict(parse_fasta(res['MuscleOut']))
new_alignment = {}
for k, v in alignment.items():
new_alignment[int_keys[k]] = v
new_alignment = Alignment(new_alignment, MolType=moltype)
res.cleanUp()
del (seq_collection, int_map, int_keys, app, res, alignment, params)
return new_alignment
| align_unaligned_seqs | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(seqs, moltype=DNA, params=None):\n if not params:\n params = {}\n seq_collection = SequenceCollection(seqs, MolType=moltype)\n int_map, int_keys = seq_collection.getIntMap()\n int_map = SequenceCollection(int_map, MolType=moltype)\n params.update({'-out': get_tmp_filename()})\n app = Muscle(InputHandler='_input_as_multiline_string', params=params,\n WorkingDir=tempfile.gettempdir())\n res = app(int_map.toFasta())\n alignment = dict(parse_fasta(res['MuscleOut']))\n new_alignment = {}\n for k, v in alignment.items():\n new_alignment[int_keys[k]] = v\n new_alignment = Alignment(new_alignment, MolType=moltype)\n res.cleanUp()\n del (seq_collection, int_map, int_keys, app, res, alignment, params)\n return new_alignment\n",
"role": "user"
},
{
"content": "align_unaligned_seqs",
"role": "assistant"
}
] |
def x(seqs, moltype, best_tree=False, params=None):
aln = align_unaligned_seqs(seqs, moltype=moltype, params=params)
tree = build_tree_from_alignment(aln, moltype, best_tree, params)
return {'Align': aln, 'Tree': tree}
| align_and_build_tree | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(seqs, moltype, best_tree=False, params=None):\n aln = align_unaligned_seqs(seqs, moltype=moltype, params=params)\n tree = build_tree_from_alignment(aln, moltype, best_tree, params)\n return {'Align': aln, 'Tree': tree}\n",
"role": "user"
},
{
"content": "align_and_build_tree",
"role": "assistant"
}
] |
def x(aln, moltype=DNA, best_tree=False, params=None):
app = Muscle(InputHandler='_input_as_multiline_string', params=params,
WorkingDir=tempfile.gettempdir())
app.Parameters['-clusteronly'].on()
app.Parameters['-tree1'].on(get_tmp_filename(app.WorkingDir))
app.Parameters['-seqtype'].on(moltype.label)
seq_collection = SequenceCollection(aln, MolType=moltype)
int_map, int_keys = seq_collection.getIntMap()
int_map = SequenceCollection(int_map, MolType=moltype)
result = app(int_map.toFasta())
tree = DndParser(result['Tree1Out'].read(), constructor=PhyloNode)
for tip in tree.tips():
tip.Name = int_keys[tip.Name]
result.cleanUp()
del (seq_collection, app, result)
return tree
| build_tree_from_alignment | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(aln, moltype=DNA, best_tree=False, params=None):\n app = Muscle(InputHandler='_input_as_multiline_string', params=params,\n WorkingDir=tempfile.gettempdir())\n app.Parameters['-clusteronly'].on()\n app.Parameters['-tree1'].on(get_tmp_filename(app.WorkingDir))\n app.Parameters['-seqtype'].on(moltype.label)\n seq_collection = SequenceCollection(aln, MolType=moltype)\n int_map, int_keys = seq_collection.getIntMap()\n int_map = SequenceCollection(int_map, MolType=moltype)\n result = app(int_map.toFasta())\n tree = DndParser(result['Tree1Out'].read(), constructor=PhyloNode)\n for tip in tree.tips():\n tip.Name = int_keys[tip.Name]\n result.cleanUp()\n del (seq_collection, app, result)\n return tree\n",
"role": "user"
},
{
"content": "build_tree_from_alignment",
"role": "assistant"
}
] |
def x(seqs, aln, params=None):
if not params:
params = {}
seqs_collection = SequenceCollection(seqs)
seqs_int_map, seqs_int_keys = seqs_collection.getIntMap(prefix='seq_')
seqs_int_map = SequenceCollection(seqs_int_map)
aln_collection = SequenceCollection(aln)
aln_int_map, aln_int_keys = aln_collection.getIntMap(prefix='aln_')
aln_int_map = SequenceCollection(aln_int_map)
params.update({'-out': get_tmp_filename(), '-profile': True})
seqs_filename = get_tmp_filename()
seqs_out = open(seqs_filename, 'w')
seqs_out.write(seqs_int_map.toFasta())
seqs_out.close()
aln_filename = get_tmp_filename()
aln_out = open(aln_filename, 'w')
aln_out.write(aln_int_map.toFasta())
aln_out.close()
app = Muscle(InputHandler='_input_as_multifile', params=params,
WorkingDir=tempfile.gettempdir())
res = app((aln_filename, seqs_filename))
alignment = dict(parse_fasta(res['MuscleOut']))
new_alignment = {}
for k, v in alignment.items():
if k in seqs_int_keys:
new_alignment[seqs_int_keys[k]] = v
else:
new_alignment[aln_int_keys[k]] = v
new_alignment = Alignment(new_alignment)
res.cleanUp()
del (seqs_collection, seqs_int_map, seqs_int_keys)
del (aln_collection, aln_int_map, aln_int_keys)
del (app, res, alignment, params)
remove(seqs_filename)
remove(aln_filename)
return new_alignment
| add_seqs_to_alignment | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(seqs, aln, params=None):\n if not params:\n params = {}\n seqs_collection = SequenceCollection(seqs)\n seqs_int_map, seqs_int_keys = seqs_collection.getIntMap(prefix='seq_')\n seqs_int_map = SequenceCollection(seqs_int_map)\n aln_collection = SequenceCollection(aln)\n aln_int_map, aln_int_keys = aln_collection.getIntMap(prefix='aln_')\n aln_int_map = SequenceCollection(aln_int_map)\n params.update({'-out': get_tmp_filename(), '-profile': True})\n seqs_filename = get_tmp_filename()\n seqs_out = open(seqs_filename, 'w')\n seqs_out.write(seqs_int_map.toFasta())\n seqs_out.close()\n aln_filename = get_tmp_filename()\n aln_out = open(aln_filename, 'w')\n aln_out.write(aln_int_map.toFasta())\n aln_out.close()\n app = Muscle(InputHandler='_input_as_multifile', params=params,\n WorkingDir=tempfile.gettempdir())\n res = app((aln_filename, seqs_filename))\n alignment = dict(parse_fasta(res['MuscleOut']))\n new_alignment = {}\n for k, v in alignment.items():\n if k in seqs_int_keys:\n new_alignment[seqs_int_keys[k]] = v\n else:\n new_alignment[aln_int_keys[k]] = v\n new_alignment = Alignment(new_alignment)\n res.cleanUp()\n del (seqs_collection, seqs_int_map, seqs_int_keys)\n del (aln_collection, aln_int_map, aln_int_keys)\n del (app, res, alignment, params)\n remove(seqs_filename)\n remove(aln_filename)\n return new_alignment\n",
"role": "user"
},
{
"content": "add_seqs_to_alignment",
"role": "assistant"
}
] |
def x(aln1, aln2, params=None):
if not params:
params = {}
aln1_collection = SequenceCollection(aln1)
aln1_int_map, aln1_int_keys = aln1_collection.getIntMap(prefix='aln1_')
aln1_int_map = SequenceCollection(aln1_int_map)
aln2_collection = SequenceCollection(aln2)
aln2_int_map, aln2_int_keys = aln2_collection.getIntMap(prefix='aln2_')
aln2_int_map = SequenceCollection(aln2_int_map)
params.update({'-out': get_tmp_filename(), '-profile': True})
aln1_filename = get_tmp_filename()
aln1_out = open(aln1_filename, 'w')
aln1_out.write(aln1_int_map.toFasta())
aln1_out.close()
aln2_filename = get_tmp_filename()
aln2_out = open(aln2_filename, 'w')
aln2_out.write(aln2_int_map.toFasta())
aln2_out.close()
app = Muscle(InputHandler='_input_as_multifile', params=params,
WorkingDir=tempfile.gettempdir())
res = app((aln1_filename, aln2_filename))
alignment = dict(parse_fasta(res['MuscleOut']))
new_alignment = {}
for k, v in alignment.items():
if k in aln1_int_keys:
new_alignment[aln1_int_keys[k]] = v
else:
new_alignment[aln2_int_keys[k]] = v
new_alignment = Alignment(new_alignment)
res.cleanUp()
del (aln1_collection, aln1_int_map, aln1_int_keys)
del (aln2_collection, aln2_int_map, aln2_int_keys)
del (app, res, alignment, params)
remove(aln1_filename)
remove(aln2_filename)
return new_alignment
| align_two_alignments | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(aln1, aln2, params=None):\n if not params:\n params = {}\n aln1_collection = SequenceCollection(aln1)\n aln1_int_map, aln1_int_keys = aln1_collection.getIntMap(prefix='aln1_')\n aln1_int_map = SequenceCollection(aln1_int_map)\n aln2_collection = SequenceCollection(aln2)\n aln2_int_map, aln2_int_keys = aln2_collection.getIntMap(prefix='aln2_')\n aln2_int_map = SequenceCollection(aln2_int_map)\n params.update({'-out': get_tmp_filename(), '-profile': True})\n aln1_filename = get_tmp_filename()\n aln1_out = open(aln1_filename, 'w')\n aln1_out.write(aln1_int_map.toFasta())\n aln1_out.close()\n aln2_filename = get_tmp_filename()\n aln2_out = open(aln2_filename, 'w')\n aln2_out.write(aln2_int_map.toFasta())\n aln2_out.close()\n app = Muscle(InputHandler='_input_as_multifile', params=params,\n WorkingDir=tempfile.gettempdir())\n res = app((aln1_filename, aln2_filename))\n alignment = dict(parse_fasta(res['MuscleOut']))\n new_alignment = {}\n for k, v in alignment.items():\n if k in aln1_int_keys:\n new_alignment[aln1_int_keys[k]] = v\n else:\n new_alignment[aln2_int_keys[k]] = v\n new_alignment = Alignment(new_alignment)\n res.cleanUp()\n del (aln1_collection, aln1_int_map, aln1_int_keys)\n del (aln2_collection, aln2_int_map, aln2_int_keys)\n del (app, res, alignment, params)\n remove(aln1_filename)\n remove(aln2_filename)\n return new_alignment\n",
"role": "user"
},
{
"content": "align_two_alignments",
"role": "assistant"
}
] |
def x():
features = random.sample(string.ascii_letters, NUM_SAMPLES)
num_capitalized = len([letter for letter in features if letter in
string.ascii_uppercase])
num_lowercase = len([letter for letter in features if letter in string.
ascii_lowercase])
if num_capitalized > num_lowercase:
label = 1
else:
label = -1
return label, features
| get_example | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n features = random.sample(string.ascii_letters, NUM_SAMPLES)\n num_capitalized = len([letter for letter in features if letter in\n string.ascii_uppercase])\n num_lowercase = len([letter for letter in features if letter in string.\n ascii_lowercase])\n if num_capitalized > num_lowercase:\n label = 1\n else:\n label = -1\n return label, features\n",
"role": "user"
},
{
"content": "get_example",
"role": "assistant"
}
] |
def x(redis, path):
file_hash = get_cache_buster(redis, path)
return '%s/%s?v=%s' % (oz.settings['static_host'], path, file_hash)
| static_url | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(redis, path):\n file_hash = get_cache_buster(redis, path)\n return '%s/%s?v=%s' % (oz.settings['static_host'], path, file_hash)\n",
"role": "user"
},
{
"content": "static_url",
"role": "assistant"
}
] |
def x(redis, path):
return escape.to_unicode(redis.hget('cache-buster:{}:v3'.format(oz.
settings['s3_bucket']), path))
| get_cache_buster | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(redis, path):\n return escape.to_unicode(redis.hget('cache-buster:{}:v3'.format(oz.\n settings['s3_bucket']), path))\n",
"role": "user"
},
{
"content": "get_cache_buster",
"role": "assistant"
}
] |
def x(redis, path, hash):
redis.hset('cache-buster:{}:v3'.format(oz.settings['s3_bucket']), path,
hash)
| set_cache_buster | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(redis, path, hash):\n redis.hset('cache-buster:{}:v3'.format(oz.settings['s3_bucket']), path,\n hash)\n",
"role": "user"
},
{
"content": "set_cache_buster",
"role": "assistant"
}
] |
def x(s3_bucket=None, validate=False):
global S3Connection
if S3Connection != None:
settings = oz.settings
s3_bucket = s3_bucket or settings['s3_bucket']
opts = {}
if settings['s3_host']:
opts['host'] = settings['s3_host']
if settings['aws_access_key'] and settings['aws_secret_key']:
opts['aws_access_key_id'] = settings['aws_access_key']
opts['aws_secret_access_key'] = settings['aws_secret_key']
return S3Connection(**opts).get_bucket(s3_bucket, validate=validate)
else:
raise Exception(
'S3 not supported in this environment as boto is not installed')
| get_bucket | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(s3_bucket=None, validate=False):\n global S3Connection\n if S3Connection != None:\n settings = oz.settings\n s3_bucket = s3_bucket or settings['s3_bucket']\n opts = {}\n if settings['s3_host']:\n opts['host'] = settings['s3_host']\n if settings['aws_access_key'] and settings['aws_secret_key']:\n opts['aws_access_key_id'] = settings['aws_access_key']\n opts['aws_secret_access_key'] = settings['aws_secret_key']\n return S3Connection(**opts).get_bucket(s3_bucket, validate=validate)\n else:\n raise Exception(\n 'S3 not supported in this environment as boto is not installed')\n",
"role": "user"
},
{
"content": "get_bucket",
"role": "assistant"
}
] |
def x(path, s3_bucket=None):
bucket_name = s3_bucket or oz.settings['s3_bucket']
if bucket_name:
bucket = get_bucket(bucket_name)
key = bucket.get_key(path)
if not key:
key = bucket.new_key(path)
return S3File(key)
else:
return LocalFile(oz.settings['static_path'], path)
| get_file | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(path, s3_bucket=None):\n bucket_name = s3_bucket or oz.settings['s3_bucket']\n if bucket_name:\n bucket = get_bucket(bucket_name)\n key = bucket.get_key(path)\n if not key:\n key = bucket.new_key(path)\n return S3File(key)\n else:\n return LocalFile(oz.settings['static_path'], path)\n",
"role": "user"
},
{
"content": "get_file",
"role": "assistant"
}
] |
def x(self, new_path, replace=False):
new_full_path = os.path.join(self.static_path, new_path)
if replace or not os.path.exists(new_full_path):
shutil.copy2(self.full_path, new_full_path)
return True
return False
| copy | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, new_path, replace=False):\n new_full_path = os.path.join(self.static_path, new_path)\n if replace or not os.path.exists(new_full_path):\n shutil.copy2(self.full_path, new_full_path)\n return True\n return False\n",
"role": "user"
},
{
"content": "copy",
"role": "assistant"
}
] |
def x(self, new_path, replace=False):
if replace or not get_file(new_path).exists():
self.key.copy(self.key.bucket, new_path)
return True
return False
| copy | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, new_path, replace=False):\n if replace or not get_file(new_path).exists():\n self.key.copy(self.key.bucket, new_path)\n return True\n return False\n",
"role": "user"
},
{
"content": "copy",
"role": "assistant"
}
] |
def x():
global _cached_connection
settings = oz.settings
if settings['redis_cache_connections'] and _cached_connection != None:
return _cached_connection
else:
conn = redis.StrictRedis(host=settings['redis_host'], port=settings
['redis_port'], db=settings['redis_db'], password=settings[
'redis_password'], decode_responses=settings[
'redis_decode_responses'], ssl=settings['redis_use_ssl'],
ssl_keyfile=settings['redis_ssl_keyfile'], ssl_certfile=
settings['redis_ssl_certfile'], ssl_cert_reqs=settings[
'redis_ssl_cert_reqs'], ssl_ca_certs=settings['redis_ssl_ca_certs']
)
if settings['redis_cache_connections']:
_cached_connection = conn
return conn
| create_connection | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n global _cached_connection\n settings = oz.settings\n if settings['redis_cache_connections'] and _cached_connection != None:\n return _cached_connection\n else:\n conn = redis.StrictRedis(host=settings['redis_host'], port=settings\n ['redis_port'], db=settings['redis_db'], password=settings[\n 'redis_password'], decode_responses=settings[\n 'redis_decode_responses'], ssl=settings['redis_use_ssl'],\n ssl_keyfile=settings['redis_ssl_keyfile'], ssl_certfile=\n settings['redis_ssl_certfile'], ssl_cert_reqs=settings[\n 'redis_ssl_cert_reqs'], ssl_ca_certs=settings['redis_ssl_ca_certs']\n )\n if settings['redis_cache_connections']:\n _cached_connection = conn\n return conn\n",
"role": "user"
},
{
"content": "create_connection",
"role": "assistant"
}
] |
def x(uc_lines, otu_id_field=1):
hit_miss_index = 0
cluster_id_index = otu_id_field
seq_label_index = 8
otus = {}
unassigned_seqs = []
for line in uc_lines:
if line.startswith('#') or len(line.strip()) == 0:
continue
curr_line = line.split('\t')
if curr_line[hit_miss_index] == 'N':
unassigned_seqs.append(curr_line[seq_label_index].split()[0])
if curr_line[hit_miss_index] == 'H':
curr_seq_label = curr_line[seq_label_index].split()[0]
curr_otu_id = curr_line[cluster_id_index].split()[0]
try:
otus[curr_otu_id].append(curr_seq_label)
except KeyError:
otus[curr_otu_id] = [curr_seq_label]
return otus, unassigned_seqs
| clusters_from_blast_uc_file | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(uc_lines, otu_id_field=1):\n hit_miss_index = 0\n cluster_id_index = otu_id_field\n seq_label_index = 8\n otus = {}\n unassigned_seqs = []\n for line in uc_lines:\n if line.startswith('#') or len(line.strip()) == 0:\n continue\n curr_line = line.split('\\t')\n if curr_line[hit_miss_index] == 'N':\n unassigned_seqs.append(curr_line[seq_label_index].split()[0])\n if curr_line[hit_miss_index] == 'H':\n curr_seq_label = curr_line[seq_label_index].split()[0]\n curr_otu_id = curr_line[cluster_id_index].split()[0]\n try:\n otus[curr_otu_id].append(curr_seq_label)\n except KeyError:\n otus[curr_otu_id] = [curr_seq_label]\n return otus, unassigned_seqs\n",
"role": "user"
},
{
"content": "clusters_from_blast_uc_file",
"role": "assistant"
}
] |
def x(fasta_filepath, output_filepath=None, log_name='sortlen.log',
HALT_EXEC=False, save_intermediate_files=False, remove_usearch_logs=
False, working_dir=None):
if not output_filepath:
_, output_filepath = mkstemp(prefix='usearch_fasta_sort', suffix=
'.fasta')
log_filepath = join(working_dir, log_name)
params = {}
app = Usearch(params, WorkingDir=working_dir, HALT_EXEC=HALT_EXEC)
data = {'--mergesort': fasta_filepath, '--output': output_filepath}
if not remove_usearch_logs:
data['--log'] = log_filepath
app_result = app(data)
return app_result, output_filepath
| usearch_fasta_sort_from_filepath | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fasta_filepath, output_filepath=None, log_name='sortlen.log',\n HALT_EXEC=False, save_intermediate_files=False, remove_usearch_logs=\n False, working_dir=None):\n if not output_filepath:\n _, output_filepath = mkstemp(prefix='usearch_fasta_sort', suffix=\n '.fasta')\n log_filepath = join(working_dir, log_name)\n params = {}\n app = Usearch(params, WorkingDir=working_dir, HALT_EXEC=HALT_EXEC)\n data = {'--mergesort': fasta_filepath, '--output': output_filepath}\n if not remove_usearch_logs:\n data['--log'] = log_filepath\n app_result = app(data)\n return app_result, output_filepath\n",
"role": "user"
},
{
"content": "usearch_fasta_sort_from_filepath",
"role": "assistant"
}
] |
def x(fasta_filepath, output_filepath=None, sizein=True, sizeout=True,
minsize=0, log_name='abundance_sort.log', usersort=False, HALT_EXEC=
False, save_intermediate_files=False, remove_usearch_logs=False,
working_dir=None):
if not output_filepath:
_, output_filepath = mkstemp(prefix='usearch_abundance_sorted',
suffix='.fasta')
log_filepath = join(working_dir, 'minsize_' + str(minsize) + '_' + log_name
)
params = {}
app = Usearch(params, WorkingDir=working_dir, HALT_EXEC=HALT_EXEC)
if usersort:
app.Parameters['--usersort'].on()
if minsize:
app.Parameters['--minsize'].on(minsize)
if sizein:
app.Parameters['--sizein'].on()
if sizeout:
app.Parameters['--sizeout'].on()
data = {'--sortsize': fasta_filepath, '--output': output_filepath}
if not remove_usearch_logs:
data['--log'] = log_filepath
try:
app_result = app(data)
except ApplicationError:
raise ValueError('No data following filter steps, please check ' +
'parameter settings for usearch_qf.')
return app_result, output_filepath
| usearch_sort_by_abundance | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fasta_filepath, output_filepath=None, sizein=True, sizeout=True,\n minsize=0, log_name='abundance_sort.log', usersort=False, HALT_EXEC=\n False, save_intermediate_files=False, remove_usearch_logs=False,\n working_dir=None):\n if not output_filepath:\n _, output_filepath = mkstemp(prefix='usearch_abundance_sorted',\n suffix='.fasta')\n log_filepath = join(working_dir, 'minsize_' + str(minsize) + '_' + log_name\n )\n params = {}\n app = Usearch(params, WorkingDir=working_dir, HALT_EXEC=HALT_EXEC)\n if usersort:\n app.Parameters['--usersort'].on()\n if minsize:\n app.Parameters['--minsize'].on(minsize)\n if sizein:\n app.Parameters['--sizein'].on()\n if sizeout:\n app.Parameters['--sizeout'].on()\n data = {'--sortsize': fasta_filepath, '--output': output_filepath}\n if not remove_usearch_logs:\n data['--log'] = log_filepath\n try:\n app_result = app(data)\n except ApplicationError:\n raise ValueError('No data following filter steps, please check ' +\n 'parameter settings for usearch_qf.')\n return app_result, output_filepath\n",
"role": "user"
},
{
"content": "usearch_sort_by_abundance",
"role": "assistant"
}
] |
def x(fasta_filepath, output_filepath=None, output_uc_filepath=None,
percent_id_err=0.97, sizein=True, sizeout=True, w=64, slots=16769023,
maxrejects=64, log_name='usearch_cluster_err_corrected.log', usersort=
False, HALT_EXEC=False, save_intermediate_files=False,
remove_usearch_logs=False, working_dir=None):
if not output_filepath:
_, output_filepath = mkstemp(prefix='usearch_cluster_err_corrected',
suffix='.fasta')
log_filepath = join(working_dir, log_name)
params = {'--sizein': sizein, '--sizeout': sizeout, '--id':
percent_id_err, '--w': w, '--slots': slots, '--maxrejects': maxrejects}
app = Usearch(params, WorkingDir=working_dir, HALT_EXEC=HALT_EXEC)
if usersort:
app.Parameters['--usersort'].on()
data = {'--cluster': fasta_filepath, '--consout': output_filepath}
if not remove_usearch_logs:
data['--log'] = log_filepath
if output_uc_filepath:
data['--uc'] = output_uc_filepath
app_result = app(data)
return app_result, output_filepath
| usearch_cluster_error_correction | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fasta_filepath, output_filepath=None, output_uc_filepath=None,\n percent_id_err=0.97, sizein=True, sizeout=True, w=64, slots=16769023,\n maxrejects=64, log_name='usearch_cluster_err_corrected.log', usersort=\n False, HALT_EXEC=False, save_intermediate_files=False,\n remove_usearch_logs=False, working_dir=None):\n if not output_filepath:\n _, output_filepath = mkstemp(prefix='usearch_cluster_err_corrected',\n suffix='.fasta')\n log_filepath = join(working_dir, log_name)\n params = {'--sizein': sizein, '--sizeout': sizeout, '--id':\n percent_id_err, '--w': w, '--slots': slots, '--maxrejects': maxrejects}\n app = Usearch(params, WorkingDir=working_dir, HALT_EXEC=HALT_EXEC)\n if usersort:\n app.Parameters['--usersort'].on()\n data = {'--cluster': fasta_filepath, '--consout': output_filepath}\n if not remove_usearch_logs:\n data['--log'] = log_filepath\n if output_uc_filepath:\n data['--uc'] = output_uc_filepath\n app_result = app(data)\n return app_result, output_filepath\n",
"role": "user"
},
{
"content": "usearch_cluster_error_correction",
"role": "assistant"
}
] |
def x(fasta_filepath, output_chimera_filepath=None,
output_non_chimera_filepath=None, abundance_skew=2.0, log_name=
'uchime_de_novo_chimera_filtering.log', usersort=False, HALT_EXEC=False,
save_intermediate_files=False, remove_usearch_logs=False, working_dir=None
):
if not output_chimera_filepath:
_, output_chimera_filepath = mkstemp(prefix='uchime_chimeras_',
suffix='.fasta')
if not output_non_chimera_filepath:
_, output_non_chimera_filepath = mkstemp(prefix=
'uchime_non_chimeras_', suffix='.fasta')
log_filepath = join(working_dir, log_name)
params = {'--abskew': abundance_skew}
app = Usearch(params, WorkingDir=working_dir, HALT_EXEC=HALT_EXEC)
if usersort:
app.Parameters['--usersort'].on()
data = {'--uchime': fasta_filepath, '--chimeras':
output_chimera_filepath, '--nonchimeras': output_non_chimera_filepath}
if not remove_usearch_logs:
data['--log'] = log_filepath
app_result = app(data)
if not save_intermediate_files:
remove_files([output_chimera_filepath])
return app_result, output_non_chimera_filepath
| usearch_chimera_filter_de_novo | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fasta_filepath, output_chimera_filepath=None,\n output_non_chimera_filepath=None, abundance_skew=2.0, log_name=\n 'uchime_de_novo_chimera_filtering.log', usersort=False, HALT_EXEC=False,\n save_intermediate_files=False, remove_usearch_logs=False, working_dir=None\n ):\n if not output_chimera_filepath:\n _, output_chimera_filepath = mkstemp(prefix='uchime_chimeras_',\n suffix='.fasta')\n if not output_non_chimera_filepath:\n _, output_non_chimera_filepath = mkstemp(prefix=\n 'uchime_non_chimeras_', suffix='.fasta')\n log_filepath = join(working_dir, log_name)\n params = {'--abskew': abundance_skew}\n app = Usearch(params, WorkingDir=working_dir, HALT_EXEC=HALT_EXEC)\n if usersort:\n app.Parameters['--usersort'].on()\n data = {'--uchime': fasta_filepath, '--chimeras':\n output_chimera_filepath, '--nonchimeras': output_non_chimera_filepath}\n if not remove_usearch_logs:\n data['--log'] = log_filepath\n app_result = app(data)\n if not save_intermediate_files:\n remove_files([output_chimera_filepath])\n return app_result, output_non_chimera_filepath\n",
"role": "user"
},
{
"content": "usearch_chimera_filter_de_novo",
"role": "assistant"
}
] |
def x(fasta_filepath, output_filepath=None, percent_id=0.97, sizein=True,
sizeout=True, w=64, slots=16769023, maxrejects=64, log_name=
'usearch_cluster_seqs.log', usersort=True, HALT_EXEC=False,
save_intermediate_files=False, remove_usearch_logs=False,
suppress_new_clusters=False, refseqs_fp=None, output_dir=None,
working_dir=None, rev=False):
if not output_filepath:
_, output_filepath = mkstemp(prefix='usearch_cluster_ref_based',
suffix='.uc')
log_filepath = join(working_dir, log_name)
uc_filepath = join(working_dir, 'clustered_seqs_post_chimera.uc')
params = {'--sizein': sizein, '--sizeout': sizeout, '--id': percent_id,
'--w': w, '--slots': slots, '--maxrejects': maxrejects}
app = Usearch(params, WorkingDir=working_dir, HALT_EXEC=HALT_EXEC)
if usersort:
app.Parameters['--usersort'].on()
if rev:
app.Parameters['--rev'].on()
data = {'--query': fasta_filepath, '--uc': uc_filepath, '--db': refseqs_fp}
if not remove_usearch_logs:
data['--log'] = log_filepath
app_result = app(data)
files_to_remove = []
if suppress_new_clusters:
output_fna_filepath = join(output_dir, 'ref_clustered_seqs.fasta')
output_filepath, labels_hits = get_fasta_from_uc_file(fasta_filepath,
uc_filepath, hit_type='H', output_dir=output_dir,
output_fna_filepath=output_fna_filepath)
files_to_remove.append(uc_filepath)
else:
output_fna_clustered = join(output_dir, 'ref_clustered_seqs.fasta')
output_filepath_ref_clusters, labels_hits = get_fasta_from_uc_file(
fasta_filepath, uc_filepath, hit_type='H', output_dir=
output_dir, output_fna_filepath=output_fna_clustered)
output_fna_failures = join(output_dir,
'ref_clustered_seqs_failures.fasta')
output_filepath_failures, labels_hits = get_fasta_from_uc_file(
fasta_filepath, uc_filepath, hit_type='N', output_dir=
output_dir, output_fna_filepath=output_fna_failures)
app_result, output_filepath_clustered_failures = usearch_cluster_seqs(
output_fna_failures, output_filepath=join(output_dir,
'clustered_seqs_reference_failures.fasta'), percent_id=
percent_id, sizein=sizein, sizeout=sizeout, w=w, slots=slots,
maxrejects=maxrejects, save_intermediate_files=
save_intermediate_files, remove_usearch_logs=
remove_usearch_logs, working_dir=working_dir)
output_filepath = concatenate_fastas(output_fna_clustered,
output_fna_failures, output_concat_filepath=join(output_dir,
'concatenated_reference_denovo_clusters.fasta'))
files_to_remove.append(output_fna_clustered)
files_to_remove.append(output_fna_failures)
files_to_remove.append(output_filepath_clustered_failures)
if not save_intermediate_files:
remove_files(files_to_remove)
return app_result, output_filepath
| usearch_cluster_seqs_ref | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fasta_filepath, output_filepath=None, percent_id=0.97, sizein=True,\n sizeout=True, w=64, slots=16769023, maxrejects=64, log_name=\n 'usearch_cluster_seqs.log', usersort=True, HALT_EXEC=False,\n save_intermediate_files=False, remove_usearch_logs=False,\n suppress_new_clusters=False, refseqs_fp=None, output_dir=None,\n working_dir=None, rev=False):\n if not output_filepath:\n _, output_filepath = mkstemp(prefix='usearch_cluster_ref_based',\n suffix='.uc')\n log_filepath = join(working_dir, log_name)\n uc_filepath = join(working_dir, 'clustered_seqs_post_chimera.uc')\n params = {'--sizein': sizein, '--sizeout': sizeout, '--id': percent_id,\n '--w': w, '--slots': slots, '--maxrejects': maxrejects}\n app = Usearch(params, WorkingDir=working_dir, HALT_EXEC=HALT_EXEC)\n if usersort:\n app.Parameters['--usersort'].on()\n if rev:\n app.Parameters['--rev'].on()\n data = {'--query': fasta_filepath, '--uc': uc_filepath, '--db': refseqs_fp}\n if not remove_usearch_logs:\n data['--log'] = log_filepath\n app_result = app(data)\n files_to_remove = []\n if suppress_new_clusters:\n output_fna_filepath = join(output_dir, 'ref_clustered_seqs.fasta')\n output_filepath, labels_hits = get_fasta_from_uc_file(fasta_filepath,\n uc_filepath, hit_type='H', output_dir=output_dir,\n output_fna_filepath=output_fna_filepath)\n files_to_remove.append(uc_filepath)\n else:\n output_fna_clustered = join(output_dir, 'ref_clustered_seqs.fasta')\n output_filepath_ref_clusters, labels_hits = get_fasta_from_uc_file(\n fasta_filepath, uc_filepath, hit_type='H', output_dir=\n output_dir, output_fna_filepath=output_fna_clustered)\n output_fna_failures = join(output_dir,\n 'ref_clustered_seqs_failures.fasta')\n output_filepath_failures, labels_hits = get_fasta_from_uc_file(\n fasta_filepath, uc_filepath, hit_type='N', output_dir=\n output_dir, output_fna_filepath=output_fna_failures)\n app_result, output_filepath_clustered_failures = usearch_cluster_seqs(\n output_fna_failures, output_filepath=join(output_dir,\n 'clustered_seqs_reference_failures.fasta'), percent_id=\n percent_id, sizein=sizein, sizeout=sizeout, w=w, slots=slots,\n maxrejects=maxrejects, save_intermediate_files=\n save_intermediate_files, remove_usearch_logs=\n remove_usearch_logs, working_dir=working_dir)\n output_filepath = concatenate_fastas(output_fna_clustered,\n output_fna_failures, output_concat_filepath=join(output_dir,\n 'concatenated_reference_denovo_clusters.fasta'))\n files_to_remove.append(output_fna_clustered)\n files_to_remove.append(output_fna_failures)\n files_to_remove.append(output_filepath_clustered_failures)\n if not save_intermediate_files:\n remove_files(files_to_remove)\n return app_result, output_filepath\n",
"role": "user"
},
{
"content": "usearch_cluster_seqs_ref",
"role": "assistant"
}
] |
def x(output_fna_clustered, output_fna_failures, output_concat_filepath):
output_fp = open(output_concat_filepath, 'w')
for label, seq in parse_fasta(open(output_fna_clustered, 'U')):
output_fp.write('>%s\n%s\n' % (label, seq))
for label, seq in parse_fasta(open(output_fna_failures, 'U')):
output_fp.write('>%s\n%s\n' % (label, seq))
return output_concat_filepath
| concatenate_fastas | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(output_fna_clustered, output_fna_failures, output_concat_filepath):\n output_fp = open(output_concat_filepath, 'w')\n for label, seq in parse_fasta(open(output_fna_clustered, 'U')):\n output_fp.write('>%s\\n%s\\n' % (label, seq))\n for label, seq in parse_fasta(open(output_fna_failures, 'U')):\n output_fp.write('>%s\\n%s\\n' % (label, seq))\n return output_concat_filepath\n",
"role": "user"
},
{
"content": "concatenate_fastas",
"role": "assistant"
}
] |
def x(fasta_filepath, output_filepath=None, label_prefix='', label_suffix=
'', retain_label_as_comment=False, count_start=0):
fasta_i = open(fasta_filepath, 'U')
if not output_filepath:
_, output_filepath = mkstemp(prefix='enumerated_seqs_', suffix='.fasta'
)
fasta_o = open(output_filepath, 'w')
for label, seq in parse_fasta(fasta_i):
curr_label = '>' + label_prefix + str(count_start) + label_suffix
if retain_label_as_comment:
curr_label += '\t' + label
fasta_o.write(curr_label.strip() + '\n')
fasta_o.write(seq.strip() + '\n')
count_start += 1
return output_filepath
| enumerate_otus | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fasta_filepath, output_filepath=None, label_prefix='', label_suffix=\n '', retain_label_as_comment=False, count_start=0):\n fasta_i = open(fasta_filepath, 'U')\n if not output_filepath:\n _, output_filepath = mkstemp(prefix='enumerated_seqs_', suffix='.fasta'\n )\n fasta_o = open(output_filepath, 'w')\n for label, seq in parse_fasta(fasta_i):\n curr_label = '>' + label_prefix + str(count_start) + label_suffix\n if retain_label_as_comment:\n curr_label += '\\t' + label\n fasta_o.write(curr_label.strip() + '\\n')\n fasta_o.write(seq.strip() + '\\n')\n count_start += 1\n return output_filepath\n",
"role": "user"
},
{
"content": "enumerate_otus",
"role": "assistant"
}
] |
def x(fasta_filepath, uc_filepath, hit_type='H', output_fna_filepath=None,
label_prefix='', output_dir=None):
hit_type_index = 0
seq_label_index = 8
target_label_index = 9
labels_hits = {}
labels_to_keep = []
for line in open(uc_filepath, 'U'):
if line.startswith('#') or len(line.strip()) == 0:
continue
curr_line = line.split('\t')
if curr_line[0] == hit_type:
labels_hits[curr_line[seq_label_index]] = curr_line[
target_label_index].strip()
labels_to_keep.append(curr_line[seq_label_index])
labels_to_keep = set(labels_to_keep)
out_fna = open(output_fna_filepath, 'w')
for label, seq in parse_fasta(open(fasta_filepath, 'U')):
if label in labels_to_keep:
if hit_type == 'H':
out_fna.write('>' + labels_hits[label] + '\n%s\n' % seq)
if hit_type == 'N':
out_fna.write('>' + label + '\n%s\n' % seq)
return output_fna_filepath, labels_hits
| get_fasta_from_uc_file | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fasta_filepath, uc_filepath, hit_type='H', output_fna_filepath=None,\n label_prefix='', output_dir=None):\n hit_type_index = 0\n seq_label_index = 8\n target_label_index = 9\n labels_hits = {}\n labels_to_keep = []\n for line in open(uc_filepath, 'U'):\n if line.startswith('#') or len(line.strip()) == 0:\n continue\n curr_line = line.split('\\t')\n if curr_line[0] == hit_type:\n labels_hits[curr_line[seq_label_index]] = curr_line[\n target_label_index].strip()\n labels_to_keep.append(curr_line[seq_label_index])\n labels_to_keep = set(labels_to_keep)\n out_fna = open(output_fna_filepath, 'w')\n for label, seq in parse_fasta(open(fasta_filepath, 'U')):\n if label in labels_to_keep:\n if hit_type == 'H':\n out_fna.write('>' + labels_hits[label] + '\\n%s\\n' % seq)\n if hit_type == 'N':\n out_fna.write('>' + label + '\\n%s\\n' % seq)\n return output_fna_filepath, labels_hits\n",
"role": "user"
},
{
"content": "get_fasta_from_uc_file",
"role": "assistant"
}
] |
def x(output_fp_de_novo_nonchimeras, output_fp_ref_nonchimeras,
output_combined_fp, chimeras_retention='union'):
de_novo_non_chimeras = []
reference_non_chimeras = []
de_novo_nonchimeras_f = open(output_fp_de_novo_nonchimeras, 'U')
reference_nonchimeras_f = open(output_fp_ref_nonchimeras, 'U')
output_combined_f = open(output_combined_fp, 'w')
for label, seq in parse_fasta(de_novo_nonchimeras_f):
de_novo_non_chimeras.append(label)
de_novo_nonchimeras_f.close()
for label, seq in parse_fasta(reference_nonchimeras_f):
reference_non_chimeras.append(label)
reference_nonchimeras_f.close()
de_novo_non_chimeras = set(de_novo_non_chimeras)
reference_non_chimeras = set(reference_non_chimeras)
if chimeras_retention == 'union':
all_non_chimeras = de_novo_non_chimeras.union(reference_non_chimeras)
elif chimeras_retention == 'intersection':
all_non_chimeras = de_novo_non_chimeras.intersection(
reference_non_chimeras)
de_novo_nonchimeras_f = open(output_fp_de_novo_nonchimeras, 'U')
reference_nonchimeras_f = open(output_fp_ref_nonchimeras, 'U')
labels_written = []
for label, seq in parse_fasta(de_novo_nonchimeras_f):
if label in all_non_chimeras:
if label not in labels_written:
output_combined_f.write('>%s\n%s\n' % (label, seq))
labels_written.append(label)
de_novo_nonchimeras_f.close()
for label, seq in parse_fasta(reference_nonchimeras_f):
if label in all_non_chimeras:
if label not in labels_written:
output_combined_f.write('>%s\n%s\n' % (label, seq))
labels_written.append(label)
reference_nonchimeras_f.close()
output_combined_f.close()
return output_combined_fp
| get_retained_chimeras | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(output_fp_de_novo_nonchimeras, output_fp_ref_nonchimeras,\n output_combined_fp, chimeras_retention='union'):\n de_novo_non_chimeras = []\n reference_non_chimeras = []\n de_novo_nonchimeras_f = open(output_fp_de_novo_nonchimeras, 'U')\n reference_nonchimeras_f = open(output_fp_ref_nonchimeras, 'U')\n output_combined_f = open(output_combined_fp, 'w')\n for label, seq in parse_fasta(de_novo_nonchimeras_f):\n de_novo_non_chimeras.append(label)\n de_novo_nonchimeras_f.close()\n for label, seq in parse_fasta(reference_nonchimeras_f):\n reference_non_chimeras.append(label)\n reference_nonchimeras_f.close()\n de_novo_non_chimeras = set(de_novo_non_chimeras)\n reference_non_chimeras = set(reference_non_chimeras)\n if chimeras_retention == 'union':\n all_non_chimeras = de_novo_non_chimeras.union(reference_non_chimeras)\n elif chimeras_retention == 'intersection':\n all_non_chimeras = de_novo_non_chimeras.intersection(\n reference_non_chimeras)\n de_novo_nonchimeras_f = open(output_fp_de_novo_nonchimeras, 'U')\n reference_nonchimeras_f = open(output_fp_ref_nonchimeras, 'U')\n labels_written = []\n for label, seq in parse_fasta(de_novo_nonchimeras_f):\n if label in all_non_chimeras:\n if label not in labels_written:\n output_combined_f.write('>%s\\n%s\\n' % (label, seq))\n labels_written.append(label)\n de_novo_nonchimeras_f.close()\n for label, seq in parse_fasta(reference_nonchimeras_f):\n if label in all_non_chimeras:\n if label not in labels_written:\n output_combined_f.write('>%s\\n%s\\n' % (label, seq))\n labels_written.append(label)\n reference_nonchimeras_f.close()\n output_combined_f.close()\n return output_combined_fp\n",
"role": "user"
},
{
"content": "get_retained_chimeras",
"role": "assistant"
}
] |
def x(original_fasta, filtered_fasta, output_filepath=None, log_name=
'assign_reads_to_otus.log', perc_id_blast=0.97, global_alignment=True,
HALT_EXEC=False, save_intermediate_files=False, remove_usearch_logs=
False, working_dir=None):
if not output_filepath:
_, output_filepath = mkstemp(prefix='assign_reads_to_otus', suffix=
'.uc')
log_filepath = join(working_dir, log_name)
params = {'--id': perc_id_blast, '--global': global_alignment}
app = Usearch(params, WorkingDir=working_dir, HALT_EXEC=HALT_EXEC)
data = {'--query': original_fasta, '--db': filtered_fasta, '--uc':
output_filepath}
if not remove_usearch_logs:
data['--log'] = log_filepath
app_result = app(data)
return app_result, output_filepath
| assign_reads_to_otus | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(original_fasta, filtered_fasta, output_filepath=None, log_name=\n 'assign_reads_to_otus.log', perc_id_blast=0.97, global_alignment=True,\n HALT_EXEC=False, save_intermediate_files=False, remove_usearch_logs=\n False, working_dir=None):\n if not output_filepath:\n _, output_filepath = mkstemp(prefix='assign_reads_to_otus', suffix=\n '.uc')\n log_filepath = join(working_dir, log_name)\n params = {'--id': perc_id_blast, '--global': global_alignment}\n app = Usearch(params, WorkingDir=working_dir, HALT_EXEC=HALT_EXEC)\n data = {'--query': original_fasta, '--db': filtered_fasta, '--uc':\n output_filepath}\n if not remove_usearch_logs:\n data['--log'] = log_filepath\n app_result = app(data)\n return app_result, output_filepath\n",
"role": "user"
},
{
"content": "assign_reads_to_otus",
"role": "assistant"
}
] |
def x(seq_path, output_dir='.', rev=False, minlen=64, remove_usearch_logs=
False, HALT_EXEC=False, output_fna_filepath=None, output_uc_filepath=
None, log_name='abundance_sorted.log', threads=1.0):
if not output_fna_filepath:
_, output_fna_filepath = mkstemp(prefix='abundance_sorted', suffix=
'.fna')
if not output_uc_filepath:
_, output_uc_filepath = mkstemp(prefix='abundance_sorted', suffix='.uc'
)
log_filepath = join(output_dir, log_name)
params = {'--minseqlength': minlen, '--sizeout': True,
'--derep_fulllength': seq_path, '--output': output_fna_filepath,
'--uc': output_uc_filepath, '--threads': threads}
if rev:
params['--strand'] = 'both'
if not remove_usearch_logs:
params['--log'] = log_filepath
app = Usearch61(params, WorkingDir=output_dir, HALT_EXEC=HALT_EXEC)
app_result = app()
return output_fna_filepath, output_uc_filepath, app_result
| sort_by_abundance_usearch61 | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(seq_path, output_dir='.', rev=False, minlen=64, remove_usearch_logs=\n False, HALT_EXEC=False, output_fna_filepath=None, output_uc_filepath=\n None, log_name='abundance_sorted.log', threads=1.0):\n if not output_fna_filepath:\n _, output_fna_filepath = mkstemp(prefix='abundance_sorted', suffix=\n '.fna')\n if not output_uc_filepath:\n _, output_uc_filepath = mkstemp(prefix='abundance_sorted', suffix='.uc'\n )\n log_filepath = join(output_dir, log_name)\n params = {'--minseqlength': minlen, '--sizeout': True,\n '--derep_fulllength': seq_path, '--output': output_fna_filepath,\n '--uc': output_uc_filepath, '--threads': threads}\n if rev:\n params['--strand'] = 'both'\n if not remove_usearch_logs:\n params['--log'] = log_filepath\n app = Usearch61(params, WorkingDir=output_dir, HALT_EXEC=HALT_EXEC)\n app_result = app()\n return output_fna_filepath, output_uc_filepath, app_result\n",
"role": "user"
},
{
"content": "sort_by_abundance_usearch61",
"role": "assistant"
}
] |
def x(seq_path, output_dir='.', minlen=64, remove_usearch_logs=False,
HALT_EXEC=False, output_fna_filepath=None, log_name='length_sorted.log'):
if not output_fna_filepath:
_, output_fna_filepath = mkstemp(prefix='length_sorted', suffix='.fna')
log_filepath = join(output_dir, log_name)
params = {'--minseqlength': minlen, '--sortbylength': seq_path,
'--output': output_fna_filepath}
if not remove_usearch_logs:
params['--log'] = log_filepath
app = Usearch61(params, WorkingDir=output_dir, HALT_EXEC=HALT_EXEC)
app_result = app()
return output_fna_filepath, app_result
| sort_by_length_usearch61 | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(seq_path, output_dir='.', minlen=64, remove_usearch_logs=False,\n HALT_EXEC=False, output_fna_filepath=None, log_name='length_sorted.log'):\n if not output_fna_filepath:\n _, output_fna_filepath = mkstemp(prefix='length_sorted', suffix='.fna')\n log_filepath = join(output_dir, log_name)\n params = {'--minseqlength': minlen, '--sortbylength': seq_path,\n '--output': output_fna_filepath}\n if not remove_usearch_logs:\n params['--log'] = log_filepath\n app = Usearch61(params, WorkingDir=output_dir, HALT_EXEC=HALT_EXEC)\n app_result = app()\n return output_fna_filepath, app_result\n",
"role": "user"
},
{
"content": "sort_by_length_usearch61",
"role": "assistant"
}
] |
def x(intermediate_fasta, refseqs_fp, percent_id=0.97, rev=False, minlen=64,
output_dir='.', remove_usearch_logs=False, wordlength=8,
usearch61_maxrejects=32, usearch61_maxaccepts=1, HALT_EXEC=False,
output_uc_filepath=None, log_filepath='ref_clustered.log', threads=1.0):
log_filepath = join(output_dir, log_filepath)
params = {'--usearch_global': intermediate_fasta, '--db': refseqs_fp,
'--minseqlength': minlen, '--id': percent_id, '--uc':
output_uc_filepath, '--wordlength': wordlength, '--maxrejects':
usearch61_maxrejects, '--maxaccepts': usearch61_maxaccepts,
'--threads': threads}
if not remove_usearch_logs:
params['--log'] = log_filepath
if rev:
params['--strand'] = 'both'
else:
params['--strand'] = 'plus'
clusters_fp = output_uc_filepath
app = Usearch61(params, WorkingDir=output_dir, HALT_EXEC=HALT_EXEC)
app_result = app()
return clusters_fp, app_result
| usearch61_cluster_ref | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(intermediate_fasta, refseqs_fp, percent_id=0.97, rev=False, minlen=64,\n output_dir='.', remove_usearch_logs=False, wordlength=8,\n usearch61_maxrejects=32, usearch61_maxaccepts=1, HALT_EXEC=False,\n output_uc_filepath=None, log_filepath='ref_clustered.log', threads=1.0):\n log_filepath = join(output_dir, log_filepath)\n params = {'--usearch_global': intermediate_fasta, '--db': refseqs_fp,\n '--minseqlength': minlen, '--id': percent_id, '--uc':\n output_uc_filepath, '--wordlength': wordlength, '--maxrejects':\n usearch61_maxrejects, '--maxaccepts': usearch61_maxaccepts,\n '--threads': threads}\n if not remove_usearch_logs:\n params['--log'] = log_filepath\n if rev:\n params['--strand'] = 'both'\n else:\n params['--strand'] = 'plus'\n clusters_fp = output_uc_filepath\n app = Usearch61(params, WorkingDir=output_dir, HALT_EXEC=HALT_EXEC)\n app_result = app()\n return clusters_fp, app_result\n",
"role": "user"
},
{
"content": "usearch61_cluster_ref",
"role": "assistant"
}
] |
def x(intermediate_fasta, percent_id=0.97, minlen=64, output_dir='.',
remove_usearch_logs=False, wordlength=8, usearch61_maxrejects=8,
usearch61_maxaccepts=1, HALT_EXEC=False, output_uc_filepath=None,
log_name='fast_clustered.log', threads=1.0):
log_filepath = join(output_dir, log_name)
params = {'--minseqlength': minlen, '--cluster_fast':
intermediate_fasta, '--id': percent_id, '--uc': output_uc_filepath,
'--wordlength': wordlength, '--maxrejects': usearch61_maxrejects,
'--maxaccepts': usearch61_maxaccepts, '--usersort': True,
'--threads': threads}
if not remove_usearch_logs:
params['--log'] = log_filepath
clusters_fp = output_uc_filepath
app = Usearch61(params, WorkingDir=output_dir, HALT_EXEC=HALT_EXEC)
app_result = app()
return clusters_fp, app_result
| usearch61_fast_cluster | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(intermediate_fasta, percent_id=0.97, minlen=64, output_dir='.',\n remove_usearch_logs=False, wordlength=8, usearch61_maxrejects=8,\n usearch61_maxaccepts=1, HALT_EXEC=False, output_uc_filepath=None,\n log_name='fast_clustered.log', threads=1.0):\n log_filepath = join(output_dir, log_name)\n params = {'--minseqlength': minlen, '--cluster_fast':\n intermediate_fasta, '--id': percent_id, '--uc': output_uc_filepath,\n '--wordlength': wordlength, '--maxrejects': usearch61_maxrejects,\n '--maxaccepts': usearch61_maxaccepts, '--usersort': True,\n '--threads': threads}\n if not remove_usearch_logs:\n params['--log'] = log_filepath\n clusters_fp = output_uc_filepath\n app = Usearch61(params, WorkingDir=output_dir, HALT_EXEC=HALT_EXEC)\n app_result = app()\n return clusters_fp, app_result\n",
"role": "user"
},
{
"content": "usearch61_fast_cluster",
"role": "assistant"
}
] |
def x(intermediate_fasta, percent_id=0.97, minlen=64, rev=False, output_dir
='.', remove_usearch_logs=False, wordlength=8, usearch61_maxrejects=32,
usearch61_maxaccepts=1, sizeorder=False, HALT_EXEC=False,
output_uc_filepath=None, log_name='smallmem_clustered.log', sizeout=
False, consout_filepath=None):
log_filepath = join(output_dir, log_name)
params = {'--minseqlength': minlen, '--cluster_smallmem':
intermediate_fasta, '--id': percent_id, '--uc': output_uc_filepath,
'--wordlength': wordlength, '--maxrejects': usearch61_maxrejects,
'--maxaccepts': usearch61_maxaccepts, '--usersort': True}
if sizeorder:
params['--sizeorder'] = True
if not remove_usearch_logs:
params['--log'] = log_filepath
if rev:
params['--strand'] = 'both'
else:
params['--strand'] = 'plus'
if sizeout:
params['--sizeout'] = True
if consout_filepath:
params['--consout'] = consout_filepath
clusters_fp = output_uc_filepath
app = Usearch61(params, WorkingDir=output_dir, HALT_EXEC=HALT_EXEC)
app_result = app()
return clusters_fp, app_result
| usearch61_smallmem_cluster | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(intermediate_fasta, percent_id=0.97, minlen=64, rev=False, output_dir\n ='.', remove_usearch_logs=False, wordlength=8, usearch61_maxrejects=32,\n usearch61_maxaccepts=1, sizeorder=False, HALT_EXEC=False,\n output_uc_filepath=None, log_name='smallmem_clustered.log', sizeout=\n False, consout_filepath=None):\n log_filepath = join(output_dir, log_name)\n params = {'--minseqlength': minlen, '--cluster_smallmem':\n intermediate_fasta, '--id': percent_id, '--uc': output_uc_filepath,\n '--wordlength': wordlength, '--maxrejects': usearch61_maxrejects,\n '--maxaccepts': usearch61_maxaccepts, '--usersort': True}\n if sizeorder:\n params['--sizeorder'] = True\n if not remove_usearch_logs:\n params['--log'] = log_filepath\n if rev:\n params['--strand'] = 'both'\n else:\n params['--strand'] = 'plus'\n if sizeout:\n params['--sizeout'] = True\n if consout_filepath:\n params['--consout'] = consout_filepath\n clusters_fp = output_uc_filepath\n app = Usearch61(params, WorkingDir=output_dir, HALT_EXEC=HALT_EXEC)\n app_result = app()\n return clusters_fp, app_result\n",
"role": "user"
},
{
"content": "usearch61_smallmem_cluster",
"role": "assistant"
}
] |
def x(abundance_fp, uchime_denovo_fp, minlen=64, output_dir='.',
remove_usearch_logs=False, uchime_denovo_log_fp='uchime_denovo.log',
usearch61_minh=0.28, usearch61_xn=8.0, usearch61_dn=1.4,
usearch61_mindiffs=3, usearch61_mindiv=0.8, usearch61_abundance_skew=
2.0, HALT_EXEC=False):
params = {'--minseqlength': minlen, '--uchime_denovo': abundance_fp,
'--uchimeout': uchime_denovo_fp, '--minh': usearch61_minh, '--xn':
usearch61_xn, '--dn': usearch61_dn, '--mindiffs':
usearch61_mindiffs, '--mindiv': usearch61_mindiv, '--abskew':
usearch61_abundance_skew}
if not remove_usearch_logs:
params['--log'] = uchime_denovo_log_fp
app = Usearch61(params, WorkingDir=output_dir, HALT_EXEC=HALT_EXEC)
app_result = app()
return uchime_denovo_fp, app_result
| usearch61_chimera_check_denovo | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(abundance_fp, uchime_denovo_fp, minlen=64, output_dir='.',\n remove_usearch_logs=False, uchime_denovo_log_fp='uchime_denovo.log',\n usearch61_minh=0.28, usearch61_xn=8.0, usearch61_dn=1.4,\n usearch61_mindiffs=3, usearch61_mindiv=0.8, usearch61_abundance_skew=\n 2.0, HALT_EXEC=False):\n params = {'--minseqlength': minlen, '--uchime_denovo': abundance_fp,\n '--uchimeout': uchime_denovo_fp, '--minh': usearch61_minh, '--xn':\n usearch61_xn, '--dn': usearch61_dn, '--mindiffs':\n usearch61_mindiffs, '--mindiv': usearch61_mindiv, '--abskew':\n usearch61_abundance_skew}\n if not remove_usearch_logs:\n params['--log'] = uchime_denovo_log_fp\n app = Usearch61(params, WorkingDir=output_dir, HALT_EXEC=HALT_EXEC)\n app_result = app()\n return uchime_denovo_fp, app_result\n",
"role": "user"
},
{
"content": "usearch61_chimera_check_denovo",
"role": "assistant"
}
] |
def x(abundance_fp, uchime_ref_fp, reference_seqs_fp, minlen=64, output_dir
='.', remove_usearch_logs=False, uchime_ref_log_fp='uchime_ref.log',
usearch61_minh=0.28, usearch61_xn=8.0, usearch61_dn=1.4,
usearch61_mindiffs=3, usearch61_mindiv=0.8, threads=1.0, HALT_EXEC=False):
params = {'--minseqlength': minlen, '--uchime_ref': abundance_fp,
'--uchimeout': uchime_ref_fp, '--db': reference_seqs_fp, '--minh':
usearch61_minh, '--xn': usearch61_xn, '--dn': usearch61_dn,
'--mindiffs': usearch61_mindiffs, '--mindiv': usearch61_mindiv,
'--strand': 'plus', '--threads': threads}
if not remove_usearch_logs:
params['--log'] = uchime_ref_log_fp
app = Usearch61(params, WorkingDir=output_dir, HALT_EXEC=HALT_EXEC)
app_result = app()
return uchime_ref_fp, app_result
| usearch61_chimera_check_ref | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(abundance_fp, uchime_ref_fp, reference_seqs_fp, minlen=64, output_dir\n ='.', remove_usearch_logs=False, uchime_ref_log_fp='uchime_ref.log',\n usearch61_minh=0.28, usearch61_xn=8.0, usearch61_dn=1.4,\n usearch61_mindiffs=3, usearch61_mindiv=0.8, threads=1.0, HALT_EXEC=False):\n params = {'--minseqlength': minlen, '--uchime_ref': abundance_fp,\n '--uchimeout': uchime_ref_fp, '--db': reference_seqs_fp, '--minh':\n usearch61_minh, '--xn': usearch61_xn, '--dn': usearch61_dn,\n '--mindiffs': usearch61_mindiffs, '--mindiv': usearch61_mindiv,\n '--strand': 'plus', '--threads': threads}\n if not remove_usearch_logs:\n params['--log'] = uchime_ref_log_fp\n app = Usearch61(params, WorkingDir=output_dir, HALT_EXEC=HALT_EXEC)\n app_result = app()\n return uchime_ref_fp, app_result\n",
"role": "user"
},
{
"content": "usearch61_chimera_check_ref",
"role": "assistant"
}
] |
def x(dereplicated_uc_lines):
dereplicated_clusters = {}
seed_hit_ix = 0
seq_id_ix = 8
seed_id_ix = 9
for line in dereplicated_uc_lines:
if line.startswith('#') or len(line.strip()) == 0:
continue
curr_line = line.strip().split('\t')
if curr_line[seed_hit_ix] == 'S':
dereplicated_clusters[curr_line[seq_id_ix]] = []
if curr_line[seed_hit_ix] == 'H':
curr_seq_id = curr_line[seq_id_ix]
dereplicated_clusters[curr_line[seed_id_ix]].append(curr_seq_id)
return dereplicated_clusters
| parse_dereplicated_uc | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(dereplicated_uc_lines):\n dereplicated_clusters = {}\n seed_hit_ix = 0\n seq_id_ix = 8\n seed_id_ix = 9\n for line in dereplicated_uc_lines:\n if line.startswith('#') or len(line.strip()) == 0:\n continue\n curr_line = line.strip().split('\\t')\n if curr_line[seed_hit_ix] == 'S':\n dereplicated_clusters[curr_line[seq_id_ix]] = []\n if curr_line[seed_hit_ix] == 'H':\n curr_seq_id = curr_line[seq_id_ix]\n dereplicated_clusters[curr_line[seed_id_ix]].append(curr_seq_id)\n return dereplicated_clusters\n",
"role": "user"
},
{
"content": "parse_dereplicated_uc",
"role": "assistant"
}
] |
def x(clustered_uc_lines, otu_prefix='denovo', ref_clustered=False):
clusters = {}
failures = []
seed_hit_ix = 0
otu_id_ix = 1
seq_id_ix = 8
ref_id_ix = 9
for line in clustered_uc_lines:
if line.startswith('#') or len(line.strip()) == 0:
continue
curr_line = line.strip().split('\t')
if curr_line[seed_hit_ix] == 'S':
clusters[otu_prefix + curr_line[otu_id_ix]] = [curr_line[
seq_id_ix].split(';')[0].split()[0]]
if curr_line[seed_hit_ix] == 'H':
curr_id = curr_line[seq_id_ix].split(';')[0].split()[0]
if ref_clustered:
try:
clusters[otu_prefix + curr_line[ref_id_ix]].append(curr_id)
except KeyError:
clusters[otu_prefix + curr_line[ref_id_ix]] = [curr_id]
else:
clusters[otu_prefix + curr_line[otu_id_ix]].append(curr_id)
if curr_line[seed_hit_ix] == 'N':
failures.append(curr_line[seq_id_ix].split(';')[0])
return clusters, failures
| parse_usearch61_clusters | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(clustered_uc_lines, otu_prefix='denovo', ref_clustered=False):\n clusters = {}\n failures = []\n seed_hit_ix = 0\n otu_id_ix = 1\n seq_id_ix = 8\n ref_id_ix = 9\n for line in clustered_uc_lines:\n if line.startswith('#') or len(line.strip()) == 0:\n continue\n curr_line = line.strip().split('\\t')\n if curr_line[seed_hit_ix] == 'S':\n clusters[otu_prefix + curr_line[otu_id_ix]] = [curr_line[\n seq_id_ix].split(';')[0].split()[0]]\n if curr_line[seed_hit_ix] == 'H':\n curr_id = curr_line[seq_id_ix].split(';')[0].split()[0]\n if ref_clustered:\n try:\n clusters[otu_prefix + curr_line[ref_id_ix]].append(curr_id)\n except KeyError:\n clusters[otu_prefix + curr_line[ref_id_ix]] = [curr_id]\n else:\n clusters[otu_prefix + curr_line[otu_id_ix]].append(curr_id)\n if curr_line[seed_hit_ix] == 'N':\n failures.append(curr_line[seq_id_ix].split(';')[0])\n return clusters, failures\n",
"role": "user"
},
{
"content": "parse_usearch61_clusters",
"role": "assistant"
}
] |
def x(de_novo_clusters, dereplicated_clusters):
clusters = {}
for curr_denovo_key in de_novo_clusters.keys():
clusters[curr_denovo_key] = de_novo_clusters[curr_denovo_key]
curr_clusters = []
for curr_denovo_id in de_novo_clusters[curr_denovo_key]:
curr_clusters += dereplicated_clusters[curr_denovo_id]
clusters[curr_denovo_key] += curr_clusters
return clusters
| merge_clusters_dereplicated_seqs | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(de_novo_clusters, dereplicated_clusters):\n clusters = {}\n for curr_denovo_key in de_novo_clusters.keys():\n clusters[curr_denovo_key] = de_novo_clusters[curr_denovo_key]\n curr_clusters = []\n for curr_denovo_id in de_novo_clusters[curr_denovo_key]:\n curr_clusters += dereplicated_clusters[curr_denovo_id]\n clusters[curr_denovo_key] += curr_clusters\n return clusters\n",
"role": "user"
},
{
"content": "merge_clusters_dereplicated_seqs",
"role": "assistant"
}
] |
def x(failures, dereplicated_clusters):
curr_failures = set(failures)
dereplicated_ids = set(dereplicated_clusters)
for curr_failure in curr_failures:
if curr_failure in dereplicated_ids:
failures += dereplicated_clusters[curr_failure]
return failures
| merge_failures_dereplicated_seqs | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(failures, dereplicated_clusters):\n curr_failures = set(failures)\n dereplicated_ids = set(dereplicated_clusters)\n for curr_failure in curr_failures:\n if curr_failure in dereplicated_ids:\n failures += dereplicated_clusters[curr_failure]\n return failures\n",
"role": "user"
},
{
"content": "merge_failures_dereplicated_seqs",
"role": "assistant"
}
] |
def x(seq_path, failures, output_fasta_fp):
parsed_out = open(output_fasta_fp, 'w')
for label, seq in parse_fasta(open(seq_path), 'U'):
curr_label = label.split()[0]
if curr_label in failures:
parsed_out.write('>%s\n%s\n' % (label, seq))
parsed_out.close()
return output_fasta_fp
| parse_usearch61_failures | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(seq_path, failures, output_fasta_fp):\n parsed_out = open(output_fasta_fp, 'w')\n for label, seq in parse_fasta(open(seq_path), 'U'):\n curr_label = label.split()[0]\n if curr_label in failures:\n parsed_out.write('>%s\\n%s\\n' % (label, seq))\n parsed_out.close()\n return output_fasta_fp\n",
"role": "user"
},
{
"content": "parse_usearch61_failures",
"role": "assistant"
}
] |
def x(self, status_code, **kwargs):
if oz.settings.get('debug'):
exception_type, exception_value, tback = sys.exc_info()
is_breakpoint = isinstance(exception_value, oz.error_pages.
DebugBreakException)
frames = oz.error_pages.get_frames(tback, is_breakpoint)
frames.reverse()
if is_breakpoint:
exception_type = 'Debug breakpoint'
exception_value = ''
self.render(oz.settings['error_pages_template'], exception_type=
exception_type, exception_value=exception_value, frames=frames,
request_input=self.request.body, request_cookies=self.cookies,
request_headers=self.request.headers, request_path=self.request
.uri, request_method=self.request.method, response_output=''.
join(self._write_buffer), response_headers=self._headers,
prettify_object=oz.error_pages.prettify_object)
return oz.break_trigger
| on_error_page_write_error | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, status_code, **kwargs):\n if oz.settings.get('debug'):\n exception_type, exception_value, tback = sys.exc_info()\n is_breakpoint = isinstance(exception_value, oz.error_pages.\n DebugBreakException)\n frames = oz.error_pages.get_frames(tback, is_breakpoint)\n frames.reverse()\n if is_breakpoint:\n exception_type = 'Debug breakpoint'\n exception_value = ''\n self.render(oz.settings['error_pages_template'], exception_type=\n exception_type, exception_value=exception_value, frames=frames,\n request_input=self.request.body, request_cookies=self.cookies,\n request_headers=self.request.headers, request_path=self.request\n .uri, request_method=self.request.method, response_output=''.\n join(self._write_buffer), response_headers=self._headers,\n prettify_object=oz.error_pages.prettify_object)\n return oz.break_trigger\n",
"role": "user"
},
{
"content": "on_error_page_write_error",
"role": "assistant"
}
] |
def x(self, name):
value = self.get_cookie(name)
if value != None:
self.clear_cookie(name)
return escape.url_unescape(value)
| get_blink_cookie | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, name):\n value = self.get_cookie(name)\n if value != None:\n self.clear_cookie(name)\n return escape.url_unescape(value)\n",
"role": "user"
},
{
"content": "get_blink_cookie",
"role": "assistant"
}
] |
def x(self, message, type='info'):
self.set_cookie('blink_message', escape.url_escape(message), httponly=True)
self.set_cookie('blink_type', escape.url_escape(type), httponly=True)
| set_blink | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, message, type='info'):\n self.set_cookie('blink_message', escape.url_escape(message), httponly=True)\n self.set_cookie('blink_type', escape.url_escape(type), httponly=True)\n",
"role": "user"
},
{
"content": "set_blink",
"role": "assistant"
}
] |
def x(lines):
clusters = []
curr_cluster = []
for l in lines:
if l.startswith('>Cluster'):
if not curr_cluster:
continue
clusters.append(curr_cluster)
curr_cluster = []
else:
curr_cluster.append(clean_cluster_seq_id(l.split()[2]))
if curr_cluster:
clusters.append(curr_cluster)
return clusters
| parse_cdhit_clstr_file | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(lines):\n clusters = []\n curr_cluster = []\n for l in lines:\n if l.startswith('>Cluster'):\n if not curr_cluster:\n continue\n clusters.append(curr_cluster)\n curr_cluster = []\n else:\n curr_cluster.append(clean_cluster_seq_id(l.split()[2]))\n if curr_cluster:\n clusters.append(curr_cluster)\n return clusters\n",
"role": "user"
},
{
"content": "parse_cdhit_clstr_file",
"role": "assistant"
}
] |
def x(self, data):
if data:
self.Parameters['-i'].on(super(CD_HIT, self).
_input_as_multiline_string(data))
return ''
| input_as_multiline_string | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, data):\n if data:\n self.Parameters['-i'].on(super(CD_HIT, self).\n _input_as_multiline_string(data))\n return ''\n",
"role": "user"
},
{
"content": "input_as_multiline_string",
"role": "assistant"
}
] |
def x(self, data):
if data:
self.Parameters['-i'].on(super(CD_HIT, self)._input_as_lines(data))
return ''
| input_as_lines | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, data):\n if data:\n self.Parameters['-i'].on(super(CD_HIT, self)._input_as_lines(data))\n return ''\n",
"role": "user"
},
{
"content": "input_as_lines",
"role": "assistant"
}
] |
def x(self, data):
result = {}
result['FASTA'] = ResultPath(Path=self._get_seqs_outfile())
result['CLSTR'] = ResultPath(Path=self._get_clstr_outfile())
return result
| get_result_paths | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, data):\n result = {}\n result['FASTA'] = ResultPath(Path=self._get_seqs_outfile())\n result['CLSTR'] = ResultPath(Path=self._get_clstr_outfile())\n return result\n",
"role": "user"
},
{
"content": "get_result_paths",
"role": "assistant"
}
] |
def x(self, tokens):
look_ahead = LookAhead(tokens)
segments = Segment()
while not look_ahead.empty():
if look_ahead.peek().type not in self.whitespace:
paragraph = MatchableSegment(look_ahead.i)
while not look_ahead.empty() and look_ahead.peek(
).type not in self.paragraph_end:
if look_ahead.peek().type == 'tab_open':
tab_depth = 1
sentence = MatchableSegment(look_ahead.i, [next(
look_ahead)])
while not look_ahead.empty() and tab_depth > 0:
tab_depth += look_ahead.peek().type == 'tab_open'
tab_depth -= look_ahead.peek().type == 'tab_close'
sentence.append(next(look_ahead))
paragraph.append(sentence)
elif look_ahead.peek().type not in self.whitespace:
sentence = MatchableSegment(look_ahead.i, [next(
look_ahead)])
sub_depth = int(sentence[0].type in SUB_OPEN)
while not look_ahead.empty():
sub_depth += look_ahead.peek().type in SUB_OPEN
sub_depth -= look_ahead.peek().type in SUB_CLOSE
sentence.append(next(look_ahead))
if sentence[-1
].type in self.sentence_end and sub_depth <= 0:
non_whitespace = sum(s.type not in self.
whitespace for s in sentence)
if non_whitespace >= self.min_sentence:
break
paragraph.append(sentence)
else:
whitespace = Segment(look_ahead.i, [next(look_ahead)])
paragraph.append(whitespace)
segments.append(paragraph)
else:
whitespace = Segment(look_ahead.i, [next(look_ahead)])
segments.append(whitespace)
return segments
| segment | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, tokens):\n look_ahead = LookAhead(tokens)\n segments = Segment()\n while not look_ahead.empty():\n if look_ahead.peek().type not in self.whitespace:\n paragraph = MatchableSegment(look_ahead.i)\n while not look_ahead.empty() and look_ahead.peek(\n ).type not in self.paragraph_end:\n if look_ahead.peek().type == 'tab_open':\n tab_depth = 1\n sentence = MatchableSegment(look_ahead.i, [next(\n look_ahead)])\n while not look_ahead.empty() and tab_depth > 0:\n tab_depth += look_ahead.peek().type == 'tab_open'\n tab_depth -= look_ahead.peek().type == 'tab_close'\n sentence.append(next(look_ahead))\n paragraph.append(sentence)\n elif look_ahead.peek().type not in self.whitespace:\n sentence = MatchableSegment(look_ahead.i, [next(\n look_ahead)])\n sub_depth = int(sentence[0].type in SUB_OPEN)\n while not look_ahead.empty():\n sub_depth += look_ahead.peek().type in SUB_OPEN\n sub_depth -= look_ahead.peek().type in SUB_CLOSE\n sentence.append(next(look_ahead))\n if sentence[-1\n ].type in self.sentence_end and sub_depth <= 0:\n non_whitespace = sum(s.type not in self.\n whitespace for s in sentence)\n if non_whitespace >= self.min_sentence:\n break\n paragraph.append(sentence)\n else:\n whitespace = Segment(look_ahead.i, [next(look_ahead)])\n paragraph.append(whitespace)\n segments.append(paragraph)\n else:\n whitespace = Segment(look_ahead.i, [next(look_ahead)])\n segments.append(whitespace)\n return segments\n",
"role": "user"
},
{
"content": "segment",
"role": "assistant"
}
] |
def x(self, payload, token, expiration=None, priority=None, identifier=None):
created_conn = False
while not created_conn:
if len(self.conns) == 0:
self.conns.append(PushConnection(self, self.address, self.
certfile, self.keyfile))
created_conn = True
conn = random.choice(self.conns)
try:
conn.send(payload, token, expiration=expiration, priority=
priority, identifier=identifier)
return
except:
logger.info('Connection died: removing')
self.conns.remove(conn)
raise SendFailedException()
| send | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, payload, token, expiration=None, priority=None, identifier=None):\n created_conn = False\n while not created_conn:\n if len(self.conns) == 0:\n self.conns.append(PushConnection(self, self.address, self.\n certfile, self.keyfile))\n created_conn = True\n conn = random.choice(self.conns)\n try:\n conn.send(payload, token, expiration=expiration, priority=\n priority, identifier=identifier)\n return\n except:\n logger.info('Connection died: removing')\n self.conns.remove(conn)\n raise SendFailedException()\n",
"role": "user"
},
{
"content": "send",
"role": "assistant"
}
] |
def x(self):
if not self.fbaddress:
raise Exception(
'Attempted to fetch feedback but no feedback_address supplied')
fbconn = FeedbackConnection(self, self.fbaddress, self.certfile, self.
keyfile)
return fbconn.get_all()
| get_all_feedback | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if not self.fbaddress:\n raise Exception(\n 'Attempted to fetch feedback but no feedback_address supplied')\n fbconn = FeedbackConnection(self, self.fbaddress, self.certfile, self.\n keyfile)\n return fbconn.get_all()\n",
"role": "user"
},
{
"content": "get_all_feedback",
"role": "assistant"
}
] |
def x(self, response):
if response.startswith('\n'):
response = response[1:]
tree = etree.fromstring(response)
return tree
| parse_xml | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, response):\n if response.startswith('\\n'):\n response = response[1:]\n tree = etree.fromstring(response)\n return tree\n",
"role": "user"
},
{
"content": "parse_xml",
"role": "assistant"
}
] |
def x(self, error):
status_code = error.response.status_code
message = error.message
if 403 == status_code:
raise NewRelicInvalidApiKeyException(message)
elif 404 == status_code:
raise NewRelicUnknownApplicationException(message)
elif 422 == status_code:
raise NewRelicInvalidParameterException(message)
else:
raise NewRelicApiException(message)
| handle_api_error | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, error):\n status_code = error.response.status_code\n message = error.message\n if 403 == status_code:\n raise NewRelicInvalidApiKeyException(message)\n elif 404 == status_code:\n raise NewRelicUnknownApplicationException(message)\n elif 422 == status_code:\n raise NewRelicInvalidParameterException(message)\n else:\n raise NewRelicApiException(message)\n",
"role": "user"
},
{
"content": "handle_api_error",
"role": "assistant"
}
] |
def x(self, api_call, window=60):
current = datetime.datetime.now()
try:
previous = getattr(self, api_call.__name__ + '_window')
previous.__str__
except AttributeError:
now = datetime.datetime.now()
outside_window = datetime.timedelta(seconds=window + 1)
previous = now - outside_window
if current - previous > datetime.timedelta(seconds=window):
setattr(self, api_call.__name__ + '_window', current)
else:
timeout = window - (current - previous).seconds
raise NewRelicApiRateLimitException(str(timeout))
| api_rate_limit_exceeded | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, api_call, window=60):\n current = datetime.datetime.now()\n try:\n previous = getattr(self, api_call.__name__ + '_window')\n previous.__str__\n except AttributeError:\n now = datetime.datetime.now()\n outside_window = datetime.timedelta(seconds=window + 1)\n previous = now - outside_window\n if current - previous > datetime.timedelta(seconds=window):\n setattr(self, api_call.__name__ + '_window', current)\n else:\n timeout = window - (current - previous).seconds\n raise NewRelicApiRateLimitException(str(timeout))\n",
"role": "user"
},
{
"content": "api_rate_limit_exceeded",
"role": "assistant"
}
] |
def x(self):
endpoint = 'https://rpm.newrelic.com'
uri = '{endpoint}/accounts/{id}/applications.xml'.format(endpoint=
endpoint, id=self.account_id)
response = self._make_get_request(uri)
applications = []
for application in response.findall('.//application'):
application_properties = {}
for field in application:
application_properties[field.tag] = field.text
applications.append(Application(application_properties))
return applications
| view_applications | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n endpoint = 'https://rpm.newrelic.com'\n uri = '{endpoint}/accounts/{id}/applications.xml'.format(endpoint=\n endpoint, id=self.account_id)\n response = self._make_get_request(uri)\n applications = []\n for application in response.findall('.//application'):\n application_properties = {}\n for field in application:\n application_properties[field.tag] = field.text\n applications.append(Application(application_properties))\n return applications\n",
"role": "user"
},
{
"content": "view_applications",
"role": "assistant"
}
] |
def x(self, applications):
endpoint = 'https://api.newrelic.com'
uri = ('{endpoint}/api/v1/accounts/{account_id}/applications/delete.xml'
.format(endpoint=endpoint, account_id=self.account_id))
payload = applications
response = self._make_post_request(uri, payload)
failed_deletions = {}
for application in response.findall('.//application'):
if not 'deleted' in application.findall('.//result')[0].text:
failed_deletions['app_id'] = application.get('id')
return failed_deletions
| delete_applications | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, applications):\n endpoint = 'https://api.newrelic.com'\n uri = ('{endpoint}/api/v1/accounts/{account_id}/applications/delete.xml'\n .format(endpoint=endpoint, account_id=self.account_id))\n payload = applications\n response = self._make_post_request(uri, payload)\n failed_deletions = {}\n for application in response.findall('.//application'):\n if not 'deleted' in application.findall('.//result')[0].text:\n failed_deletions['app_id'] = application.get('id')\n return failed_deletions\n",
"role": "user"
},
{
"content": "delete_applications",
"role": "assistant"
}
] |
def x(self, application_id=None, application_name=None, description=None,
revision=None, changelog=None, user=None):
endpoint = 'https://rpm.newrelic.com'
uri = '{endpoint}/deployments.xml'.format(endpoint=endpoint)
deploy_event = {}
if not application_id is None:
deploy_event['deployment[application_id]'] = application_id
elif not application_name is None:
deploy_event['deployment[app_name]'] = application_name
else:
raise NewRelicInvalidParameterException(
'Must specify either application_id or application_name.')
if not description is None:
deploy_event['deployment[description]'] = description
if not revision is None:
deploy_event['deployment[revision]'] = revision
if not changelog is None:
deploy_event['deployment[changelog]'] = changelog
if not user is None:
deploy_event['deployment[user]'] = user
response = self._make_post_request(uri, deploy_event)
result = {}
for value in response:
result[value.tag] = value.text
return result
| notify_deployment | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, application_id=None, application_name=None, description=None,\n revision=None, changelog=None, user=None):\n endpoint = 'https://rpm.newrelic.com'\n uri = '{endpoint}/deployments.xml'.format(endpoint=endpoint)\n deploy_event = {}\n if not application_id is None:\n deploy_event['deployment[application_id]'] = application_id\n elif not application_name is None:\n deploy_event['deployment[app_name]'] = application_name\n else:\n raise NewRelicInvalidParameterException(\n 'Must specify either application_id or application_name.')\n if not description is None:\n deploy_event['deployment[description]'] = description\n if not revision is None:\n deploy_event['deployment[revision]'] = revision\n if not changelog is None:\n deploy_event['deployment[changelog]'] = changelog\n if not user is None:\n deploy_event['deployment[user]'] = user\n response = self._make_post_request(uri, deploy_event)\n result = {}\n for value in response:\n result[value.tag] = value.text\n return result\n",
"role": "user"
},
{
"content": "notify_deployment",
"role": "assistant"
}
] |
def x(self, agent_id, re=None, limit=5000):
self._api_rate_limit_exceeded(self.get_metric_names)
parameters = {'re': re, 'limit': limit}
endpoint = 'https://api.newrelic.com'
uri = '{endpoint}/api/v1/applications/{agent_id}/metrics.xml'.format(
endpoint=endpoint, agent_id=agent_id)
response = self._make_get_request(uri, parameters=parameters, timeout=
max(self.timeout, 5.0))
metrics = {}
for metric in response.findall('.//metric'):
fields = []
for field in metric.findall('.//field'):
fields.append(field.get('name'))
metrics[metric.get('name')] = fields
return metrics
| get_metric_names | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, agent_id, re=None, limit=5000):\n self._api_rate_limit_exceeded(self.get_metric_names)\n parameters = {'re': re, 'limit': limit}\n endpoint = 'https://api.newrelic.com'\n uri = '{endpoint}/api/v1/applications/{agent_id}/metrics.xml'.format(\n endpoint=endpoint, agent_id=agent_id)\n response = self._make_get_request(uri, parameters=parameters, timeout=\n max(self.timeout, 5.0))\n metrics = {}\n for metric in response.findall('.//metric'):\n fields = []\n for field in metric.findall('.//field'):\n fields.append(field.get('name'))\n metrics[metric.get('name')] = fields\n return metrics\n",
"role": "user"
},
{
"content": "get_metric_names",
"role": "assistant"
}
] |
def x(self, applications, metrics, field, begin, end, summary=False):
self._api_rate_limit_exceeded(self.get_metric_data)
parameters = {}
try:
int(applications[0])
except ValueError:
app_string = 'app'
else:
app_string = 'app_id'
if len(applications) > 1:
app_string = app_string + '[]'
parameters[app_string] = applications
parameters['metrics[]'] = metrics
parameters['field'] = field
parameters['begin'] = begin
parameters['end'] = end
parameters['summary'] = int(summary)
endpoint = 'https://api.newrelic.com'
uri = '{endpoint}/api/v1/accounts/{account_id}/metrics/data.xml'.format(
endpoint=endpoint, account_id=self.account_id)
response = self._make_get_request(uri, parameters=parameters, timeout=
max(self.timeout, 5.0))
metrics = []
for metric in response.findall('.//metric'):
metrics.append(Metric(metric))
return metrics
| get_metric_data | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, applications, metrics, field, begin, end, summary=False):\n self._api_rate_limit_exceeded(self.get_metric_data)\n parameters = {}\n try:\n int(applications[0])\n except ValueError:\n app_string = 'app'\n else:\n app_string = 'app_id'\n if len(applications) > 1:\n app_string = app_string + '[]'\n parameters[app_string] = applications\n parameters['metrics[]'] = metrics\n parameters['field'] = field\n parameters['begin'] = begin\n parameters['end'] = end\n parameters['summary'] = int(summary)\n endpoint = 'https://api.newrelic.com'\n uri = '{endpoint}/api/v1/accounts/{account_id}/metrics/data.xml'.format(\n endpoint=endpoint, account_id=self.account_id)\n response = self._make_get_request(uri, parameters=parameters, timeout=\n max(self.timeout, 5.0))\n metrics = []\n for metric in response.findall('.//metric'):\n metrics.append(Metric(metric))\n return metrics\n",
"role": "user"
},
{
"content": "get_metric_data",
"role": "assistant"
}
] |
def x(self, application_id):
endpoint = 'https://rpm.newrelic.com'
remote_file = 'threshold_values.xml'
uri = ('{endpoint}/accounts/{account_id}/applications/{app_id}/{xml}'.
format(endpoint=endpoint, account_id=self.account_id, app_id=
application_id, xml=remote_file))
response = self._make_get_request(uri)
thresholds = []
for threshold_value in response.findall('.//threshold_value'):
properties = {}
for tag, text in threshold_value.items():
properties[tag] = text
thresholds.append(Threshold(properties))
return thresholds
| get_threshold_values | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, application_id):\n endpoint = 'https://rpm.newrelic.com'\n remote_file = 'threshold_values.xml'\n uri = ('{endpoint}/accounts/{account_id}/applications/{app_id}/{xml}'.\n format(endpoint=endpoint, account_id=self.account_id, app_id=\n application_id, xml=remote_file))\n response = self._make_get_request(uri)\n thresholds = []\n for threshold_value in response.findall('.//threshold_value'):\n properties = {}\n for tag, text in threshold_value.items():\n properties[tag] = text\n thresholds.append(Threshold(properties))\n return thresholds\n",
"role": "user"
},
{
"content": "get_threshold_values",
"role": "assistant"
}
] |
def x(self):
endpoint = 'https://api.newrelic.com'
uri = '{endpoint}/api/v1/accounts/{id}/servers.xml'.format(endpoint=
endpoint, id=self.account_id)
response = self._make_get_request(uri)
servers = []
for server in response.findall('.//server'):
server_properties = {}
for field in server:
server_properties[field.tag] = field.text
servers.append(Server(server_properties))
return servers
| view_servers | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n endpoint = 'https://api.newrelic.com'\n uri = '{endpoint}/api/v1/accounts/{id}/servers.xml'.format(endpoint=\n endpoint, id=self.account_id)\n response = self._make_get_request(uri)\n servers = []\n for server in response.findall('.//server'):\n server_properties = {}\n for field in server:\n server_properties[field.tag] = field.text\n servers.append(Server(server_properties))\n return servers\n",
"role": "user"
},
{
"content": "view_servers",
"role": "assistant"
}
] |
def x(self, server_id):
endpoint = 'https://api.newrelic.com'
uri = ('{endpoint}/api/v1/accounts/{account_id}/servers/{server_id}.xml'
.format(endpoint=endpoint, account_id=self.account_id, server_id=
server_id))
response = self._make_delete_request(uri)
failed_deletions = []
for server in response.findall('.//server'):
if not 'deleted' in server.findall('.//result')[0].text:
failed_deletions.append({'server_id': server.get('id')})
return failed_deletions
| delete_servers | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, server_id):\n endpoint = 'https://api.newrelic.com'\n uri = ('{endpoint}/api/v1/accounts/{account_id}/servers/{server_id}.xml'\n .format(endpoint=endpoint, account_id=self.account_id, server_id=\n server_id))\n response = self._make_delete_request(uri)\n failed_deletions = []\n for server in response.findall('.//server'):\n if not 'deleted' in server.findall('.//result')[0].text:\n failed_deletions.append({'server_id': server.get('id')})\n return failed_deletions\n",
"role": "user"
},
{
"content": "delete_servers",
"role": "assistant"
}
] |
def x(reads1_infile_path, reads2_infile_path, perc_max_diff=None,
min_overlap=None, outfile_label='fastqjoin', params={}, working_dir=
tempfile.gettempdir(), SuppressStderr=True, SuppressStdout=True,
HALT_EXEC=False):
abs_r1_path = os.path.abspath(reads1_infile_path)
abs_r2_path = os.path.abspath(reads2_infile_path)
infile_paths = [abs_r1_path, abs_r2_path]
for p in infile_paths:
if not os.path.exists(p):
raise IOError('File not found at: %s' % p)
fastq_join_app = FastqJoin(params=params, WorkingDir=working_dir,
SuppressStderr=SuppressStderr, SuppressStdout=SuppressStdout,
HALT_EXEC=HALT_EXEC)
if perc_max_diff is not None:
if isinstance(perc_max_diff, int) and 0 <= perc_max_diff <= 100:
fastq_join_app.Parameters['-p'].on(perc_max_diff)
else:
raise ValueError('perc_max_diff must be int between 0-100!')
if min_overlap is not None:
if isinstance(min_overlap, int) and 0 < min_overlap:
fastq_join_app.Parameters['-m'].on(min_overlap)
else:
raise ValueError('min_overlap must be an int >= 0!')
if outfile_label is not None:
if isinstance(outfile_label, str):
fastq_join_app.Parameters['-o'].on(outfile_label + '.')
else:
raise ValueError('outfile_label must be a string!')
else:
pass
result = fastq_join_app(infile_paths)
path_dict = {}
path_dict['Assembled'] = result['Assembled'].name
path_dict['UnassembledReads1'] = result['UnassembledReads1'].name
path_dict['UnassembledReads2'] = result['UnassembledReads2'].name
for path in path_dict.values():
if not os.path.exists(path):
raise IOError('Output file not found at: %s' % path)
for key, file_path in path_dict.items():
new_file_path = file_path + '.fastq'
shutil.move(file_path, new_file_path)
path_dict[key] = new_file_path
return path_dict
| join_paired_end_reads_fastqjoin | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(reads1_infile_path, reads2_infile_path, perc_max_diff=None,\n min_overlap=None, outfile_label='fastqjoin', params={}, working_dir=\n tempfile.gettempdir(), SuppressStderr=True, SuppressStdout=True,\n HALT_EXEC=False):\n abs_r1_path = os.path.abspath(reads1_infile_path)\n abs_r2_path = os.path.abspath(reads2_infile_path)\n infile_paths = [abs_r1_path, abs_r2_path]\n for p in infile_paths:\n if not os.path.exists(p):\n raise IOError('File not found at: %s' % p)\n fastq_join_app = FastqJoin(params=params, WorkingDir=working_dir,\n SuppressStderr=SuppressStderr, SuppressStdout=SuppressStdout,\n HALT_EXEC=HALT_EXEC)\n if perc_max_diff is not None:\n if isinstance(perc_max_diff, int) and 0 <= perc_max_diff <= 100:\n fastq_join_app.Parameters['-p'].on(perc_max_diff)\n else:\n raise ValueError('perc_max_diff must be int between 0-100!')\n if min_overlap is not None:\n if isinstance(min_overlap, int) and 0 < min_overlap:\n fastq_join_app.Parameters['-m'].on(min_overlap)\n else:\n raise ValueError('min_overlap must be an int >= 0!')\n if outfile_label is not None:\n if isinstance(outfile_label, str):\n fastq_join_app.Parameters['-o'].on(outfile_label + '.')\n else:\n raise ValueError('outfile_label must be a string!')\n else:\n pass\n result = fastq_join_app(infile_paths)\n path_dict = {}\n path_dict['Assembled'] = result['Assembled'].name\n path_dict['UnassembledReads1'] = result['UnassembledReads1'].name\n path_dict['UnassembledReads2'] = result['UnassembledReads2'].name\n for path in path_dict.values():\n if not os.path.exists(path):\n raise IOError('Output file not found at: %s' % path)\n for key, file_path in path_dict.items():\n new_file_path = file_path + '.fastq'\n shutil.move(file_path, new_file_path)\n path_dict[key] = new_file_path\n return path_dict\n",
"role": "user"
},
{
"content": "join_paired_end_reads_fastqjoin",
"role": "assistant"
}
] |
def x(self):
if self.Parameters['-o'].isOn():
output_path = self._absolute(str(self.Parameters['-o'].Value))
else:
raise ValueError('No output path specified.')
return output_path
| get_output_path | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.Parameters['-o'].isOn():\n output_path = self._absolute(str(self.Parameters['-o'].Value))\n else:\n raise ValueError('No output path specified.')\n return output_path\n",
"role": "user"
},
{
"content": "get_output_path",
"role": "assistant"
}
] |
def x(self):
if self.Parameters['-r'].isOn():
stitch_path = self._absolute(str(self.Parameters['-r'].Value))
return stitch_path
elif self.Parameters['-r'].isOff():
return None
| get_stitch_report_path | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.Parameters['-r'].isOn():\n stitch_path = self._absolute(str(self.Parameters['-r'].Value))\n return stitch_path\n elif self.Parameters['-r'].isOff():\n return None\n",
"role": "user"
},
{
"content": "get_stitch_report_path",
"role": "assistant"
}
] |
def x(self, data):
output_path = self._get_output_path()
result = {}
result['Assembled'] = ResultPath(Path=output_path + 'join', IsWritten=True)
result['UnassembledReads1'] = ResultPath(Path=output_path + 'un1',
IsWritten=True)
result['UnassembledReads2'] = ResultPath(Path=output_path + 'un2',
IsWritten=True)
stitch_path = self._get_stitch_report_path()
if stitch_path:
result['Report'] = ResultPath(Path=stitch_path, IsWritten=True)
mate_path_string = output_path + 'join2'
mate_unassembled_path_string = output_path + 'un3'
if os.path.exists(mate_path_string) and os.path.exists(
mate_unassembled_path_string):
result['Mate'] = ResultPath(Path=mate_path_string, IsWritten=True)
result['MateUnassembled'] = ResultPath(Path=
mate_unassembled_path_string, IsWritten=True)
else:
pass
return result
| get_result_paths | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, data):\n output_path = self._get_output_path()\n result = {}\n result['Assembled'] = ResultPath(Path=output_path + 'join', IsWritten=True)\n result['UnassembledReads1'] = ResultPath(Path=output_path + 'un1',\n IsWritten=True)\n result['UnassembledReads2'] = ResultPath(Path=output_path + 'un2',\n IsWritten=True)\n stitch_path = self._get_stitch_report_path()\n if stitch_path:\n result['Report'] = ResultPath(Path=stitch_path, IsWritten=True)\n mate_path_string = output_path + 'join2'\n mate_unassembled_path_string = output_path + 'un3'\n if os.path.exists(mate_path_string) and os.path.exists(\n mate_unassembled_path_string):\n result['Mate'] = ResultPath(Path=mate_path_string, IsWritten=True)\n result['MateUnassembled'] = ResultPath(Path=\n mate_unassembled_path_string, IsWritten=True)\n else:\n pass\n return result\n",
"role": "user"
},
{
"content": "get_result_paths",
"role": "assistant"
}
] |
def x(fields, as_dict=False):
_fields = {}
if fields in ['~', None, False, True, {}, []]:
_fields = {}
elif isinstance(fields, dict):
_fields.update({unicode(k).strip(): int(v) for k, v in fields.
iteritems()})
elif isinstance(fields, basestring):
_fields.update({unicode(s).strip(): (1) for s in fields.split(',')})
elif isinstance(fields, (list, tuple)):
_fields.update({unicode(s).strip(): (1) for s in fields})
else:
raise ValueError('invalid fields value')
if as_dict:
return _fields
else:
return sorted(_fields.keys())
| parse_fields | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fields, as_dict=False):\n _fields = {}\n if fields in ['~', None, False, True, {}, []]:\n _fields = {}\n elif isinstance(fields, dict):\n _fields.update({unicode(k).strip(): int(v) for k, v in fields.\n iteritems()})\n elif isinstance(fields, basestring):\n _fields.update({unicode(s).strip(): (1) for s in fields.split(',')})\n elif isinstance(fields, (list, tuple)):\n _fields.update({unicode(s).strip(): (1) for s in fields})\n else:\n raise ValueError('invalid fields value')\n if as_dict:\n return _fields\n else:\n return sorted(_fields.keys())\n",
"role": "user"
},
{
"content": "parse_fields",
"role": "assistant"
}
] |
def x(date, func='date'):
if isinstance(date, basestring):
date = date.strip()
if not date:
return '_end == None'
if date == '~':
return ''
before = lambda d: '_start < %s("%s")' % (func, ts2dt(d) if d else None)
after = lambda d: '(_end >= %s("%s") or _end == None)' % (func, ts2dt(d
) if d else None)
split = date.split('~')
split = [re.sub('\\+\\d\\d:\\d\\d', '', d.replace('T', ' ')) for d in split
]
if len(split) == 1:
return '%s and %s' % (before(split[0]), after(split[0]))
elif split[0] in ['', None]:
return before(split[1])
elif split[1] in ['', None]:
return after(split[0])
else:
return '%s and %s' % (before(split[1]), after(split[0]))
| date_range | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(date, func='date'):\n if isinstance(date, basestring):\n date = date.strip()\n if not date:\n return '_end == None'\n if date == '~':\n return ''\n before = lambda d: '_start < %s(\"%s\")' % (func, ts2dt(d) if d else None)\n after = lambda d: '(_end >= %s(\"%s\") or _end == None)' % (func, ts2dt(d\n ) if d else None)\n split = date.split('~')\n split = [re.sub('\\\\+\\\\d\\\\d:\\\\d\\\\d', '', d.replace('T', ' ')) for d in split\n ]\n if len(split) == 1:\n return '%s and %s' % (before(split[0]), after(split[0]))\n elif split[0] in ['', None]:\n return before(split[1])\n elif split[1] in ['', None]:\n return after(split[0])\n else:\n return '%s and %s' % (before(split[1]), after(split[0]))\n",
"role": "user"
},
{
"content": "date_range",
"role": "assistant"
}
] |
def x(table, query=None, date=None, fields=None, distinct=False, limit=None,
alias=None):
date = date_range(date)
limit = int(limit or -1)
if query and date:
query = '%s and %s' % (query, date)
elif date:
query = date
elif query:
pass
else:
query = None
fields = parse_fields(fields=fields) or None
fields = fields if fields else table.columns
msg = 'parse(query=%s, fields=%s)' % (query, fields)
logger.debug(msg)
kwargs = {}
if query:
interpreter = MQLInterpreter(table)
query = interpreter.parse(query)
kwargs['whereclause'] = query
if distinct:
kwargs['distinct'] = distinct
query = select(fields, from_obj=table, **kwargs)
if limit >= 1:
query = query.limit(limit)
if alias:
query = query.alias(alias)
return query
| parse | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(table, query=None, date=None, fields=None, distinct=False, limit=None,\n alias=None):\n date = date_range(date)\n limit = int(limit or -1)\n if query and date:\n query = '%s and %s' % (query, date)\n elif date:\n query = date\n elif query:\n pass\n else:\n query = None\n fields = parse_fields(fields=fields) or None\n fields = fields if fields else table.columns\n msg = 'parse(query=%s, fields=%s)' % (query, fields)\n logger.debug(msg)\n kwargs = {}\n if query:\n interpreter = MQLInterpreter(table)\n query = interpreter.parse(query)\n kwargs['whereclause'] = query\n if distinct:\n kwargs['distinct'] = distinct\n query = select(fields, from_obj=table, **kwargs)\n if limit >= 1:\n query = query.limit(limit)\n if alias:\n query = query.alias(alias)\n return query\n",
"role": "user"
},
{
"content": "parse",
"role": "assistant"
}
] |
def x(self, func, *func_args):
worker = self.pick_sticky(0)
args = (func,) + func_args
coro = worker.enqueue(enums.Task.FUNC, args)
asyncio.ensure_future(coro)
| enqueue_sync | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, func, *func_args):\n worker = self.pick_sticky(0)\n args = (func,) + func_args\n coro = worker.enqueue(enums.Task.FUNC, args)\n asyncio.ensure_future(coro)\n",
"role": "user"
},
{
"content": "enqueue_sync",
"role": "assistant"
}
] |
async def async_enqueue_sync(self, func, *func_args):
"""
Enqueue an arbitrary synchronous function.
"""
worker = self.pick_sticky(0)
args = (func,) + func_args
await worker.enqueue(enums.Task.FUNC, args)
| async_enqueue_sync | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def async_enqueue_sync(self, func, *func_args):\n \"\"\"\n Enqueue an arbitrary synchronous function.\n \"\"\"\n worker = self.pick_sticky(0)\n args = (func,) + func_args\n await worker.enqueue(enums.Task.FUNC, args)\n",
"role": "user"
},
{
"content": "async_enqueue_sync",
"role": "assistant"
}
] |
def x(self, resource):
worker = self.pick_sticky(resource.url_string)
coro = worker.enqueue(enums.Task.DOWNLOAD, (resource,))
asyncio.ensure_future(coro)
| enqueue_download | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, resource):\n worker = self.pick_sticky(resource.url_string)\n coro = worker.enqueue(enums.Task.DOWNLOAD, (resource,))\n asyncio.ensure_future(coro)\n",
"role": "user"
},
{
"content": "enqueue_download",
"role": "assistant"
}
] |
async def async_enqueue_download(self, resource):
"""
Enqueue the download of the given foreign resource.
"""
worker = self.pick_sticky(resource.url_string)
await worker.enqueue(enums.Task.DOWNLOAD, (resource,))
| async_enqueue_download | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def async_enqueue_download(self, resource):\n \"\"\"\n Enqueue the download of the given foreign resource.\n \"\"\"\n worker = self.pick_sticky(resource.url_string)\n await worker.enqueue(enums.Task.DOWNLOAD, (resource,))\n",
"role": "user"
},
{
"content": "async_enqueue_download",
"role": "assistant"
}
] |
def x(self, converter, from_resource, to_resource):
worker = self.pick_sticky(from_resource.url_string)
args = converter, from_resource, to_resource
coro = worker.enqueue(enums.Task.CONVERT, args)
asyncio.ensure_future(coro)
| enqueue_convert | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, converter, from_resource, to_resource):\n worker = self.pick_sticky(from_resource.url_string)\n args = converter, from_resource, to_resource\n coro = worker.enqueue(enums.Task.CONVERT, args)\n asyncio.ensure_future(coro)\n",
"role": "user"
},
{
"content": "enqueue_convert",
"role": "assistant"
}
] |
async def async_enqueue_convert(self, converter, from_, to):
"""
Enqueue use of the given converter to convert to given
from and to resources.
"""
worker = self.pick_sticky(from_.url_string)
args = converter, from_, to
await worker.enqueue(enums.Task.CONVERT, args)
| async_enqueue_convert | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def async_enqueue_convert(self, converter, from_, to):\n \"\"\"\n Enqueue use of the given converter to convert to given\n from and to resources.\n \"\"\"\n worker = self.pick_sticky(from_.url_string)\n args = converter, from_, to\n await worker.enqueue(enums.Task.CONVERT, args)\n",
"role": "user"
},
{
"content": "async_enqueue_convert",
"role": "assistant"
}
] |
async def async_enqueue_multiconvert(self, url_string, to_type):
"""
Enqueue a multi-step conversion process, from the given URL string
(which is assumed to have been downloaded / resolved)
"""
worker = self.pick_sticky(url_string)
args = url_string, to_type
await worker.enqueue(enums.Task.MULTICONVERT, args)
| async_enqueue_multiconvert | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def async_enqueue_multiconvert(self, url_string, to_type):\n \"\"\"\n Enqueue a multi-step conversion process, from the given URL string\n (which is assumed to have been downloaded / resolved)\n \"\"\"\n worker = self.pick_sticky(url_string)\n args = url_string, to_type\n await worker.enqueue(enums.Task.MULTICONVERT, args)\n",
"role": "user"
},
{
"content": "async_enqueue_multiconvert",
"role": "assistant"
}
] |
def x(self, **kwargs):
if self._strain_out_of_plane is not None:
return self._strain_out_of_plane / -2.0 * (self.unstrained.c11(**
kwargs) / self.unstrained.c12(**kwargs))
else:
return 1 - self.unstrained.a(**kwargs) / self.substrate.a(**kwargs)
| strain_in_plane | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, **kwargs):\n if self._strain_out_of_plane is not None:\n return self._strain_out_of_plane / -2.0 * (self.unstrained.c11(**\n kwargs) / self.unstrained.c12(**kwargs))\n else:\n return 1 - self.unstrained.a(**kwargs) / self.substrate.a(**kwargs)\n",
"role": "user"
},
{
"content": "strain_in_plane",
"role": "assistant"
}
] |
def x(self, **kwargs):
if self._strain_out_of_plane is not None:
return self._strain_out_of_plane
else:
return -2 * self.unstrained.c12(**kwargs) / self.unstrained.c11(**
kwargs) * self.strain_in_plane(**kwargs)
| strain_out_of_plane | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, **kwargs):\n if self._strain_out_of_plane is not None:\n return self._strain_out_of_plane\n else:\n return -2 * self.unstrained.c12(**kwargs) / self.unstrained.c11(**\n kwargs) * self.strain_in_plane(**kwargs)\n",
"role": "user"
},
{
"content": "strain_out_of_plane",
"role": "assistant"
}
] |
def x(self, **kwargs):
if self.substrate is not None:
return self.substrate.a(**kwargs)
else:
return self.unstrained.a(**kwargs) / (1.0 - self.strain_in_plane(**
kwargs))
| substrate_a | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, **kwargs):\n if self.substrate is not None:\n return self.substrate.a(**kwargs)\n else:\n return self.unstrained.a(**kwargs) / (1.0 - self.strain_in_plane(**\n kwargs))\n",
"role": "user"
},
{
"content": "substrate_a",
"role": "assistant"
}
] |
def x(self, **kwargs):
return self.unstrained.CBO(**kwargs) + self.CBO_strain_shift(**kwargs)
| CBO | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, **kwargs):\n return self.unstrained.CBO(**kwargs) + self.CBO_strain_shift(**kwargs)\n",
"role": "user"
},
{
"content": "CBO",
"role": "assistant"
}
] |
def x(self, **kwargs):
return self.unstrained.CBO_Gamma(**kwargs) + self.CBO_strain_shift(**kwargs
)
| CBO_Gamma | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, **kwargs):\n return self.unstrained.CBO_Gamma(**kwargs) + self.CBO_strain_shift(**kwargs\n )\n",
"role": "user"
},
{
"content": "CBO_Gamma",
"role": "assistant"
}
] |
def x(self, **kwargs):
return self.unstrained.CBO_L(**kwargs) + self.CBO_strain_shift(**kwargs)
| CBO_L | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, **kwargs):\n return self.unstrained.CBO_L(**kwargs) + self.CBO_strain_shift(**kwargs)\n",
"role": "user"
},
{
"content": "CBO_L",
"role": "assistant"
}
] |
def x(self, **kwargs):
return self.unstrained.CBO_X(**kwargs) + self.CBO_strain_shift(**kwargs)
| CBO_X | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, **kwargs):\n return self.unstrained.CBO_X(**kwargs) + self.CBO_strain_shift(**kwargs)\n",
"role": "user"
},
{
"content": "CBO_X",
"role": "assistant"
}
] |
def x(self, **kwargs):
return self.unstrained.Eg(**kwargs) + self.Eg_strain_shift(**kwargs)
| Eg | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, **kwargs):\n return self.unstrained.Eg(**kwargs) + self.Eg_strain_shift(**kwargs)\n",
"role": "user"
},
{
"content": "Eg",
"role": "assistant"
}
] |
def x(self):
for subsegment_or_token in self:
if isinstance(subsegment_or_token, Segment):
subsegment = subsegment_or_token
for token in subsegment.tokens():
yield token
else:
token = subsegment_or_token
yield token
| tokens | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n for subsegment_or_token in self:\n if isinstance(subsegment_or_token, Segment):\n subsegment = subsegment_or_token\n for token in subsegment.tokens():\n yield token\n else:\n token = subsegment_or_token\n yield token\n",
"role": "user"
},
{
"content": "tokens",
"role": "assistant"
}
] |
def x(self, status_code, **kwargs):
return_error = {'code': self.get_status()}
exc_info = kwargs.get('exc_info')
if exc_info and isinstance(exc_info[1], oz.json_api.ApiError):
return_error['error'] = exc_info[1].message
else:
return_error['error'] = API_ERROR_CODE_MAP.get(self.get_status(),
'Unknown error')
if oz.settings.get('debug'):
return_error['trace'] = ''.join(traceback.format_exception(*exc_info))
self.finish(return_error)
return oz.break_trigger
| api_on_write_error | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, status_code, **kwargs):\n return_error = {'code': self.get_status()}\n exc_info = kwargs.get('exc_info')\n if exc_info and isinstance(exc_info[1], oz.json_api.ApiError):\n return_error['error'] = exc_info[1].message\n else:\n return_error['error'] = API_ERROR_CODE_MAP.get(self.get_status(),\n 'Unknown error')\n if oz.settings.get('debug'):\n return_error['trace'] = ''.join(traceback.format_exception(*exc_info))\n self.finish(return_error)\n return oz.break_trigger\n",
"role": "user"
},
{
"content": "api_on_write_error",
"role": "assistant"
}
] |
def x(self, obj):
callback = self.get_argument('callback', None) if oz.settings['allow_jsonp'
] else None
if callback and not CALLBACK_VALIDATOR.match(callback):
raise oz.json_api.ApiError(
'Invalid callback identifier - only functions with ASCII characters are allowed'
)
json = escape.json_encode(obj)
if callback:
self.set_header('Content-Type', 'application/javascript; charset=UTF-8'
)
self.finish('%s(%s)' % (callback, json))
else:
self.set_header('Content-Type', 'application/json; charset=UTF-8')
self.finish(json)
| respond | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, obj):\n callback = self.get_argument('callback', None) if oz.settings['allow_jsonp'\n ] else None\n if callback and not CALLBACK_VALIDATOR.match(callback):\n raise oz.json_api.ApiError(\n 'Invalid callback identifier - only functions with ASCII characters are allowed'\n )\n json = escape.json_encode(obj)\n if callback:\n self.set_header('Content-Type', 'application/javascript; charset=UTF-8'\n )\n self.finish('%s(%s)' % (callback, json))\n else:\n self.set_header('Content-Type', 'application/json; charset=UTF-8')\n self.finish(json)\n",
"role": "user"
},
{
"content": "respond",
"role": "assistant"
}
] |
def x(self):
if self._decoded_body == None:
raw_content_type = self.request.headers.get('content-type') or ''
content_type = raw_content_type.split(';')[0].strip().lower()
if content_type == 'application/json':
try:
self._decoded_body = escape.json_decode(self.request.body)
except:
raise oz.json_api.ApiError('Bad JSON body')
else:
raise oz.json_api.ApiError('JSON body expected')
return self._decoded_body
| body | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self._decoded_body == None:\n raw_content_type = self.request.headers.get('content-type') or ''\n content_type = raw_content_type.split(';')[0].strip().lower()\n if content_type == 'application/json':\n try:\n self._decoded_body = escape.json_decode(self.request.body)\n except:\n raise oz.json_api.ApiError('Bad JSON body')\n else:\n raise oz.json_api.ApiError('JSON body expected')\n return self._decoded_body\n",
"role": "user"
},
{
"content": "body",
"role": "assistant"
}
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.