source_code stringlengths 45 7.24k | function_name stringlengths 1 111 | messages list |
|---|---|---|
def x(obj):
try:
return pprint.pformat(str(obj))
except UnicodeDecodeError as e:
raise
except Exception as e:
return '[could not display: <%s: %s>]' % (e.__class__.__name__, str(e))
| prettify_object | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(obj):\n try:\n return pprint.pformat(str(obj))\n except UnicodeDecodeError as e:\n raise\n except Exception as e:\n return '[could not display: <%s: %s>]' % (e.__class__.__name__, str(e))\n",
"role": "user"
},
{
"content": "prettify_object",
"role": "assistant"
}
] |
def x(repo_url):
REPO_PREFIX = 'djangodeployer-cache-'
REPO_POSTFIX_UUID = str(uuid.uuid4()).split('-')[-1]
REPO_CACHE_NAME = REPO_PREFIX + REPO_POSTFIX_UUID
REPO_CACHE_LOCATION = '/tmp/%s' % REPO_CACHE_NAME
repo = git.Repo.clone_from(repo_url, REPO_CACHE_LOCATION)
return REPO_CACHE_LOCATION
| clone_git_repo | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(repo_url):\n REPO_PREFIX = 'djangodeployer-cache-'\n REPO_POSTFIX_UUID = str(uuid.uuid4()).split('-')[-1]\n REPO_CACHE_NAME = REPO_PREFIX + REPO_POSTFIX_UUID\n REPO_CACHE_LOCATION = '/tmp/%s' % REPO_CACHE_NAME\n repo = git.Repo.clone_from(repo_url, REPO_CACHE_LOCATION)\n return REPO_CACHE_LOCATION\n",
"role": "user"
},
{
"content": "clone_git_repo",
"role": "assistant"
}
] |
def x(repo_path, ignore_files=[], ignore_folders=[]):
default_ignore_files = ['.gitignore']
default_ignore_folders = ['.git']
ignore_files += default_ignore_files
ignore_folders += default_ignore_folders
filelist = []
for root, folders, files in os.walk(repo_path):
for ignore_file in ignore_files:
if ignore_file in files:
files.remove(ignore_file)
for ignore_folder in ignore_folders:
if ignore_folder in folders:
folders.remove(ignore_folder)
for file_name in files:
filelist.append('%s/%s' % (root, file_name))
return filelist
| get_template_filelist | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(repo_path, ignore_files=[], ignore_folders=[]):\n default_ignore_files = ['.gitignore']\n default_ignore_folders = ['.git']\n ignore_files += default_ignore_files\n ignore_folders += default_ignore_folders\n filelist = []\n for root, folders, files in os.walk(repo_path):\n for ignore_file in ignore_files:\n if ignore_file in files:\n files.remove(ignore_file)\n for ignore_folder in ignore_folders:\n if ignore_folder in folders:\n folders.remove(ignore_folder)\n for file_name in files:\n filelist.append('%s/%s' % (root, file_name))\n return filelist\n",
"role": "user"
},
{
"content": "get_template_filelist",
"role": "assistant"
}
] |
def x(repo_path, to_path, template_params, settings_dir):
TEMPLATE_PROJECT_FOLDER_PLACEHOLDER_NAME = 'deployer_project'
repo_path = repo_path.rstrip('/')
to_path = to_path.rstrip('/')
files_to_render = get_template_filelist(repo_path, ignore_folders=[
TEMPLATE_PROJECT_FOLDER_PLACEHOLDER_NAME])
for single_file_path in files_to_render:
source_file_path = single_file_path
dest_file_path = source_file_path.replace(repo_path, to_path)
render_from_single_file(source_file_path, dest_file_path,
template_params)
settings_template_dir = os.path.join(repo_path,
TEMPLATE_PROJECT_FOLDER_PLACEHOLDER_NAME)
settings_files = get_template_filelist(settings_template_dir)
for single_file_path in settings_files:
source = single_file_path
dest = single_file_path.replace(settings_template_dir, settings_dir)
render_from_single_file(source, dest, template_params)
| render_from_repo | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(repo_path, to_path, template_params, settings_dir):\n TEMPLATE_PROJECT_FOLDER_PLACEHOLDER_NAME = 'deployer_project'\n repo_path = repo_path.rstrip('/')\n to_path = to_path.rstrip('/')\n files_to_render = get_template_filelist(repo_path, ignore_folders=[\n TEMPLATE_PROJECT_FOLDER_PLACEHOLDER_NAME])\n for single_file_path in files_to_render:\n source_file_path = single_file_path\n dest_file_path = source_file_path.replace(repo_path, to_path)\n render_from_single_file(source_file_path, dest_file_path,\n template_params)\n settings_template_dir = os.path.join(repo_path,\n TEMPLATE_PROJECT_FOLDER_PLACEHOLDER_NAME)\n settings_files = get_template_filelist(settings_template_dir)\n for single_file_path in settings_files:\n source = single_file_path\n dest = single_file_path.replace(settings_template_dir, settings_dir)\n render_from_single_file(source, dest, template_params)\n",
"role": "user"
},
{
"content": "render_from_repo",
"role": "assistant"
}
] |
def x(name: str, log_level: int=logging.INFO) ->logging.Logger:
logger = logging.getLogger(name)
logger.setLevel(log_level)
shandler: logging.StreamHandler = logging.StreamHandler(sys.stdout)
fmt: str = '\x1b[1;32m%(levelname)-5s %(module)s:%(funcName)s():'
fmt += '%(lineno)d %(asctime)s\x1b[0m| %(message)s'
shandler.setFormatter(logging.Formatter(fmt))
logger.addHandler(shandler)
return logger
| gen_logger | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(name: str, log_level: int=logging.INFO) ->logging.Logger:\n logger = logging.getLogger(name)\n logger.setLevel(log_level)\n shandler: logging.StreamHandler = logging.StreamHandler(sys.stdout)\n fmt: str = '\\x1b[1;32m%(levelname)-5s %(module)s:%(funcName)s():'\n fmt += '%(lineno)d %(asctime)s\\x1b[0m| %(message)s'\n shandler.setFormatter(logging.Formatter(fmt))\n logger.addHandler(shandler)\n return logger\n",
"role": "user"
},
{
"content": "gen_logger",
"role": "assistant"
}
] |
def x() ->Dict[str, str]:
ua_list: List[str] = [
'Mozilla/5.0 (Windows NT 6.3; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/33.0.1750.117 Safari/537.36'
]
headers: Dict[str, str] = {'User-Agent': ua_list[random.randint(0, len(
ua_list) - 1)]}
return headers
| gen_headers | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x() ->Dict[str, str]:\n ua_list: List[str] = [\n 'Mozilla/5.0 (Windows NT 6.3; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/33.0.1750.117 Safari/537.36'\n ]\n headers: Dict[str, str] = {'User-Agent': ua_list[random.randint(0, len(\n ua_list) - 1)]}\n return headers\n",
"role": "user"
},
{
"content": "gen_headers",
"role": "assistant"
}
] |
def x(results: str, domain: str, fuzzy: bool) ->List[str]:
pattern: Pattern = re.compile('([\\w.-]+@[\\w.-]+)')
hits: List[str] = pattern.findall(results)
if fuzzy:
seed = domain.split('.')[0]
emails: List[str] = [x.lower() for x in hits if x.split('@')[1].
__contains__(seed)]
else:
emails: List[str] = [x.lower() for x in hits if x.endswith(domain)]
return list(set(emails))
| extract_emails | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(results: str, domain: str, fuzzy: bool) ->List[str]:\n pattern: Pattern = re.compile('([\\\\w.-]+@[\\\\w.-]+)')\n hits: List[str] = pattern.findall(results)\n if fuzzy:\n seed = domain.split('.')[0]\n emails: List[str] = [x.lower() for x in hits if x.split('@')[1].\n __contains__(seed)]\n else:\n emails: List[str] = [x.lower() for x in hits if x.endswith(domain)]\n return list(set(emails))\n",
"role": "user"
},
{
"content": "extract_emails",
"role": "assistant"
}
] |
def x(seqs, blast_constructor, blast_db=None, blast_mat_root=None, params={
}, add_seq_names=True, out_filename=None, WorkingDir=None,
SuppressStderr=None, SuppressStdout=None, input_handler=None, HALT_EXEC
=False):
if blast_db:
params['-d'] = blast_db
if out_filename:
params['-o'] = out_filename
ih = input_handler or guess_input_handler(seqs, add_seq_names)
blast_app = blast_constructor(params=params, blast_mat_root=
blast_mat_root, InputHandler=ih, WorkingDir=WorkingDir,
SuppressStderr=SuppressStderr, SuppressStdout=SuppressStdout,
HALT_EXEC=HALT_EXEC)
return blast_app(seqs)
| blast_seqs | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(seqs, blast_constructor, blast_db=None, blast_mat_root=None, params={\n }, add_seq_names=True, out_filename=None, WorkingDir=None,\n SuppressStderr=None, SuppressStdout=None, input_handler=None, HALT_EXEC\n =False):\n if blast_db:\n params['-d'] = blast_db\n if out_filename:\n params['-o'] = out_filename\n ih = input_handler or guess_input_handler(seqs, add_seq_names)\n blast_app = blast_constructor(params=params, blast_mat_root=\n blast_mat_root, InputHandler=ih, WorkingDir=WorkingDir,\n SuppressStderr=SuppressStderr, SuppressStdout=SuppressStdout,\n HALT_EXEC=HALT_EXEC)\n return blast_app(seqs)\n",
"role": "user"
},
{
"content": "blast_seqs",
"role": "assistant"
}
] |
def x(acc_list, blast_db=None, is_protein=None, out_filename=None, params={
}, WorkingDir=tempfile.gettempdir(), SuppressStderr=None,
SuppressStdout=None):
if is_protein is None:
params['-p'] = 'G'
elif is_protein:
params['-p'] = 'T'
else:
params['-p'] = 'F'
if blast_db:
params['-d'] = blast_db
if out_filename:
params['-o'] = out_filename
params['-a'] = 'F'
fasta_cmd = FastaCmd(params=params, InputHandler='_input_as_string',
WorkingDir=WorkingDir, SuppressStderr=SuppressStderr,
SuppressStdout=SuppressStdout)
return fasta_cmd('"%s"' % ','.join(acc_list))
| fasta_cmd_get_seqs | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(acc_list, blast_db=None, is_protein=None, out_filename=None, params={\n }, WorkingDir=tempfile.gettempdir(), SuppressStderr=None,\n SuppressStdout=None):\n if is_protein is None:\n params['-p'] = 'G'\n elif is_protein:\n params['-p'] = 'T'\n else:\n params['-p'] = 'F'\n if blast_db:\n params['-d'] = blast_db\n if out_filename:\n params['-o'] = out_filename\n params['-a'] = 'F'\n fasta_cmd = FastaCmd(params=params, InputHandler='_input_as_string',\n WorkingDir=WorkingDir, SuppressStderr=SuppressStderr,\n SuppressStdout=SuppressStdout)\n return fasta_cmd('\"%s\"' % ','.join(acc_list))\n",
"role": "user"
},
{
"content": "fasta_cmd_get_seqs",
"role": "assistant"
}
] |
def x(acc_list, blast_db, is_protein=True):
fasta_cmd_res = fasta_cmd_get_seqs(acc_list, blast_db=blast_db,
is_protein=is_protein)
recs = FastaCmdFinder(fasta_cmd_res['StdOut'])
result = {}
for rec in recs:
try:
result[rec[0][1:].strip()] = ''.join(map(strip, rec[1:]))
except IndexError:
pass
fasta_cmd_res.cleanUp()
return result
| seqs_from_fastacmd | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(acc_list, blast_db, is_protein=True):\n fasta_cmd_res = fasta_cmd_get_seqs(acc_list, blast_db=blast_db,\n is_protein=is_protein)\n recs = FastaCmdFinder(fasta_cmd_res['StdOut'])\n result = {}\n for rec in recs:\n try:\n result[rec[0][1:].strip()] = ''.join(map(strip, rec[1:]))\n except IndexError:\n pass\n fasta_cmd_res.cleanUp()\n return result\n",
"role": "user"
},
{
"content": "seqs_from_fastacmd",
"role": "assistant"
}
] |
def x(seq, n, max_iterations, app, core_threshold, extra_threshold,
lower_threshold, second_db=None):
app.Parameters['-h'].on(core_threshold)
app.Parameters['-e'].on(core_threshold)
checkpoints = []
ids = []
last_num_ids = None
for i in range(max_iterations):
if checkpoints:
app.Parameters['-R'].on(checkpoints[-1])
curr_check = 'checkpoint_%s.chk' % i
app.Parameters['-C'].on(curr_check)
output = app(seq)
if not access(curr_check, F_OK):
break
checkpoints.append(curr_check)
result = list(output.get('BlastOut', output['StdOut']))
output.cleanUp()
if result:
ids = LastProteinIds9(result, keep_values=True, filter_identity
=False)
num_ids = len(ids)
if num_ids >= n:
break
if num_ids == last_num_ids:
break
last_num_ids = num_ids
if not checkpoints:
return ids
if not second_db and num_ids >= n:
return ids
app.Parameters['-h'].on(extra_threshold)
app.Parameters['-e'].on(lower_threshold)
if second_db:
app.Parameters['-d'].on(second_db)
for i in range(max_iterations):
app.Parameters['-R'].on(checkpoints[-1])
curr_check = 'checkpoint_b_%s.chk' % i
app.Parameters['-C'].on(curr_check)
output = app(seq)
if not access(curr_check, F_OK):
break
checkpoints.append(curr_check)
result = list(output.get('BlastOut', output['StdOut']))
if result:
ids = LastProteinIds9(result, keep_values=True, filter_identity
=False)
num_ids = len(ids)
if num_ids >= n:
break
if num_ids == last_num_ids:
break
last_num_ids = num_ids
for c in checkpoints:
remove(c)
return ids
| ids_from_seq_two_step | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(seq, n, max_iterations, app, core_threshold, extra_threshold,\n lower_threshold, second_db=None):\n app.Parameters['-h'].on(core_threshold)\n app.Parameters['-e'].on(core_threshold)\n checkpoints = []\n ids = []\n last_num_ids = None\n for i in range(max_iterations):\n if checkpoints:\n app.Parameters['-R'].on(checkpoints[-1])\n curr_check = 'checkpoint_%s.chk' % i\n app.Parameters['-C'].on(curr_check)\n output = app(seq)\n if not access(curr_check, F_OK):\n break\n checkpoints.append(curr_check)\n result = list(output.get('BlastOut', output['StdOut']))\n output.cleanUp()\n if result:\n ids = LastProteinIds9(result, keep_values=True, filter_identity\n =False)\n num_ids = len(ids)\n if num_ids >= n:\n break\n if num_ids == last_num_ids:\n break\n last_num_ids = num_ids\n if not checkpoints:\n return ids\n if not second_db and num_ids >= n:\n return ids\n app.Parameters['-h'].on(extra_threshold)\n app.Parameters['-e'].on(lower_threshold)\n if second_db:\n app.Parameters['-d'].on(second_db)\n for i in range(max_iterations):\n app.Parameters['-R'].on(checkpoints[-1])\n curr_check = 'checkpoint_b_%s.chk' % i\n app.Parameters['-C'].on(curr_check)\n output = app(seq)\n if not access(curr_check, F_OK):\n break\n checkpoints.append(curr_check)\n result = list(output.get('BlastOut', output['StdOut']))\n if result:\n ids = LastProteinIds9(result, keep_values=True, filter_identity\n =False)\n num_ids = len(ids)\n if num_ids >= n:\n break\n if num_ids == last_num_ids:\n break\n last_num_ids = num_ids\n for c in checkpoints:\n remove(c)\n return ids\n",
"role": "user"
},
{
"content": "ids_from_seq_two_step",
"role": "assistant"
}
] |
def x(seq, n, max_iterations, app, core_threshold, lower_threshold, step=100):
last_num_ids = None
checkpoints = []
cp_name_base = make_unique_str()
all_ids = {}
try:
i = 0
while 1:
app.Parameters['-h'].on(core_threshold)
app.Parameters['-e'].on(core_threshold)
if core_threshold > lower_threshold:
raise ThresholdFound
if checkpoints:
app.Parameters['-R'].on(checkpoints[-1])
curr_check = 'checkpoint_' + cp_name_base + '_' + str(i) + '.chk'
app.Parameters['-C'].on(curr_check)
output = app(seq)
result = list(output.get('BlastOut', output['StdOut']))
while not result:
output = app(seq)
result = list(output.get('BlastOut', output['StdOut']))
ids = LastProteinIds9(result, keep_values=True, filter_identity
=False)
output.cleanUp()
all_ids[i + 1] = core_threshold, copy(ids)
if not access(curr_check, F_OK):
raise ThresholdFound
checkpoints.append(curr_check)
num_ids = len(ids)
if num_ids >= n:
raise ThresholdFound
last_num_ids = num_ids
core_threshold *= step
if i >= max_iterations - 1:
raise ThresholdFound
i += 1
except ThresholdFound:
for c in checkpoints:
remove(c)
app.Parameters['-R'].off()
return ids, i + 1, all_ids
| ids_from_seq_lower_threshold | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(seq, n, max_iterations, app, core_threshold, lower_threshold, step=100):\n last_num_ids = None\n checkpoints = []\n cp_name_base = make_unique_str()\n all_ids = {}\n try:\n i = 0\n while 1:\n app.Parameters['-h'].on(core_threshold)\n app.Parameters['-e'].on(core_threshold)\n if core_threshold > lower_threshold:\n raise ThresholdFound\n if checkpoints:\n app.Parameters['-R'].on(checkpoints[-1])\n curr_check = 'checkpoint_' + cp_name_base + '_' + str(i) + '.chk'\n app.Parameters['-C'].on(curr_check)\n output = app(seq)\n result = list(output.get('BlastOut', output['StdOut']))\n while not result:\n output = app(seq)\n result = list(output.get('BlastOut', output['StdOut']))\n ids = LastProteinIds9(result, keep_values=True, filter_identity\n =False)\n output.cleanUp()\n all_ids[i + 1] = core_threshold, copy(ids)\n if not access(curr_check, F_OK):\n raise ThresholdFound\n checkpoints.append(curr_check)\n num_ids = len(ids)\n if num_ids >= n:\n raise ThresholdFound\n last_num_ids = num_ids\n core_threshold *= step\n if i >= max_iterations - 1:\n raise ThresholdFound\n i += 1\n except ThresholdFound:\n for c in checkpoints:\n remove(c)\n app.Parameters['-R'].off()\n return ids, i + 1, all_ids\n",
"role": "user"
},
{
"content": "ids_from_seq_lower_threshold",
"role": "assistant"
}
] |
def x(num_chars=20):
chars = 'abcdefghigklmnopqrstuvwxyz'
all_chars = chars + chars.upper() + '01234567890'
picks = list(all_chars)
return ''.join([choice(picks) for i in range(num_chars)])
| make_unique_str | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(num_chars=20):\n chars = 'abcdefghigklmnopqrstuvwxyz'\n all_chars = chars + chars.upper() + '01234567890'\n picks = list(all_chars)\n return ''.join([choice(picks) for i in range(num_chars)])\n",
"role": "user"
},
{
"content": "make_unique_str",
"role": "assistant"
}
] |
def x(checked_ids):
result = checked_ids.keys()
for i in checked_ids.values():
result.extend(i.keys())
return dict.fromkeys(result).keys()
| keep_everything_scorer | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(checked_ids):\n result = checked_ids.keys()\n for i in checked_ids.values():\n result.extend(i.keys())\n return dict.fromkeys(result).keys()\n",
"role": "user"
},
{
"content": "keep_everything_scorer",
"role": "assistant"
}
] |
def x(seqs, app, query_parser, scorer=keep_everything_scorer,
max_iterations=None, blast_db=None, max_seqs=None):
if isinstance(scorer, int):
scorer = make_shotgun_scorer(scorer)
seqs_to_check = list(seqs)
checked_ids = {}
curr_iteration = 0
while seqs_to_check:
unchecked_ids = {}
all_output = app(seqs_to_check)
output = all_output.get('BlastOut', all_output['StdOut'])
for query_id, match_id, match_score in query_parser(output):
if query_id not in checked_ids:
checked_ids[query_id] = {}
checked_ids[query_id][match_id] = match_score
if match_id not in checked_ids:
unchecked_ids[match_id] = True
all_output.cleanUp()
if unchecked_ids:
seq_file = fasta_cmd_get_seqs(unchecked_ids.keys(), app.
Parameters['-d'].Value)['StdOut']
seqs_to_check = []
for s in FastaCmdFinder(fasta_cmd_get_seqs(unchecked_ids.keys(),
app.Parameters['-d'].Value)['StdOut']):
seqs_to_check.extend(s)
else:
seqs_to_check = []
curr_iteration += 1
if max_iterations and curr_iteration >= max_iterations:
break
if max_seqs:
curr = scorer(checked_ids)
if len(curr) >= max_seqs:
return curr
return scorer(checked_ids)
| ids_from_seqs_iterative | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(seqs, app, query_parser, scorer=keep_everything_scorer,\n max_iterations=None, blast_db=None, max_seqs=None):\n if isinstance(scorer, int):\n scorer = make_shotgun_scorer(scorer)\n seqs_to_check = list(seqs)\n checked_ids = {}\n curr_iteration = 0\n while seqs_to_check:\n unchecked_ids = {}\n all_output = app(seqs_to_check)\n output = all_output.get('BlastOut', all_output['StdOut'])\n for query_id, match_id, match_score in query_parser(output):\n if query_id not in checked_ids:\n checked_ids[query_id] = {}\n checked_ids[query_id][match_id] = match_score\n if match_id not in checked_ids:\n unchecked_ids[match_id] = True\n all_output.cleanUp()\n if unchecked_ids:\n seq_file = fasta_cmd_get_seqs(unchecked_ids.keys(), app.\n Parameters['-d'].Value)['StdOut']\n seqs_to_check = []\n for s in FastaCmdFinder(fasta_cmd_get_seqs(unchecked_ids.keys(),\n app.Parameters['-d'].Value)['StdOut']):\n seqs_to_check.extend(s)\n else:\n seqs_to_check = []\n curr_iteration += 1\n if max_iterations and curr_iteration >= max_iterations:\n break\n if max_seqs:\n curr = scorer(checked_ids)\n if len(curr) >= max_seqs:\n return curr\n return scorer(checked_ids)\n",
"role": "user"
},
{
"content": "ids_from_seqs_iterative",
"role": "assistant"
}
] |
def x(seqs, blast_db='nr', e_value='1e-20', max_hits=200, working_dir=
tempfile.gettempdir(), blast_mat_root=None, extra_params={}):
params = {'-M': 'BLOSUM62', '-a': '1', '-e': e_value, '-b': max_hits,
'-v': max_hits, '-p': 'blastp'}
params.update(extra_params)
blast_res = blast_seqs(seqs, Blastall, blast_mat_root=blast_mat_root,
blast_db=blast_db, params=params, add_seq_names=False, WorkingDir=
working_dir)
if blast_res['StdOut']:
lines = [x for x in blast_res['StdOut']]
return BlastResult(lines)
return None
| blastp | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(seqs, blast_db='nr', e_value='1e-20', max_hits=200, working_dir=\n tempfile.gettempdir(), blast_mat_root=None, extra_params={}):\n params = {'-M': 'BLOSUM62', '-a': '1', '-e': e_value, '-b': max_hits,\n '-v': max_hits, '-p': 'blastp'}\n params.update(extra_params)\n blast_res = blast_seqs(seqs, Blastall, blast_mat_root=blast_mat_root,\n blast_db=blast_db, params=params, add_seq_names=False, WorkingDir=\n working_dir)\n if blast_res['StdOut']:\n lines = [x for x in blast_res['StdOut']]\n return BlastResult(lines)\n return None\n",
"role": "user"
},
{
"content": "blastp",
"role": "assistant"
}
] |
def x(fname):
global fh
fh = open(fname, 'w')
if os.name == 'nt':
import win32con
import win32file
import pywintypes
LOCK_EX = win32con.LOCKFILE_EXCLUSIVE_LOCK
LOCK_SH = 0
LOCK_NB = win32con.LOCKFILE_FAIL_IMMEDIATELY
__overlapped = pywintypes.OVERLAPPED()
hfile = win32file._get_osfhandle(fh.fileno())
try:
win32file.LockFileEx(hfile, LOCK_EX | LOCK_NB, 0, -65536,
__overlapped)
except pywintypes.error as exc_value:
if exc_value[0] == 33:
return False
else:
from fcntl import flock, LOCK_EX, LOCK_NB
try:
flock(fh.fileno(), LOCK_EX | LOCK_NB)
except Exception as ex:
return False
fh.write(str(os.getpid()))
fh.flush()
return True
| set_lock | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fname):\n global fh\n fh = open(fname, 'w')\n if os.name == 'nt':\n import win32con\n import win32file\n import pywintypes\n LOCK_EX = win32con.LOCKFILE_EXCLUSIVE_LOCK\n LOCK_SH = 0\n LOCK_NB = win32con.LOCKFILE_FAIL_IMMEDIATELY\n __overlapped = pywintypes.OVERLAPPED()\n hfile = win32file._get_osfhandle(fh.fileno())\n try:\n win32file.LockFileEx(hfile, LOCK_EX | LOCK_NB, 0, -65536,\n __overlapped)\n except pywintypes.error as exc_value:\n if exc_value[0] == 33:\n return False\n else:\n from fcntl import flock, LOCK_EX, LOCK_NB\n try:\n flock(fh.fileno(), LOCK_EX | LOCK_NB)\n except Exception as ex:\n return False\n fh.write(str(os.getpid()))\n fh.flush()\n return True\n",
"role": "user"
},
{
"content": "set_lock",
"role": "assistant"
}
] |
def x(fname):
if not set_lock(fname):
logger.error('File {} is already locked. Terminating.'.format(fname))
sys.exit()
| assert_lock | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fname):\n if not set_lock(fname):\n logger.error('File {} is already locked. Terminating.'.format(fname))\n sys.exit()\n",
"role": "user"
},
{
"content": "assert_lock",
"role": "assistant"
}
] |
def x(fasta_path, is_protein=False, output_dir=None, HALT_EXEC=False):
fasta_dir, fasta_filename = split(fasta_path)
if not output_dir:
output_dir = fasta_dir or '.'
fasta_path = fasta_filename
if not output_dir.endswith('/'):
db_name = output_dir + '/' + fasta_filename
else:
db_name = output_dir + fasta_filename
fdb = FormatDb(WorkingDir=output_dir, HALT_EXEC=HALT_EXEC)
if is_protein:
fdb.Parameters['-p'].on('T')
else:
fdb.Parameters['-p'].on('F')
app_result = fdb(fasta_path)
db_filepaths = []
for v in app_result.values():
try:
db_filepaths.append(v.name)
except AttributeError:
pass
return db_name, db_filepaths
| build_blast_db_from_fasta_path | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fasta_path, is_protein=False, output_dir=None, HALT_EXEC=False):\n fasta_dir, fasta_filename = split(fasta_path)\n if not output_dir:\n output_dir = fasta_dir or '.'\n fasta_path = fasta_filename\n if not output_dir.endswith('/'):\n db_name = output_dir + '/' + fasta_filename\n else:\n db_name = output_dir + fasta_filename\n fdb = FormatDb(WorkingDir=output_dir, HALT_EXEC=HALT_EXEC)\n if is_protein:\n fdb.Parameters['-p'].on('T')\n else:\n fdb.Parameters['-p'].on('F')\n app_result = fdb(fasta_path)\n db_filepaths = []\n for v in app_result.values():\n try:\n db_filepaths.append(v.name)\n except AttributeError:\n pass\n return db_name, db_filepaths\n",
"role": "user"
},
{
"content": "build_blast_db_from_fasta_path",
"role": "assistant"
}
] |
def x(fasta_file, is_protein=False, output_dir=None, HALT_EXEC=False):
output_dir = output_dir or '.'
_, fasta_path = mkstemp(dir=output_dir, prefix='BLAST_temp_db_', suffix
='.fasta')
fasta_f = open(fasta_path, 'w')
for line in fasta_file:
fasta_f.write('%s\n' % line.strip())
fasta_f.close()
blast_db, db_filepaths = build_blast_db_from_fasta_path(fasta_path,
is_protein=is_protein, output_dir=None, HALT_EXEC=HALT_EXEC)
db_filepaths.append(fasta_path)
return blast_db, db_filepaths
| build_blast_db_from_fasta_file | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fasta_file, is_protein=False, output_dir=None, HALT_EXEC=False):\n output_dir = output_dir or '.'\n _, fasta_path = mkstemp(dir=output_dir, prefix='BLAST_temp_db_', suffix\n ='.fasta')\n fasta_f = open(fasta_path, 'w')\n for line in fasta_file:\n fasta_f.write('%s\\n' % line.strip())\n fasta_f.close()\n blast_db, db_filepaths = build_blast_db_from_fasta_path(fasta_path,\n is_protein=is_protein, output_dir=None, HALT_EXEC=HALT_EXEC)\n db_filepaths.append(fasta_path)\n return blast_db, db_filepaths\n",
"role": "user"
},
{
"content": "build_blast_db_from_fasta_file",
"role": "assistant"
}
] |
def x(seqs, is_protein=False, output_dir='./', HALT_EXEC=False):
_, tmp_fasta_filepath = mkstemp(prefix='Blast_tmp_db', suffix='.fasta')
tmp_fasta_file = open(tmp_fasta_filepath, 'w')
tmp_fasta_file.write(seqs.toFasta())
tmp_fasta_file.close()
db_name, db_filepaths = build_blast_db_from_fasta_path(tmp_fasta_filepath,
is_protein=is_protein, output_dir=output_dir, HALT_EXEC=HALT_EXEC)
remove(tmp_fasta_filepath)
return db_name, db_filepaths
| build_blast_db_from_seqs | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(seqs, is_protein=False, output_dir='./', HALT_EXEC=False):\n _, tmp_fasta_filepath = mkstemp(prefix='Blast_tmp_db', suffix='.fasta')\n tmp_fasta_file = open(tmp_fasta_filepath, 'w')\n tmp_fasta_file.write(seqs.toFasta())\n tmp_fasta_file.close()\n db_name, db_filepaths = build_blast_db_from_fasta_path(tmp_fasta_filepath,\n is_protein=is_protein, output_dir=output_dir, HALT_EXEC=HALT_EXEC)\n remove(tmp_fasta_filepath)\n return db_name, db_filepaths\n",
"role": "user"
},
{
"content": "build_blast_db_from_seqs",
"role": "assistant"
}
] |
def x():
usage = 'usage: %prog [options] fasta_filepath'
version = 'Version: %prog 0.1'
parser = OptionParser(usage=usage, version=version)
parser.add_option('-p', '--is_protein', action='store_true', dest=
'is_protein', default=False, help=
'Pass if building db of protein sequences [default: False, nucleotide db]'
)
parser.add_option('-o', '--output_dir', action='store', type='string',
dest='output_dir', default=None, help=
'the output directory [default: directory containing input fasta_filepath]'
)
opts, args = parser.parse_args()
num_args = 1
if len(args) != num_args:
parser.error('Must provide single filepath to build database from.')
return opts, args
| parse_command_line_parameters | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n usage = 'usage: %prog [options] fasta_filepath'\n version = 'Version: %prog 0.1'\n parser = OptionParser(usage=usage, version=version)\n parser.add_option('-p', '--is_protein', action='store_true', dest=\n 'is_protein', default=False, help=\n 'Pass if building db of protein sequences [default: False, nucleotide db]'\n )\n parser.add_option('-o', '--output_dir', action='store', type='string',\n dest='output_dir', default=None, help=\n 'the output directory [default: directory containing input fasta_filepath]'\n )\n opts, args = parser.parse_args()\n num_args = 1\n if len(args) != num_args:\n parser.error('Must provide single filepath to build database from.')\n return opts, args\n",
"role": "user"
},
{
"content": "parse_command_line_parameters",
"role": "assistant"
}
] |
def x(self, data):
self.Parameters['-i'].on(data)
input_filepath = self.Parameters['-i'].Value
input_file_dir, input_filename = split(input_filepath)
input_file_base, input_file_ext = splitext(input_filename)
self.Parameters['-l'].on(FilePath('%s.log') % input_filename)
self.Parameters['-n'].on(FilePath(input_filename))
return ''
| input_as_parameter | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, data):\n self.Parameters['-i'].on(data)\n input_filepath = self.Parameters['-i'].Value\n input_file_dir, input_filename = split(input_filepath)\n input_file_base, input_file_ext = splitext(input_filename)\n self.Parameters['-l'].on(FilePath('%s.log') % input_filename)\n self.Parameters['-n'].on(FilePath(input_filename))\n return ''\n",
"role": "user"
},
{
"content": "input_as_parameter",
"role": "assistant"
}
] |
def x(self, data):
wd = self.WorkingDir
db_name = self.Parameters['-n'].Value
log_name = self.Parameters['-l'].Value
result = {}
result['log'] = ResultPath(Path=wd + log_name, IsWritten=True)
if self.Parameters['-p'].Value == 'F':
extensions = ['nhr', 'nin', 'nsq', 'nsd', 'nsi']
else:
extensions = ['phr', 'pin', 'psq', 'psd', 'psi']
for extension in extensions:
for file_path in glob(wd + (db_name + '*' + extension)):
key = file_path.split(db_name + '.')[1]
result_path = ResultPath(Path=file_path, IsWritten=True)
result[key] = result_path
return result
| get_result_paths | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, data):\n wd = self.WorkingDir\n db_name = self.Parameters['-n'].Value\n log_name = self.Parameters['-l'].Value\n result = {}\n result['log'] = ResultPath(Path=wd + log_name, IsWritten=True)\n if self.Parameters['-p'].Value == 'F':\n extensions = ['nhr', 'nin', 'nsq', 'nsd', 'nsi']\n else:\n extensions = ['phr', 'pin', 'psq', 'psd', 'psi']\n for extension in extensions:\n for file_path in glob(wd + (db_name + '*' + extension)):\n key = file_path.split(db_name + '.')[1]\n result_path = ResultPath(Path=file_path, IsWritten=True)\n result[key] = result_path\n return result\n",
"role": "user"
},
{
"content": "get_result_paths",
"role": "assistant"
}
] |
def x(self):
if not hasattr(self, '_cached_session_key'):
session_id_bytes = self.get_secure_cookie('session_id')
session_id = None
if session_id_bytes:
try:
session_id = session_id_bytes.decode('utf-8')
except:
pass
if not session_id:
session_id = oz.redis_sessions.random_hex(20)
session_time = oz.settings['session_time']
kwargs = dict(name='session_id', value=session_id.encode('utf-8'),
domain=oz.settings.get('cookie_domain'), httponly=True)
if session_time:
kwargs['expires_days'] = round(session_time / 60 / 60 / 24)
self.set_secure_cookie(**kwargs)
password_salt = oz.settings['session_salt']
self._cached_session_key = ('session:%s:v4' % oz.redis_sessions.
password_hash(session_id, password_salt=password_salt))
return self._cached_session_key
| session_key | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if not hasattr(self, '_cached_session_key'):\n session_id_bytes = self.get_secure_cookie('session_id')\n session_id = None\n if session_id_bytes:\n try:\n session_id = session_id_bytes.decode('utf-8')\n except:\n pass\n if not session_id:\n session_id = oz.redis_sessions.random_hex(20)\n session_time = oz.settings['session_time']\n kwargs = dict(name='session_id', value=session_id.encode('utf-8'),\n domain=oz.settings.get('cookie_domain'), httponly=True)\n if session_time:\n kwargs['expires_days'] = round(session_time / 60 / 60 / 24)\n self.set_secure_cookie(**kwargs)\n password_salt = oz.settings['session_salt']\n self._cached_session_key = ('session:%s:v4' % oz.redis_sessions.\n password_hash(session_id, password_salt=password_salt))\n return self._cached_session_key\n",
"role": "user"
},
{
"content": "session_key",
"role": "assistant"
}
] |
def x(self):
session_time = oz.settings['session_time']
if session_time:
self.redis().expire(self._session_key, session_time)
| update_session_expiration | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n session_time = oz.settings['session_time']\n if session_time:\n self.redis().expire(self._session_key, session_time)\n",
"role": "user"
},
{
"content": "update_session_expiration",
"role": "assistant"
}
] |
def x(self, name, default=None):
value = self.redis().hget(self._session_key, name) or default
self._update_session_expiration()
return value
| get_session_value | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, name, default=None):\n value = self.redis().hget(self._session_key, name) or default\n self._update_session_expiration()\n return value\n",
"role": "user"
},
{
"content": "get_session_value",
"role": "assistant"
}
] |
def x(self, name, value):
self.redis().hset(self._session_key, name, value)
self._update_session_expiration()
| set_session_value | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, name, value):\n self.redis().hset(self._session_key, name, value)\n self._update_session_expiration()\n",
"role": "user"
},
{
"content": "set_session_value",
"role": "assistant"
}
] |
def x(self, name):
self.redis().hdel(self._session_key, name)
self._update_session_expiration()
| clear_session_value | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, name):\n self.redis().hdel(self._session_key, name)\n self._update_session_expiration()\n",
"role": "user"
},
{
"content": "clear_session_value",
"role": "assistant"
}
] |
async def rewrite_middleware(server, request):
"""
Sanic middleware that utilizes a security class's "rewrite" method to
check
"""
if singletons.settings.SECURITY is not None:
security_class = singletons.settings.load('SECURITY')
else:
security_class = DummySecurity
security = security_class()
try:
new_path = await security.rewrite(request)
except SecurityException as e:
msg = ''
if DEBUG:
msg = str(e)
return server.response.text(msg, status=400)
request.path = new_path
| rewrite_middleware | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def rewrite_middleware(server, request):\n \"\"\"\n Sanic middleware that utilizes a security class's \"rewrite\" method to\n check\n \"\"\"\n if singletons.settings.SECURITY is not None:\n security_class = singletons.settings.load('SECURITY')\n else:\n security_class = DummySecurity\n security = security_class()\n try:\n new_path = await security.rewrite(request)\n except SecurityException as e:\n msg = ''\n if DEBUG:\n msg = str(e)\n return server.response.text(msg, status=400)\n request.path = new_path\n",
"role": "user"
},
{
"content": "rewrite_middleware",
"role": "assistant"
}
] |
def x(self, c, details):
if c in [Settings.SUCCESS, Settings.CREATED, Settings.ACCEPTED]:
return details
elif c == Settings.BAD_REQUEST:
raise ErrAtlasBadRequest(c, details)
elif c == Settings.UNAUTHORIZED:
raise ErrAtlasUnauthorized(c, details)
elif c == Settings.FORBIDDEN:
raise ErrAtlasForbidden(c, details)
elif c == Settings.NOTFOUND:
raise ErrAtlasNotFound(c, details)
elif c == Settings.METHOD_NOT_ALLOWED:
raise ErrAtlasMethodNotAllowed(c, details)
elif c == Settings.CONFLICT:
raise ErrAtlasConflict(c, details)
else:
raise ErrAtlasServerErrors(c, details)
| answer | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, c, details):\n if c in [Settings.SUCCESS, Settings.CREATED, Settings.ACCEPTED]:\n return details\n elif c == Settings.BAD_REQUEST:\n raise ErrAtlasBadRequest(c, details)\n elif c == Settings.UNAUTHORIZED:\n raise ErrAtlasUnauthorized(c, details)\n elif c == Settings.FORBIDDEN:\n raise ErrAtlasForbidden(c, details)\n elif c == Settings.NOTFOUND:\n raise ErrAtlasNotFound(c, details)\n elif c == Settings.METHOD_NOT_ALLOWED:\n raise ErrAtlasMethodNotAllowed(c, details)\n elif c == Settings.CONFLICT:\n raise ErrAtlasConflict(c, details)\n else:\n raise ErrAtlasServerErrors(c, details)\n",
"role": "user"
},
{
"content": "answer",
"role": "assistant"
}
] |
def x(self, uri):
r = None
try:
r = requests.get(uri, allow_redirects=True, timeout=Settings.
requests_timeout, headers={}, auth=HTTPDigestAuth(self.user,
self.password))
return self.answer(r.status_code, r.json())
except:
raise
finally:
if r:
r.connection.close()
| get | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, uri):\n r = None\n try:\n r = requests.get(uri, allow_redirects=True, timeout=Settings.\n requests_timeout, headers={}, auth=HTTPDigestAuth(self.user,\n self.password))\n return self.answer(r.status_code, r.json())\n except:\n raise\n finally:\n if r:\n r.connection.close()\n",
"role": "user"
},
{
"content": "get",
"role": "assistant"
}
] |
def x(self, uri, pull=True, **kwargs):
self.repo = repo = git_clone(uri, pull=pull, reflect=True)
cmd = 'git rev-list --all'
output = sys_call(cmd, cwd=repo.path)
repo_shas = set(x.strip() for x in output.split('\n') if x)
logger.debug('Total Commits: %s' % len(repo_shas))
cmd = 'git --no-pager log --all --format=sha:%H --numstat'
output = sys_call(cmd)
all_logs = re.sub('\n+', '\n', output)
c_logs = [x for x in [s.strip() for s in all_logs.split('sha:')] if x]
_end = None
objs = []
for c_log in c_logs:
sha, s, all_changes = c_log.partition('\n')
c = repo.get_object(sha)
_start = ts2dt(c.commit_time)
obj = dict(_oid=sha, _start=_start, _end=_end, repo_uri=uri, tree=c
.tree, parents=c.parents, author=c.author, committer=c.
committer, author_time=c.author_time, message=c.message,
mergetag=c.mergetag, extra=c.extra)
for _file in all_changes.split('\n'):
_file = _file.strip()
obj.setdefault('files', {})
if not _file:
added, removed, fname = 0, 0, None
else:
added, removed, fname = _file.split('\t')
added = 0 if added == '-' else int(added)
removed = 0 if removed == '-' else int(removed)
changes = {'added': added, 'removed': removed}
obj['files'][fname] = changes
obj['added'] = sum([v.get('added', 0) for v in obj['files'].
itervalues()])
obj['removed'] = sum([v.get('removed', 0) for v in obj['files'].
itervalues()])
obj['acked_by'] = acked_by_re.findall(c.message)
obj['signed_off_by'] = signed_off_by_re.findall(c.message)
obj['resolves'] = resolves_re.findall(c.message)
obj['related'] = related_re.findall(c.message)
objs.append(obj)
self.objects.extend(objs)
return super(Commit, self).get_objects(**kwargs)
| get_objects | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, uri, pull=True, **kwargs):\n self.repo = repo = git_clone(uri, pull=pull, reflect=True)\n cmd = 'git rev-list --all'\n output = sys_call(cmd, cwd=repo.path)\n repo_shas = set(x.strip() for x in output.split('\\n') if x)\n logger.debug('Total Commits: %s' % len(repo_shas))\n cmd = 'git --no-pager log --all --format=sha:%H --numstat'\n output = sys_call(cmd)\n all_logs = re.sub('\\n+', '\\n', output)\n c_logs = [x for x in [s.strip() for s in all_logs.split('sha:')] if x]\n _end = None\n objs = []\n for c_log in c_logs:\n sha, s, all_changes = c_log.partition('\\n')\n c = repo.get_object(sha)\n _start = ts2dt(c.commit_time)\n obj = dict(_oid=sha, _start=_start, _end=_end, repo_uri=uri, tree=c\n .tree, parents=c.parents, author=c.author, committer=c.\n committer, author_time=c.author_time, message=c.message,\n mergetag=c.mergetag, extra=c.extra)\n for _file in all_changes.split('\\n'):\n _file = _file.strip()\n obj.setdefault('files', {})\n if not _file:\n added, removed, fname = 0, 0, None\n else:\n added, removed, fname = _file.split('\\t')\n added = 0 if added == '-' else int(added)\n removed = 0 if removed == '-' else int(removed)\n changes = {'added': added, 'removed': removed}\n obj['files'][fname] = changes\n obj['added'] = sum([v.get('added', 0) for v in obj['files'].\n itervalues()])\n obj['removed'] = sum([v.get('removed', 0) for v in obj['files'].\n itervalues()])\n obj['acked_by'] = acked_by_re.findall(c.message)\n obj['signed_off_by'] = signed_off_by_re.findall(c.message)\n obj['resolves'] = resolves_re.findall(c.message)\n obj['related'] = related_re.findall(c.message)\n objs.append(obj)\n self.objects.extend(objs)\n return super(Commit, self).get_objects(**kwargs)\n",
"role": "user"
},
{
"content": "get_objects",
"role": "assistant"
}
] |
def x(dictionary):
for _, value in sorted(dictionary.iteritems(), key=operator.itemgetter(0)):
yield value
| dict_values_sorted_by_key | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(dictionary):\n for _, value in sorted(dictionary.iteritems(), key=operator.itemgetter(0)):\n yield value\n",
"role": "user"
},
{
"content": "dict_values_sorted_by_key",
"role": "assistant"
}
] |
def x(f):
name = f.__name__
def x(self, *args, **kwargs):
if not args and not kwargs:
if hasattr(self, '_%s' % name):
return getattr(self, '_%s' % name)
a = f(self, *args, **kwargs)
setattr(self, '_%s' % name, a)
return a
else:
return f(self, *args, **kwargs)
func.__name__ = name
return func
| ondemand | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(f):\n name = f.__name__\n\n def x(self, *args, **kwargs):\n if not args and not kwargs:\n if hasattr(self, '_%s' % name):\n return getattr(self, '_%s' % name)\n a = f(self, *args, **kwargs)\n setattr(self, '_%s' % name, a)\n return a\n else:\n return f(self, *args, **kwargs)\n func.__name__ = name\n return func\n",
"role": "user"
},
{
"content": "ondemand",
"role": "assistant"
}
] |
def x(self, start, size):
LOG.debug('get_memory: %d bytes from %x', size, start)
return get_memory(self.ida.idaapi, start, size, default_byte=self.
default_byte)
| get_memory | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, start, size):\n LOG.debug('get_memory: %d bytes from %x', size, start)\n return get_memory(self.ida.idaapi, start, size, default_byte=self.\n default_byte)\n",
"role": "user"
},
{
"content": "get_memory",
"role": "assistant"
}
] |
def x(self, **kwargs):
fmt = ':::'.join('%%{%s}' % f for f in self._fields)
if self.ssh_host:
output = self._ssh_cmd(fmt)
else:
output = self._local_cmd(fmt)
if isinstance(output, basestring):
output = unicode(output, 'utf-8')
output = output.strip().split('\n')
lines = [l.strip().split(':::') for l in output]
now = utcnow()
host = self.ssh_host or socket.gethostname()
for line in lines:
obj = {'host': host, '_start': now}
for i, item in enumerate(line):
if item == '(none)':
item = None
obj[self._fields[i]] = item
obj['_oid'] = '%s__%s' % (host, obj['nvra'])
self.objects.add(obj)
return super(Rpm, self).get_objects(**kwargs)
| get_objects | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, **kwargs):\n fmt = ':::'.join('%%{%s}' % f for f in self._fields)\n if self.ssh_host:\n output = self._ssh_cmd(fmt)\n else:\n output = self._local_cmd(fmt)\n if isinstance(output, basestring):\n output = unicode(output, 'utf-8')\n output = output.strip().split('\\n')\n lines = [l.strip().split(':::') for l in output]\n now = utcnow()\n host = self.ssh_host or socket.gethostname()\n for line in lines:\n obj = {'host': host, '_start': now}\n for i, item in enumerate(line):\n if item == '(none)':\n item = None\n obj[self._fields[i]] = item\n obj['_oid'] = '%s__%s' % (host, obj['nvra'])\n self.objects.add(obj)\n return super(Rpm, self).get_objects(**kwargs)\n",
"role": "user"
},
{
"content": "get_objects",
"role": "assistant"
}
] |
def x(request, exception):
error = format_error(title='Resource not found', detail=str(exception))
return json(return_an_error(error), status=HTTPStatus.NOT_FOUND)
| handle_404 | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(request, exception):\n error = format_error(title='Resource not found', detail=str(exception))\n return json(return_an_error(error), status=HTTPStatus.NOT_FOUND)\n",
"role": "user"
},
{
"content": "handle_404",
"role": "assistant"
}
] |
def x(t, container, name, value):
if name in container:
raise Exception("%s '%s' already exists" % (t, name))
else:
container[name] = value
| add_to_dict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(t, container, name, value):\n if name in container:\n raise Exception(\"%s '%s' already exists\" % (t, name))\n else:\n container[name] = value\n",
"role": "user"
},
{
"content": "add_to_dict",
"role": "assistant"
}
] |
def x(config=None):
if config == None:
config = {}
config_source = None
try:
with open(os.environ.get('OZ_CONFIG', 'config.py')) as f:
config_source = f.read()
except Exception:
tornado.log.gen_log.info('Could not read config.py', exc_info=True)
if config_source != None:
tornado.util.exec_in(config_source, config, config)
for p in config.get('plugins', ['oz.core']):
plugin(p)
for key, value in config.get('app_options', {}).items():
setattr(tornado.options.options, key, value)
global settings
settings = tornado.options.options.as_dict()
settings['ui_modules'] = _uimodules
settings['project_name'] = config.get('project_name')
| initialize | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(config=None):\n if config == None:\n config = {}\n config_source = None\n try:\n with open(os.environ.get('OZ_CONFIG', 'config.py')) as f:\n config_source = f.read()\n except Exception:\n tornado.log.gen_log.info('Could not read config.py', exc_info=True)\n if config_source != None:\n tornado.util.exec_in(config_source, config, config)\n for p in config.get('plugins', ['oz.core']):\n plugin(p)\n for key, value in config.get('app_options', {}).items():\n setattr(tornado.options.options, key, value)\n global settings\n settings = tornado.options.options.as_dict()\n settings['ui_modules'] = _uimodules\n settings['project_name'] = config.get('project_name')\n",
"role": "user"
},
{
"content": "initialize",
"role": "assistant"
}
] |
def x(self, name, *args, **kwargs):
listeners = self._triggers.get(name, [])
for listener in listeners:
result = listener(*args, **kwargs)
if result == break_trigger:
return False
return True
| trigger | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, name, *args, **kwargs):\n listeners = self._triggers.get(name, [])\n for listener in listeners:\n result = listener(*args, **kwargs)\n if result == break_trigger:\n return False\n return True\n",
"role": "user"
},
{
"content": "trigger",
"role": "assistant"
}
] |
def x(self, subdir=None):
if subdir is not None:
dirname = self.cache_path
if subdir:
dirname = os.path.join(dirname, subdir)
else:
dirname = os.path.dirname(self.cache_path)
os.makedirs(dirname, exist_ok=True)
| cache_makedirs | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, subdir=None):\n if subdir is not None:\n dirname = self.cache_path\n if subdir:\n dirname = os.path.join(dirname, subdir)\n else:\n dirname = os.path.dirname(self.cache_path)\n os.makedirs(dirname, exist_ok=True)\n",
"role": "user"
},
{
"content": "cache_makedirs",
"role": "assistant"
}
] |
def x(connection_string=None):
global _session_makers
connection_string = connection_string or oz.settings['db']
if not connection_string in _session_makers:
_session_makers[connection_string] = sessionmaker(bind=engine(
connection_string=connection_string))
return _session_makers[connection_string]()
| session | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(connection_string=None):\n global _session_makers\n connection_string = connection_string or oz.settings['db']\n if not connection_string in _session_makers:\n _session_makers[connection_string] = sessionmaker(bind=engine(\n connection_string=connection_string))\n return _session_makers[connection_string]()\n",
"role": "user"
},
{
"content": "session",
"role": "assistant"
}
] |
def x(query_fasta_fp, database_fasta_fp, output_fp, params=None):
if params is None:
params = {}
if '-out' not in params:
params['-out'] = 'blast9'
blat = Blat(params=params)
result = blat([query_fasta_fp, database_fasta_fp, output_fp])
return result['output']
| assign_reads_to_database | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(query_fasta_fp, database_fasta_fp, output_fp, params=None):\n if params is None:\n params = {}\n if '-out' not in params:\n params['-out'] = 'blast9'\n blat = Blat(params=params)\n result = blat([query_fasta_fp, database_fasta_fp, output_fp])\n return result['output']\n",
"role": "user"
},
{
"content": "assign_reads_to_database",
"role": "assistant"
}
] |
def x(query_fasta_fp, database_fasta_fp, output_fp, params=None):
if params is None:
params = {}
my_params = {'-t': 'dna', '-q': 'dna'}
if '-t' in params or '-q' in params:
raise ApplicationError(
'Cannot change database or query types when ' +
'using assign_dna_reads_to_dna_database. ' +
"""Use assign_reads_to_database instead.
""")
my_params.update(params)
result = assign_reads_to_database(query_fasta_fp, database_fasta_fp,
output_fp, my_params)
return result
| assign_dna_reads_to_dna_database | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(query_fasta_fp, database_fasta_fp, output_fp, params=None):\n if params is None:\n params = {}\n my_params = {'-t': 'dna', '-q': 'dna'}\n if '-t' in params or '-q' in params:\n raise ApplicationError(\n 'Cannot change database or query types when ' +\n 'using assign_dna_reads_to_dna_database. ' +\n \"\"\"Use assign_reads_to_database instead.\n\"\"\")\n my_params.update(params)\n result = assign_reads_to_database(query_fasta_fp, database_fasta_fp,\n output_fp, my_params)\n return result\n",
"role": "user"
},
{
"content": "assign_dna_reads_to_dna_database",
"role": "assistant"
}
] |
def x(query_fasta_fp, database_fasta_fp, output_fp, temp_dir='/tmp', params
=None):
if params is None:
params = {}
my_params = {'-t': 'prot', '-q': 'prot'}
if not isabs(temp_dir):
raise ApplicationError('temp_dir must be an absolute path.')
if '-t' in params or '-q' in params:
raise ApplicationError(
'Cannot change database or query types when using assign_dna_reads_to_dna_database. Use assign_reads_to_database instead.'
)
if 'genetic_code' in params:
my_genetic_code = GeneticCodes[params['genetic_code']]
del params['genetic_code']
else:
my_genetic_code = GeneticCodes[1]
my_params.update(params)
_, tmp = mkstemp(dir=temp_dir)
tmp_out = open(tmp, 'w')
for label, sequence in parse_fasta(open(query_fasta_fp)):
seq_id = label.split()[0]
s = DNA.makeSequence(sequence)
translations = my_genetic_code.sixframes(s)
frames = [1, 2, 3, -1, -2, -3]
translations = dict(zip(frames, translations))
for frame, translation in sorted(translations.iteritems()):
entry = '>{seq_id}_frame_{frame}\n{trans}\n'
entry = entry.format(seq_id=seq_id, frame=frame, trans=translation)
tmp_out.write(entry)
tmp_out.close()
result = assign_reads_to_database(tmp, database_fasta_fp, output_fp,
params=my_params)
remove(tmp)
return result
| assign_dna_reads_to_protein_database | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(query_fasta_fp, database_fasta_fp, output_fp, temp_dir='/tmp', params\n =None):\n if params is None:\n params = {}\n my_params = {'-t': 'prot', '-q': 'prot'}\n if not isabs(temp_dir):\n raise ApplicationError('temp_dir must be an absolute path.')\n if '-t' in params or '-q' in params:\n raise ApplicationError(\n 'Cannot change database or query types when using assign_dna_reads_to_dna_database. Use assign_reads_to_database instead.'\n )\n if 'genetic_code' in params:\n my_genetic_code = GeneticCodes[params['genetic_code']]\n del params['genetic_code']\n else:\n my_genetic_code = GeneticCodes[1]\n my_params.update(params)\n _, tmp = mkstemp(dir=temp_dir)\n tmp_out = open(tmp, 'w')\n for label, sequence in parse_fasta(open(query_fasta_fp)):\n seq_id = label.split()[0]\n s = DNA.makeSequence(sequence)\n translations = my_genetic_code.sixframes(s)\n frames = [1, 2, 3, -1, -2, -3]\n translations = dict(zip(frames, translations))\n for frame, translation in sorted(translations.iteritems()):\n entry = '>{seq_id}_frame_{frame}\\n{trans}\\n'\n entry = entry.format(seq_id=seq_id, frame=frame, trans=translation)\n tmp_out.write(entry)\n tmp_out.close()\n result = assign_reads_to_database(tmp, database_fasta_fp, output_fp,\n params=my_params)\n remove(tmp)\n return result\n",
"role": "user"
},
{
"content": "assign_dna_reads_to_protein_database",
"role": "assistant"
}
] |
def x(self):
command_parts = []
cd_command = ''.join(['cd ', str(self.WorkingDir), ';'])
if self._command is None:
raise ApplicationError('_command has not been set.')
command = self._command
parameters = sorted([str(x) for x in self.Parameters.values() if str(x)])
synonyms = self._synonyms
command_parts.append(cd_command)
command_parts.append(command)
command_parts.append(self._database)
command_parts.append(self._query)
command_parts += parameters
if self._output:
command_parts.append(self._output.Path)
return self._command_delimiter.join(filter(None, command_parts)).strip()
| get_base_command | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n command_parts = []\n cd_command = ''.join(['cd ', str(self.WorkingDir), ';'])\n if self._command is None:\n raise ApplicationError('_command has not been set.')\n command = self._command\n parameters = sorted([str(x) for x in self.Parameters.values() if str(x)])\n synonyms = self._synonyms\n command_parts.append(cd_command)\n command_parts.append(command)\n command_parts.append(self._database)\n command_parts.append(self._query)\n command_parts += parameters\n if self._output:\n command_parts.append(self._output.Path)\n return self._command_delimiter.join(filter(None, command_parts)).strip()\n",
"role": "user"
},
{
"content": "get_base_command",
"role": "assistant"
}
] |
def x(self, data):
query, database, output = data
if not isabs(database) or not isabs(query) or not isabs(output):
raise ApplicationError('Only absolute paths allowed.\n%s' % ', '.
join(data))
self._database = FilePath(database)
self._query = FilePath(query)
self._output = ResultPath(output, IsWritten=True)
if self.Parameters['-t'].isOn() and self.Parameters['-q'].isOn() and (self
.Parameters['-t'].Value, self.Parameters['-q'].Value
) not in self._valid_combinations:
error_message = ('Invalid combination of database and query ' +
"types ('%s', '%s').\n" % (self.Paramters['-t'].Value, self.
Parameters['-q'].Value))
error_message += 'Must be one of: %s\n' % repr(self._valid_combinations
)
raise ApplicationError(error_message)
if self.Parameters['-t'].isOn() and self.Parameters['-t'
].Value not in self._database_types:
error_message = 'Invalid database type %s\n' % self.Parameters['-t'
].Value
error_message += 'Allowed values: %s\n' % ', '.join(self.
_database_types)
raise ApplicationError(error_message)
if self.Parameters['-q'].isOn() and self.Parameters['-q'
].Value not in self._query_types:
error_message = 'Invalid query type %s\n' % self.Parameters['-q'].Value
error_message += 'Allowed values: %s\n' % ', '.join(self._query_types)
raise ApplicationError(error_message)
if self.Parameters['-mask'].isOn() and self.Parameters['-mask'
].Value not in self._mask_types:
error_message = 'Invalid mask type %s\n' % self.Parameters['-mask']
error_message += 'Allowed Values: %s\n' % ', '.join(self._mask_types)
raise ApplicationError(error_message)
if self.Parameters['-qMask'].isOn() and self.Parameters['-qMask'
].Value not in self._mask_types:
error_message = 'Invalid qMask type %s\n' % self.Parameters['-qMask'
].Value
error_message += 'Allowed values: %s\n' % ', '.join(self._mask_types)
raise ApplicationError(error_message)
if self.Parameters['-repeats'].isOn() and self.Parameters['-repeats'
].Value not in self._mask_types:
error_message = 'Invalid repeat type %s\n' % self.Parameters['-repeat'
].Value
error_message += 'Allowed values: %s\n' % ', '.join(self._mask_types)
raise ApplicationError(error_message)
if self.Parameters['-out'].isOn() and self.Parameters['-out'
].Value not in self._out_types:
error_message = 'Invalid output type %s\n' % self.Parameters['-out']
error_message += 'Allowed values: %s\n' % ', '.join(self._out_types)
raise ApplicationError(error_message)
return ''
| input_as_list | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, data):\n query, database, output = data\n if not isabs(database) or not isabs(query) or not isabs(output):\n raise ApplicationError('Only absolute paths allowed.\\n%s' % ', '.\n join(data))\n self._database = FilePath(database)\n self._query = FilePath(query)\n self._output = ResultPath(output, IsWritten=True)\n if self.Parameters['-t'].isOn() and self.Parameters['-q'].isOn() and (self\n .Parameters['-t'].Value, self.Parameters['-q'].Value\n ) not in self._valid_combinations:\n error_message = ('Invalid combination of database and query ' + \n \"types ('%s', '%s').\\n\" % (self.Paramters['-t'].Value, self.\n Parameters['-q'].Value))\n error_message += 'Must be one of: %s\\n' % repr(self._valid_combinations\n )\n raise ApplicationError(error_message)\n if self.Parameters['-t'].isOn() and self.Parameters['-t'\n ].Value not in self._database_types:\n error_message = 'Invalid database type %s\\n' % self.Parameters['-t'\n ].Value\n error_message += 'Allowed values: %s\\n' % ', '.join(self.\n _database_types)\n raise ApplicationError(error_message)\n if self.Parameters['-q'].isOn() and self.Parameters['-q'\n ].Value not in self._query_types:\n error_message = 'Invalid query type %s\\n' % self.Parameters['-q'].Value\n error_message += 'Allowed values: %s\\n' % ', '.join(self._query_types)\n raise ApplicationError(error_message)\n if self.Parameters['-mask'].isOn() and self.Parameters['-mask'\n ].Value not in self._mask_types:\n error_message = 'Invalid mask type %s\\n' % self.Parameters['-mask']\n error_message += 'Allowed Values: %s\\n' % ', '.join(self._mask_types)\n raise ApplicationError(error_message)\n if self.Parameters['-qMask'].isOn() and self.Parameters['-qMask'\n ].Value not in self._mask_types:\n error_message = 'Invalid qMask type %s\\n' % self.Parameters['-qMask'\n ].Value\n error_message += 'Allowed values: %s\\n' % ', '.join(self._mask_types)\n raise ApplicationError(error_message)\n if self.Parameters['-repeats'].isOn() and self.Parameters['-repeats'\n ].Value not in self._mask_types:\n error_message = 'Invalid repeat type %s\\n' % self.Parameters['-repeat'\n ].Value\n error_message += 'Allowed values: %s\\n' % ', '.join(self._mask_types)\n raise ApplicationError(error_message)\n if self.Parameters['-out'].isOn() and self.Parameters['-out'\n ].Value not in self._out_types:\n error_message = 'Invalid output type %s\\n' % self.Parameters['-out']\n error_message += 'Allowed values: %s\\n' % ', '.join(self._out_types)\n raise ApplicationError(error_message)\n return ''\n",
"role": "user"
},
{
"content": "input_as_list",
"role": "assistant"
}
] |
def x(var, d, lat, lon, depth, cfg):
if lon < 0:
lon = lon + 360
url = cfg['url']
doy = int(d.strftime('%j'))
dataset = open_url(url)
dn = np.abs(doy - dataset['time'][:]).argmin()
xn = np.abs(lon - dataset['lon'][:]).argmin()
yn = np.abs(lat - dataset['lat'][:]).argmin()
if re.match('temperature\\d?$', var):
mn = ma.masked_values(dataset.t_mn.t_mn[dn, :, yn, xn].reshape(
dataset['depth'].shape[0]), dataset.t_mn.attributes['_FillValue'])
sd = ma.masked_values(dataset.t_sd.t_sd[dn, :, yn, xn].reshape(
dataset['depth'].shape[0]), dataset.t_sd.attributes['_FillValue'])
elif re.match('salinity\\d?$', var):
mn = ma.masked_values(dataset.s_mn.s_mn[dn, :, yn, xn].reshape(
dataset['depth'].shape[0]), dataset.s_mn.attributes['_FillValue'])
sd = ma.masked_values(dataset.s_sd.s_sd[dn, :, yn, xn].reshape(
dataset['depth'].shape[0]), dataset.s_sd.attributes['_FillValue'])
zwoa = ma.array(dataset.depth[:])
ind = (depth <= zwoa.max()) & (depth >= zwoa.min())
f = interp1d(zwoa[~ma.getmaskarray(mn)].compressed(), mn.compressed())
mn_interp = ma.masked_all(depth.shape)
mn_interp[ind] = f(depth[ind])
f = interp1d(zwoa[~ma.getmaskarray(sd)].compressed(), sd.compressed())
sd_interp = ma.masked_all(depth.shape)
sd_interp[ind] = f(depth[ind])
output = {'woa_an': mn_interp, 'woa_sd': sd_interp}
return output
| woa_profile_from_dap | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(var, d, lat, lon, depth, cfg):\n if lon < 0:\n lon = lon + 360\n url = cfg['url']\n doy = int(d.strftime('%j'))\n dataset = open_url(url)\n dn = np.abs(doy - dataset['time'][:]).argmin()\n xn = np.abs(lon - dataset['lon'][:]).argmin()\n yn = np.abs(lat - dataset['lat'][:]).argmin()\n if re.match('temperature\\\\d?$', var):\n mn = ma.masked_values(dataset.t_mn.t_mn[dn, :, yn, xn].reshape(\n dataset['depth'].shape[0]), dataset.t_mn.attributes['_FillValue'])\n sd = ma.masked_values(dataset.t_sd.t_sd[dn, :, yn, xn].reshape(\n dataset['depth'].shape[0]), dataset.t_sd.attributes['_FillValue'])\n elif re.match('salinity\\\\d?$', var):\n mn = ma.masked_values(dataset.s_mn.s_mn[dn, :, yn, xn].reshape(\n dataset['depth'].shape[0]), dataset.s_mn.attributes['_FillValue'])\n sd = ma.masked_values(dataset.s_sd.s_sd[dn, :, yn, xn].reshape(\n dataset['depth'].shape[0]), dataset.s_sd.attributes['_FillValue'])\n zwoa = ma.array(dataset.depth[:])\n ind = (depth <= zwoa.max()) & (depth >= zwoa.min())\n f = interp1d(zwoa[~ma.getmaskarray(mn)].compressed(), mn.compressed())\n mn_interp = ma.masked_all(depth.shape)\n mn_interp[ind] = f(depth[ind])\n f = interp1d(zwoa[~ma.getmaskarray(sd)].compressed(), sd.compressed())\n sd_interp = ma.masked_all(depth.shape)\n sd_interp[ind] = f(depth[ind])\n output = {'woa_an': mn_interp, 'woa_sd': sd_interp}\n return output\n",
"role": "user"
},
{
"content": "woa_profile_from_dap",
"role": "assistant"
}
] |
def x(d, lat, lon, filename, varnames=None):
d = np.asanyarray(d)
lat = np.asanyarray(lat)
lon = np.asanyarray(lon)
lon[lon < 0] += 360
doy = np.array([int(dd.strftime('%j')) for dd in d])
nc = netCDF4.Dataset(expanduser(filename), 'r')
if varnames is None:
varnames = {}
for v in nc.variables.keys():
if nc.variables[v].dimensions == (u'time', u'depth', u'lat', u'lon'
):
varnames[v] = v
output = {}
for v in varnames:
output[v] = []
for d_n, lat_n, lon_n in zip(doy, lat, lon):
n_d = np.abs(d_n - nc.variables['time'][:]).argmin()
n_x = np.abs(lon_n - nc.variables['lon'][:]).argmin()
n_y = np.abs(lat_n - nc.variables['lat'][:]).argmin()
for v in varnames:
output[v].append(nc.variables[varnames[v]][n_d, 0, n_y, n_x])
for v in varnames:
output[v] = ma.fix_invalid(output[v])
return output
| woa_track_from_file | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(d, lat, lon, filename, varnames=None):\n d = np.asanyarray(d)\n lat = np.asanyarray(lat)\n lon = np.asanyarray(lon)\n lon[lon < 0] += 360\n doy = np.array([int(dd.strftime('%j')) for dd in d])\n nc = netCDF4.Dataset(expanduser(filename), 'r')\n if varnames is None:\n varnames = {}\n for v in nc.variables.keys():\n if nc.variables[v].dimensions == (u'time', u'depth', u'lat', u'lon'\n ):\n varnames[v] = v\n output = {}\n for v in varnames:\n output[v] = []\n for d_n, lat_n, lon_n in zip(doy, lat, lon):\n n_d = np.abs(d_n - nc.variables['time'][:]).argmin()\n n_x = np.abs(lon_n - nc.variables['lon'][:]).argmin()\n n_y = np.abs(lat_n - nc.variables['lat'][:]).argmin()\n for v in varnames:\n output[v].append(nc.variables[varnames[v]][n_d, 0, n_y, n_x])\n for v in varnames:\n output[v] = ma.fix_invalid(output[v])\n return output\n",
"role": "user"
},
{
"content": "woa_track_from_file",
"role": "assistant"
}
] |
def x(self, doy, depth, lat, lon, var):
dims, idx = cropIndices(self.dims, lat, lon, depth, doy)
subset = {}
for v in var:
subset[v] = ma.asanyarray([self.ncs[tnn][v][0, idx['zn'], idx['yn'],
idx['xn']] for tnn in idx['tn']])
return subset, dims
| crop | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, doy, depth, lat, lon, var):\n dims, idx = cropIndices(self.dims, lat, lon, depth, doy)\n subset = {}\n for v in var:\n subset[v] = ma.asanyarray([self.ncs[tnn][v][0, idx['zn'], idx['yn'],\n idx['xn']] for tnn in idx['tn']])\n return subset, dims\n",
"role": "user"
},
{
"content": "crop",
"role": "assistant"
}
] |
def x(self, doy, depth, lat, lon, var):
subset, dims = self.crop(doy, depth, lat, lon, var)
if np.all([(d in dims['time']) for d in doy]) & np.all([(z in dims[
'depth']) for z in depth]) & np.all([(y in dims['lat']) for y in lat]
) & np.all([(x in dims['lon']) for x in lon]):
dn = np.nonzero([(d in doy) for d in dims['time']])[0]
zn = np.nonzero([(z in depth) for z in dims['depth']])[0]
yn = np.nonzero([(y in lat) for y in dims['lat']])[0]
xn = np.nonzero([(x in lon) for x in dims['lon']])[0]
output = {}
for v in subset:
output[v] = subset[v][:, :, :, xn][:, :, yn][:, zn][dn]
return output
output = {}
for v in var:
output[v] = ma.masked_all((doy.size, depth.size, lat.size, lon.size
), dtype=subset[v].dtype)
if subset[v].dtype in ['int32']:
subset[v] = subset[v].astype('f')
subset[v][ma.getmaskarray(subset[v])] = np.nan
subset[v] = subset[v].data
if not (doy == dims['time']).all():
for v in subset.keys():
f = interp1d(dims['time'], subset[v], axis=0)
subset[v] = f(doy)
dims['time'] = np.atleast_1d(doy)
if not (np.all(lat == dims['lat']) and np.all(lon == dims['lon'])):
points_out = []
for latn in lat:
for lonn in lon:
points_out.append([latn, lonn])
points_out = np.array(points_out)
for v in subset:
tmp = np.nan * np.ones((doy.size, dims['depth'].size, lat.size,
lon.size), dtype=subset[v].dtype)
for nt in range(doy.size):
for nz in range(dims['depth'].size):
data = subset[v][nt, nz]
idx = np.nonzero(~np.isnan(data))
if idx[0].size > 0:
points = np.array([dims['lat'][idx[0]], dims['lon']
[idx[1]]]).T
values = data[idx]
try:
values_out = griddata(np.atleast_1d(np.squeeze(
points)), values, np.atleast_1d(np.squeeze(
points_out)))
except:
values_out = []
for p in points_out:
try:
values_out.append(griddata(np.
atleast_1d(np.squeeze(points)),
values, np.atleast_1d(np.squeeze(p))))
except:
values_out.append(np.nan)
values_out = np.array(values_out)
idx = np.isfinite(values_out)
for [y, x], out in zip(points_out[idx], values_out[idx]
):
tmp[nt, nz, y == lat, x == lon] = out
subset[v] = tmp
same_depth = np.shape(depth) == dims['depth'].shape and np.allclose(depth,
dims['depth'])
if not same_depth:
for v in list(subset.keys()):
try:
f = interp1d(dims['depth'], subset[v], axis=1, bounds_error
=False)
subset[v] = f(np.array(depth))
except:
print("Fail to interpolate '%s' in depth" % v)
del subset[v]
for v in subset:
if output[v].dtype in ['int32']:
subset[v] = np.round(subset[v])
output[v][:] = ma.fix_invalid(subset[v][:])
return output
| interpolate | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, doy, depth, lat, lon, var):\n subset, dims = self.crop(doy, depth, lat, lon, var)\n if np.all([(d in dims['time']) for d in doy]) & np.all([(z in dims[\n 'depth']) for z in depth]) & np.all([(y in dims['lat']) for y in lat]\n ) & np.all([(x in dims['lon']) for x in lon]):\n dn = np.nonzero([(d in doy) for d in dims['time']])[0]\n zn = np.nonzero([(z in depth) for z in dims['depth']])[0]\n yn = np.nonzero([(y in lat) for y in dims['lat']])[0]\n xn = np.nonzero([(x in lon) for x in dims['lon']])[0]\n output = {}\n for v in subset:\n output[v] = subset[v][:, :, :, xn][:, :, yn][:, zn][dn]\n return output\n output = {}\n for v in var:\n output[v] = ma.masked_all((doy.size, depth.size, lat.size, lon.size\n ), dtype=subset[v].dtype)\n if subset[v].dtype in ['int32']:\n subset[v] = subset[v].astype('f')\n subset[v][ma.getmaskarray(subset[v])] = np.nan\n subset[v] = subset[v].data\n if not (doy == dims['time']).all():\n for v in subset.keys():\n f = interp1d(dims['time'], subset[v], axis=0)\n subset[v] = f(doy)\n dims['time'] = np.atleast_1d(doy)\n if not (np.all(lat == dims['lat']) and np.all(lon == dims['lon'])):\n points_out = []\n for latn in lat:\n for lonn in lon:\n points_out.append([latn, lonn])\n points_out = np.array(points_out)\n for v in subset:\n tmp = np.nan * np.ones((doy.size, dims['depth'].size, lat.size,\n lon.size), dtype=subset[v].dtype)\n for nt in range(doy.size):\n for nz in range(dims['depth'].size):\n data = subset[v][nt, nz]\n idx = np.nonzero(~np.isnan(data))\n if idx[0].size > 0:\n points = np.array([dims['lat'][idx[0]], dims['lon']\n [idx[1]]]).T\n values = data[idx]\n try:\n values_out = griddata(np.atleast_1d(np.squeeze(\n points)), values, np.atleast_1d(np.squeeze(\n points_out)))\n except:\n values_out = []\n for p in points_out:\n try:\n values_out.append(griddata(np.\n atleast_1d(np.squeeze(points)),\n values, np.atleast_1d(np.squeeze(p))))\n except:\n values_out.append(np.nan)\n values_out = np.array(values_out)\n idx = np.isfinite(values_out)\n for [y, x], out in zip(points_out[idx], values_out[idx]\n ):\n tmp[nt, nz, y == lat, x == lon] = out\n subset[v] = tmp\n same_depth = np.shape(depth) == dims['depth'].shape and np.allclose(depth,\n dims['depth'])\n if not same_depth:\n for v in list(subset.keys()):\n try:\n f = interp1d(dims['depth'], subset[v], axis=1, bounds_error\n =False)\n subset[v] = f(np.array(depth))\n except:\n print(\"Fail to interpolate '%s' in depth\" % v)\n del subset[v]\n for v in subset:\n if output[v].dtype in ['int32']:\n subset[v] = np.round(subset[v])\n output[v][:] = ma.fix_invalid(subset[v][:])\n return output\n",
"role": "user"
},
{
"content": "interpolate",
"role": "assistant"
}
] |
def x(cls, path=None, objtype=None, query=None, **kwargs):
if not objtype:
objtype = cls.OBJTYPE
if not path:
path = cls.PATH
if not query:
query = kwargs
return objtype.db(path).find(query).count()
| count | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, path=None, objtype=None, query=None, **kwargs):\n if not objtype:\n objtype = cls.OBJTYPE\n if not path:\n path = cls.PATH\n if not query:\n query = kwargs\n return objtype.db(path).find(query).count()\n",
"role": "user"
},
{
"content": "count",
"role": "assistant"
}
] |
def x(cls, path=None, objtype=None, query=None, page_size=1000, **kwargs):
if not objtype:
objtype = cls.OBJTYPE
if not path:
path = cls.PATH
db = objtype.db(path)
if not query:
query = kwargs
results = list(db.find(query).sort(ID_KEY, ASCENDING).limit(page_size))
while results:
page = [objtype(path=path, _new_object=result) for result in results]
for obj in page:
yield obj
query[ID_KEY] = {GT: results[-1][ID_KEY]}
results = list(db.find(query).sort(ID_KEY, ASCENDING).limit(page_size))
| iterator | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, path=None, objtype=None, query=None, page_size=1000, **kwargs):\n if not objtype:\n objtype = cls.OBJTYPE\n if not path:\n path = cls.PATH\n db = objtype.db(path)\n if not query:\n query = kwargs\n results = list(db.find(query).sort(ID_KEY, ASCENDING).limit(page_size))\n while results:\n page = [objtype(path=path, _new_object=result) for result in results]\n for obj in page:\n yield obj\n query[ID_KEY] = {GT: results[-1][ID_KEY]}\n results = list(db.find(query).sort(ID_KEY, ASCENDING).limit(page_size))\n",
"role": "user"
},
{
"content": "iterator",
"role": "assistant"
}
] |
def x(self, data, **kwargs):
obj = self.OBJTYPE.create(data, path=self.PATH, **kwargs)
self.append(obj)
return obj
| insert | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, data, **kwargs):\n obj = self.OBJTYPE.create(data, path=self.PATH, **kwargs)\n self.append(obj)\n return obj\n",
"role": "user"
},
{
"content": "insert",
"role": "assistant"
}
] |
def x(self, new_path):
for elt in self:
DatabaseObject.create(elt, path=new_path)
for elt in self:
elt._collection.remove({ID_KEY: elt[ID_KEY]})
| move | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, new_path):\n for elt in self:\n DatabaseObject.create(elt, path=new_path)\n for elt in self:\n elt._collection.remove({ID_KEY: elt[ID_KEY]})\n",
"role": "user"
},
{
"content": "move",
"role": "assistant"
}
] |
def x(seqs, moltype=DNA_cogent, params={}):
seqs = LoadSeqs(data=seqs, moltype=moltype, aligned=False)
try:
s1, s2 = seqs.values()
except ValueError:
raise ValueError('Pairwise aligning of seqs requires exactly two seqs.'
)
try:
gap_open = params['gap_open']
except KeyError:
gap_open = 5
try:
gap_extend = params['gap_extend']
except KeyError:
gap_extend = 2
try:
score_matrix = params['score_matrix']
except KeyError:
score_matrix = make_dna_scoring_dict(match=1, transition=-1,
transversion=-1)
return local_pairwise(s1, s2, score_matrix, gap_open, gap_extend)
| pair_hmm_align_unaligned_seqs | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(seqs, moltype=DNA_cogent, params={}):\n seqs = LoadSeqs(data=seqs, moltype=moltype, aligned=False)\n try:\n s1, s2 = seqs.values()\n except ValueError:\n raise ValueError('Pairwise aligning of seqs requires exactly two seqs.'\n )\n try:\n gap_open = params['gap_open']\n except KeyError:\n gap_open = 5\n try:\n gap_extend = params['gap_extend']\n except KeyError:\n gap_extend = 2\n try:\n score_matrix = params['score_matrix']\n except KeyError:\n score_matrix = make_dna_scoring_dict(match=1, transition=-1,\n transversion=-1)\n return local_pairwise(s1, s2, score_matrix, gap_open, gap_extend)\n",
"role": "user"
},
{
"content": "pair_hmm_align_unaligned_seqs",
"role": "assistant"
}
] |
async def viewers_js(request):
"""
Viewers determines the viewers installed based on settings, then uses the
conversion infrastructure to convert all these JS files into a single JS
bundle, that is then served. As with media, it will simply serve a cached
version if necessary.
"""
response = singletons.server.response
viewers_resource = singletons.viewers.get_resource()
url_string = viewers_resource.url_string
target_ts = TypeString('min.js')
target_resource = TypedResource(url_string, target_ts)
if target_resource.cache_exists():
return await response.file(target_resource.cache_path, headers={
'Content-Type': 'application/javascript'})
if not viewers_resource.cache_exists():
viewers_resource.save()
await singletons.workers.async_enqueue_sync(enqueue_conversion_path,
url_string, str(target_ts), singletons.workers.enqueue_convert)
return response.text(NOT_LOADED_JS, headers={'Content-Type':
'application/javascript'})
| viewers_js | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "async def viewers_js(request):\n \"\"\"\n Viewers determines the viewers installed based on settings, then uses the\n conversion infrastructure to convert all these JS files into a single JS\n bundle, that is then served. As with media, it will simply serve a cached\n version if necessary.\n \"\"\"\n response = singletons.server.response\n viewers_resource = singletons.viewers.get_resource()\n url_string = viewers_resource.url_string\n target_ts = TypeString('min.js')\n target_resource = TypedResource(url_string, target_ts)\n if target_resource.cache_exists():\n return await response.file(target_resource.cache_path, headers={\n 'Content-Type': 'application/javascript'})\n if not viewers_resource.cache_exists():\n viewers_resource.save()\n await singletons.workers.async_enqueue_sync(enqueue_conversion_path,\n url_string, str(target_ts), singletons.workers.enqueue_convert)\n return response.text(NOT_LOADED_JS, headers={'Content-Type':\n 'application/javascript'})\n",
"role": "user"
},
{
"content": "viewers_js",
"role": "assistant"
}
] |
def x(self, uri, _oid=None, _start=None, _end=None, load_kwargs=None, **kwargs
):
load_kwargs = load_kwargs or {}
objects = load(path=uri, filetype='csv', **load_kwargs)
k = itertools.count(1)
now = utcnow()
__oid = lambda o: k.next()
_oid = _oid or __oid
_start = _start or now
_end = _end or None
def x(v):
_v = type(v)
_ = True if _v is type or hasattr(v, '__call__') else False
return _
for obj in objects:
obj['_oid'] = _oid(obj) if is_callable(_oid) else _oid
obj['_start'] = _start(obj) if is_callable(_start) else _start
obj['_end'] = _end(obj) if is_callable(_end) else _end
self.container.add(obj)
return super(Rows, self).get_objects(**kwargs)
| get_objects | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, uri, _oid=None, _start=None, _end=None, load_kwargs=None, **kwargs\n ):\n load_kwargs = load_kwargs or {}\n objects = load(path=uri, filetype='csv', **load_kwargs)\n k = itertools.count(1)\n now = utcnow()\n __oid = lambda o: k.next()\n _oid = _oid or __oid\n _start = _start or now\n _end = _end or None\n\n def x(v):\n _v = type(v)\n _ = True if _v is type or hasattr(v, '__call__') else False\n return _\n for obj in objects:\n obj['_oid'] = _oid(obj) if is_callable(_oid) else _oid\n obj['_start'] = _start(obj) if is_callable(_start) else _start\n obj['_end'] = _end(obj) if is_callable(_end) else _end\n self.container.add(obj)\n return super(Rows, self).get_objects(**kwargs)\n",
"role": "user"
},
{
"content": "get_objects",
"role": "assistant"
}
] |
def x(self):
content = {'databaseName': self.databaseName, 'roles': self.roles,
'username': self.username, 'password': self.password}
return content
| getSpecs | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n content = {'databaseName': self.databaseName, 'roles': self.roles,\n 'username': self.username, 'password': self.password}\n return content\n",
"role": "user"
},
{
"content": "getSpecs",
"role": "assistant"
}
] |
def x(self, databaseName, roleNames, collectionName=None):
for roleName in roleNames:
self.add_role(databaseName, roleName, collectionName)
| add_roles | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, databaseName, roleNames, collectionName=None):\n for roleName in roleNames:\n self.add_role(databaseName, roleName, collectionName)\n",
"role": "user"
},
{
"content": "add_roles",
"role": "assistant"
}
] |
def x(self, databaseName, roleName, collectionName=None):
role = {'databaseName': databaseName, 'roleName': roleName}
if collectionName:
role['collectionName'] = collectionName
if collectionName and roleName not in [RoleSpecs.read, RoleSpecs.readWrite
]:
raise ErrRole('Permissions [%s] not available for a collection' %
roleName)
elif not collectionName and roleName not in [RoleSpecs.read, RoleSpecs.
readWrite, RoleSpecs.dbAdmin] and databaseName != 'admin':
raise ErrRole(
'Permissions [%s] is only available for admin database' % roleName)
if role not in self.roles:
self.roles.append(role)
| add_role | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, databaseName, roleName, collectionName=None):\n role = {'databaseName': databaseName, 'roleName': roleName}\n if collectionName:\n role['collectionName'] = collectionName\n if collectionName and roleName not in [RoleSpecs.read, RoleSpecs.readWrite\n ]:\n raise ErrRole('Permissions [%s] not available for a collection' %\n roleName)\n elif not collectionName and roleName not in [RoleSpecs.read, RoleSpecs.\n readWrite, RoleSpecs.dbAdmin] and databaseName != 'admin':\n raise ErrRole(\n 'Permissions [%s] is only available for admin database' % roleName)\n if role not in self.roles:\n self.roles.append(role)\n",
"role": "user"
},
{
"content": "add_role",
"role": "assistant"
}
] |
def x(self, databaseName, roleNames, collectionName=None):
for roleName in roleNames:
self.remove_role(databaseName, roleName, collectionName)
| remove_roles | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, databaseName, roleNames, collectionName=None):\n for roleName in roleNames:\n self.remove_role(databaseName, roleName, collectionName)\n",
"role": "user"
},
{
"content": "remove_roles",
"role": "assistant"
}
] |
def x(self, databaseName, roleName, collectionName=None):
role = {'databaseName': databaseName, 'roleName': roleName}
if collectionName:
role['collectionName'] = collectionName
if role in self.roles:
self.roles.remove(role)
| remove_role | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, databaseName, roleName, collectionName=None):\n role = {'databaseName': databaseName, 'roleName': roleName}\n if collectionName:\n role['collectionName'] = collectionName\n if role in self.roles:\n self.roles.remove(role)\n",
"role": "user"
},
{
"content": "remove_role",
"role": "assistant"
}
] |
def x(self):
content = {}
if len(self.roles) != 0:
content['roles'] = self.roles
if self.password:
content['password'] = self.password
return content
| getSpecs | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n content = {}\n if len(self.roles) != 0:\n content['roles'] = self.roles\n if self.password:\n content['password'] = self.password\n return content\n",
"role": "user"
},
{
"content": "getSpecs",
"role": "assistant"
}
] |
def x(seq_path=None, result_path=None, shortest_len=True, similarity=0.97,
threads=1, exact=False, HALT_EXEC=False):
if seq_path is None or not exists(seq_path):
raise ValueError(
'Error: FASTA query sequence filepath is mandatory input.')
if result_path is None or not isdir(dirname(realpath(result_path))):
raise ValueError('Error: output directory is mandatory input.')
sumaclust = Sumaclust(HALT_EXEC=HALT_EXEC)
sumaclust.Parameters['-O'].on(result_path)
if similarity is not None:
sumaclust.Parameters['-t'].on(similarity)
if exact:
sumaclust.Parameters['-e'].on()
if not shortest_len:
sumaclust.Parameters['-l'].off()
if threads > 0:
sumaclust.Parameters['-p'].on(threads)
else:
raise ValueError('Number of threads must be positive.')
app_result = sumaclust(seq_path)
f_otumap = app_result['OtuMap']
clusters = [line.strip().split('\t')[1:] for line in f_otumap]
return clusters
| sumaclust_denovo_cluster | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(seq_path=None, result_path=None, shortest_len=True, similarity=0.97,\n threads=1, exact=False, HALT_EXEC=False):\n if seq_path is None or not exists(seq_path):\n raise ValueError(\n 'Error: FASTA query sequence filepath is mandatory input.')\n if result_path is None or not isdir(dirname(realpath(result_path))):\n raise ValueError('Error: output directory is mandatory input.')\n sumaclust = Sumaclust(HALT_EXEC=HALT_EXEC)\n sumaclust.Parameters['-O'].on(result_path)\n if similarity is not None:\n sumaclust.Parameters['-t'].on(similarity)\n if exact:\n sumaclust.Parameters['-e'].on()\n if not shortest_len:\n sumaclust.Parameters['-l'].off()\n if threads > 0:\n sumaclust.Parameters['-p'].on(threads)\n else:\n raise ValueError('Number of threads must be positive.')\n app_result = sumaclust(seq_path)\n f_otumap = app_result['OtuMap']\n clusters = [line.strip().split('\\t')[1:] for line in f_otumap]\n return clusters\n",
"role": "user"
},
{
"content": "sumaclust_denovo_cluster",
"role": "assistant"
}
] |
def x(self, data):
result = {}
result['OtuMap'] = ResultPath(Path=self.Parameters['-O'].Value,
IsWritten=True)
if not isfile(result['OtuMap'].Path):
otumap_f = open(result['OtuMap'].Path, 'w')
otumap_f.close()
return result
| get_result_paths | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, data):\n result = {}\n result['OtuMap'] = ResultPath(Path=self.Parameters['-O'].Value,\n IsWritten=True)\n if not isfile(result['OtuMap'].Path):\n otumap_f = open(result['OtuMap'].Path, 'w')\n otumap_f.close()\n return result\n",
"role": "user"
},
{
"content": "get_result_paths",
"role": "assistant"
}
] |
def x(fasta_filepath, output_filepath, output_uc=False, working_dir=None,
strand='both', maxuniquesize=None, minuniquesize=None, sizein=False,
sizeout=True, log_name='derep.log', HALT_EXEC=False):
if not working_dir:
working_dir = dirname(abspath(output_filepath))
app = Vsearch(WorkingDir=working_dir, HALT_EXEC=HALT_EXEC)
log_filepath = join(working_dir, log_name)
uc_filepath = None
if output_uc:
root_name = splitext(abspath(output_filepath))[0]
uc_filepath = join(working_dir, '%s.uc' % root_name)
app.Parameters['--uc'].on(uc_filepath)
if maxuniquesize:
app.Parameters['--maxuniquesize'].on(maxuniquesize)
if minuniquesize:
app.Parameters['--minuniquesize'].on(minuniquesize)
if sizein:
app.Parameters['--sizein'].on()
if sizeout:
app.Parameters['--sizeout'].on()
if strand == 'both' or strand == 'plus':
app.Parameters['--strand'].on(strand)
else:
raise ValueError("Option --strand accepts only 'both'or 'plus' values")
app.Parameters['--derep_fulllength'].on(fasta_filepath)
app.Parameters['--output'].on(output_filepath)
app.Parameters['--log'].on(log_filepath)
app_result = app()
return output_filepath, uc_filepath, log_filepath
| vsearch_dereplicate_exact_seqs | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fasta_filepath, output_filepath, output_uc=False, working_dir=None,\n strand='both', maxuniquesize=None, minuniquesize=None, sizein=False,\n sizeout=True, log_name='derep.log', HALT_EXEC=False):\n if not working_dir:\n working_dir = dirname(abspath(output_filepath))\n app = Vsearch(WorkingDir=working_dir, HALT_EXEC=HALT_EXEC)\n log_filepath = join(working_dir, log_name)\n uc_filepath = None\n if output_uc:\n root_name = splitext(abspath(output_filepath))[0]\n uc_filepath = join(working_dir, '%s.uc' % root_name)\n app.Parameters['--uc'].on(uc_filepath)\n if maxuniquesize:\n app.Parameters['--maxuniquesize'].on(maxuniquesize)\n if minuniquesize:\n app.Parameters['--minuniquesize'].on(minuniquesize)\n if sizein:\n app.Parameters['--sizein'].on()\n if sizeout:\n app.Parameters['--sizeout'].on()\n if strand == 'both' or strand == 'plus':\n app.Parameters['--strand'].on(strand)\n else:\n raise ValueError(\"Option --strand accepts only 'both'or 'plus' values\")\n app.Parameters['--derep_fulllength'].on(fasta_filepath)\n app.Parameters['--output'].on(output_filepath)\n app.Parameters['--log'].on(log_filepath)\n app_result = app()\n return output_filepath, uc_filepath, log_filepath\n",
"role": "user"
},
{
"content": "vsearch_dereplicate_exact_seqs",
"role": "assistant"
}
] |
def x(fasta_filepath, output_filepath, working_dir=None, minsize=None,
maxsize=None, log_name='abundance_sort.log', HALT_EXEC=False):
if not working_dir:
working_dir = dirname(output_filepath)
app = Vsearch(WorkingDir=working_dir, HALT_EXEC=HALT_EXEC)
log_filepath = join(working_dir, log_name)
if minsize:
app.Parameters['--minsize'].on(minsize)
if maxsize:
app.Parameters['--maxsize'].on(maxsize)
app.Parameters['--sortbysize'].on(fasta_filepath)
app.Parameters['--output'].on(output_filepath)
app.Parameters['--log'].on(log_filepath)
app_result = app()
return output_filepath, log_filepath
| vsearch_sort_by_abundance | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fasta_filepath, output_filepath, working_dir=None, minsize=None,\n maxsize=None, log_name='abundance_sort.log', HALT_EXEC=False):\n if not working_dir:\n working_dir = dirname(output_filepath)\n app = Vsearch(WorkingDir=working_dir, HALT_EXEC=HALT_EXEC)\n log_filepath = join(working_dir, log_name)\n if minsize:\n app.Parameters['--minsize'].on(minsize)\n if maxsize:\n app.Parameters['--maxsize'].on(maxsize)\n app.Parameters['--sortbysize'].on(fasta_filepath)\n app.Parameters['--output'].on(output_filepath)\n app.Parameters['--log'].on(log_filepath)\n app_result = app()\n return output_filepath, log_filepath\n",
"role": "user"
},
{
"content": "vsearch_sort_by_abundance",
"role": "assistant"
}
] |
def x(fasta_filepath, working_dir, output_chimeras=True, output_nonchimeras
=True, output_alns=False, output_tabular=False, log_name=
'vsearch_uchime_de_novo_chimera_filtering.log', HALT_EXEC=False):
app = Vsearch(WorkingDir=working_dir, HALT_EXEC=HALT_EXEC)
if not (output_chimeras or output_nonchimeras or output_alns or
output_tabular):
raise ValueError(
'At least one output format (output_chimeras,output_nonchimeras, output_alns, output_tabular)must be selected'
)
output_chimera_filepath = None
output_non_chimera_filepath = None
output_alns_filepath = None
output_tabular_filepath = None
if output_chimeras:
output_chimera_filepath = join(working_dir, 'uchime_chimeras.fasta')
app.Parameters['--chimeras'].on(output_chimera_filepath)
if output_nonchimeras:
output_non_chimera_filepath = join(working_dir,
'uchime_non_chimeras.fasta')
app.Parameters['--nonchimeras'].on(output_non_chimera_filepath)
if output_alns:
output_alns_filepath = join(working_dir, 'uchime_alignments.txt')
app.Parameters['--uchimealns'].on(output_alns_filepath)
if output_tabular:
output_tabular_filepath = join(working_dir, 'uchime_tabular.txt')
app.Parameters['--uchimeout'].on(output_tabular_filepath)
log_filepath = join(working_dir, log_name)
app.Parameters['--uchime_denovo'].on(fasta_filepath)
app.Parameters['--log'].on(log_filepath)
app_result = app()
return (output_chimera_filepath, output_non_chimera_filepath,
output_alns_filepath, output_tabular_filepath, log_filepath)
| vsearch_chimera_filter_de_novo | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fasta_filepath, working_dir, output_chimeras=True, output_nonchimeras\n =True, output_alns=False, output_tabular=False, log_name=\n 'vsearch_uchime_de_novo_chimera_filtering.log', HALT_EXEC=False):\n app = Vsearch(WorkingDir=working_dir, HALT_EXEC=HALT_EXEC)\n if not (output_chimeras or output_nonchimeras or output_alns or\n output_tabular):\n raise ValueError(\n 'At least one output format (output_chimeras,output_nonchimeras, output_alns, output_tabular)must be selected'\n )\n output_chimera_filepath = None\n output_non_chimera_filepath = None\n output_alns_filepath = None\n output_tabular_filepath = None\n if output_chimeras:\n output_chimera_filepath = join(working_dir, 'uchime_chimeras.fasta')\n app.Parameters['--chimeras'].on(output_chimera_filepath)\n if output_nonchimeras:\n output_non_chimera_filepath = join(working_dir,\n 'uchime_non_chimeras.fasta')\n app.Parameters['--nonchimeras'].on(output_non_chimera_filepath)\n if output_alns:\n output_alns_filepath = join(working_dir, 'uchime_alignments.txt')\n app.Parameters['--uchimealns'].on(output_alns_filepath)\n if output_tabular:\n output_tabular_filepath = join(working_dir, 'uchime_tabular.txt')\n app.Parameters['--uchimeout'].on(output_tabular_filepath)\n log_filepath = join(working_dir, log_name)\n app.Parameters['--uchime_denovo'].on(fasta_filepath)\n app.Parameters['--log'].on(log_filepath)\n app_result = app()\n return (output_chimera_filepath, output_non_chimera_filepath,\n output_alns_filepath, output_tabular_filepath, log_filepath)\n",
"role": "user"
},
{
"content": "vsearch_chimera_filter_de_novo",
"role": "assistant"
}
] |
def x(self, data):
result = {}
result['Output'] = ResultPath(Path=self.Parameters['--output'].Value,
IsWritten=self.Parameters['--output'].isOn())
result['ClusterFile'] = ResultPath(Path=self.Parameters['--uc'].Value,
IsWritten=self.Parameters['--uc'].isOn())
result['Output_aln'] = ResultPath(Path=self.Parameters['--uchimealns'].
Value, IsWritten=self.Parameters['--uchimealns'].isOn())
result['Output_tabular'] = ResultPath(Path=self.Parameters[
'--uchimeout'].Value, IsWritten=self.Parameters['--uchimeout'].isOn())
result['Output_chimeras'] = ResultPath(Path=self.Parameters[
'--chimeras'].Value, IsWritten=self.Parameters['--chimeras'].isOn())
result['Output_nonchimeras'] = ResultPath(Path=self.Parameters[
'--nonchimeras'].Value, IsWritten=self.Parameters['--nonchimeras'].
isOn())
result['LogFile'] = ResultPath(Path=self.Parameters['--log'].Value,
IsWritten=self.Parameters['--log'].isOn())
return result
| get_result_paths | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, data):\n result = {}\n result['Output'] = ResultPath(Path=self.Parameters['--output'].Value,\n IsWritten=self.Parameters['--output'].isOn())\n result['ClusterFile'] = ResultPath(Path=self.Parameters['--uc'].Value,\n IsWritten=self.Parameters['--uc'].isOn())\n result['Output_aln'] = ResultPath(Path=self.Parameters['--uchimealns'].\n Value, IsWritten=self.Parameters['--uchimealns'].isOn())\n result['Output_tabular'] = ResultPath(Path=self.Parameters[\n '--uchimeout'].Value, IsWritten=self.Parameters['--uchimeout'].isOn())\n result['Output_chimeras'] = ResultPath(Path=self.Parameters[\n '--chimeras'].Value, IsWritten=self.Parameters['--chimeras'].isOn())\n result['Output_nonchimeras'] = ResultPath(Path=self.Parameters[\n '--nonchimeras'].Value, IsWritten=self.Parameters['--nonchimeras'].\n isOn())\n result['LogFile'] = ResultPath(Path=self.Parameters['--log'].Value,\n IsWritten=self.Parameters['--log'].isOn())\n return result\n",
"role": "user"
},
{
"content": "get_result_paths",
"role": "assistant"
}
] |
def x(filename, doy, latitude, longitude, depth):
assert np.size(doy) == 1
assert np.size(latitude) == 1
assert np.size(longitude) == 1
assert np.size(depth) == 1
assert (longitude >= 0) & (longitude <= 360)
assert depth >= 0
nc = netCDF4.Dataset(filename)
t = 2 * np.pi * doy / 366
Z = np.absolute(nc['depth'][:] - depth).argmin()
I = np.absolute(nc['lat'][:] - latitude).argmin()
J = np.absolute(nc['lon'][:] - longitude).argmin()
value = nc['mean'][:, I, J]
value[:64] += nc['an_cos'][Z, I, J] * np.cos(t) + nc['an_sin'][:, I, J
] * np.sin(t)
value[:55] += nc['sa_cos'][Z, I, J] * np.cos(2 * t) + nc['sa_sin'][:, I, J
] * np.sin(2 * t)
value = value[Z]
std = nc['std_dev'][Z, I, J]
return value, std
| extract | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(filename, doy, latitude, longitude, depth):\n assert np.size(doy) == 1\n assert np.size(latitude) == 1\n assert np.size(longitude) == 1\n assert np.size(depth) == 1\n assert (longitude >= 0) & (longitude <= 360)\n assert depth >= 0\n nc = netCDF4.Dataset(filename)\n t = 2 * np.pi * doy / 366\n Z = np.absolute(nc['depth'][:] - depth).argmin()\n I = np.absolute(nc['lat'][:] - latitude).argmin()\n J = np.absolute(nc['lon'][:] - longitude).argmin()\n value = nc['mean'][:, I, J]\n value[:64] += nc['an_cos'][Z, I, J] * np.cos(t) + nc['an_sin'][:, I, J\n ] * np.sin(t)\n value[:55] += nc['sa_cos'][Z, I, J] * np.cos(2 * t) + nc['sa_sin'][:, I, J\n ] * np.sin(2 * t)\n value = value[Z]\n std = nc['std_dev'][Z, I, J]\n return value, std\n",
"role": "user"
},
{
"content": "extract",
"role": "assistant"
}
] |
def x(filename, doy, latitude, longitude, depth):
assert np.size(doy) == 1
assert np.size(latitude) == 1
assert np.size(longitude) == 1
assert (longitude >= 0) & (longitude <= 360)
assert depth >= 0
nc = netCDF4.Dataset(filename)
t = 2 * np.pi * doy / 366
Z = slice(0, nc['depth'].size)
I = np.absolute(nc['lat'][:] - latitude).argmin()
J = np.absolute(nc['lon'][:] - longitude).argmin()
assert (nc['depth'][:64] == nc['depth_ann'][:]).all()
assert (nc['depth'][:55] == nc['depth_semiann'][:]).all()
value = nc['mean'][:, I, J]
value[:64] += nc['an_cos'][Z, I, J] * np.cos(t) + nc['an_sin'][:, I, J
] * np.sin(t)
value[:55] += nc['sa_cos'][Z, I, J] * np.cos(2 * t) + nc['sa_sin'][:, I, J
] * np.sin(2 * t)
value = value
output = {'depth': np.asanyarray(depth)}
from scipy.interpolate import griddata
output['value'] = griddata(nc['depth'][Z], value[Z], depth)
for v in ['std_dev']:
output[v] = griddata(nc['depth'][Z], nc[v][Z, I, J], depth)
return output
| cars_profile | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(filename, doy, latitude, longitude, depth):\n assert np.size(doy) == 1\n assert np.size(latitude) == 1\n assert np.size(longitude) == 1\n assert (longitude >= 0) & (longitude <= 360)\n assert depth >= 0\n nc = netCDF4.Dataset(filename)\n t = 2 * np.pi * doy / 366\n Z = slice(0, nc['depth'].size)\n I = np.absolute(nc['lat'][:] - latitude).argmin()\n J = np.absolute(nc['lon'][:] - longitude).argmin()\n assert (nc['depth'][:64] == nc['depth_ann'][:]).all()\n assert (nc['depth'][:55] == nc['depth_semiann'][:]).all()\n value = nc['mean'][:, I, J]\n value[:64] += nc['an_cos'][Z, I, J] * np.cos(t) + nc['an_sin'][:, I, J\n ] * np.sin(t)\n value[:55] += nc['sa_cos'][Z, I, J] * np.cos(2 * t) + nc['sa_sin'][:, I, J\n ] * np.sin(2 * t)\n value = value\n output = {'depth': np.asanyarray(depth)}\n from scipy.interpolate import griddata\n output['value'] = griddata(nc['depth'][Z], value[Z], depth)\n for v in ['std_dev']:\n output[v] = griddata(nc['depth'][Z], nc[v][Z, I, J], depth)\n return output\n",
"role": "user"
},
{
"content": "cars_profile",
"role": "assistant"
}
] |
def x(self, doy, depth, lat, lon, var):
dims, idx = cropIndices(self.dims, lat, lon, depth)
dims['time'] = np.atleast_1d(doy)
idx['tn'] = np.arange(dims['time'].size)
xn = idx['xn']
yn = idx['yn']
zn = idx['zn']
tn = idx['tn']
subset = {}
for v in var:
if v == 'mn':
mn = []
for d in doy:
t = 2 * np.pi * d / 366
value = self.ncs[0]['mean'][:, yn, xn]
value[:64] += self.ncs[0]['an_cos'][:, yn, xn] * np.cos(t
) + self.ncs[0]['an_sin'][:, yn, xn] * np.sin(t)
value[:55] += self.ncs[0]['sa_cos'][:, yn, xn] * np.cos(2 * t
) + self.ncs[0]['sa_sin'][:, yn, xn] * np.sin(2 * t)
mn.append(value[zn])
subset['mn'] = ma.asanyarray(mn)
else:
subset[v] = ma.asanyarray(doy.size * [self[v][zn, yn, xn]])
return subset, dims
| crop | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, doy, depth, lat, lon, var):\n dims, idx = cropIndices(self.dims, lat, lon, depth)\n dims['time'] = np.atleast_1d(doy)\n idx['tn'] = np.arange(dims['time'].size)\n xn = idx['xn']\n yn = idx['yn']\n zn = idx['zn']\n tn = idx['tn']\n subset = {}\n for v in var:\n if v == 'mn':\n mn = []\n for d in doy:\n t = 2 * np.pi * d / 366\n value = self.ncs[0]['mean'][:, yn, xn]\n value[:64] += self.ncs[0]['an_cos'][:, yn, xn] * np.cos(t\n ) + self.ncs[0]['an_sin'][:, yn, xn] * np.sin(t)\n value[:55] += self.ncs[0]['sa_cos'][:, yn, xn] * np.cos(2 * t\n ) + self.ncs[0]['sa_sin'][:, yn, xn] * np.sin(2 * t)\n mn.append(value[zn])\n subset['mn'] = ma.asanyarray(mn)\n else:\n subset[v] = ma.asanyarray(doy.size * [self[v][zn, yn, xn]])\n return subset, dims\n",
"role": "user"
},
{
"content": "crop",
"role": "assistant"
}
] |
def x(self, doy, depth, lat, lon, var):
subset, dims = self.crop(doy, depth, lat, lon, var)
if np.all([(d in dims['time']) for d in doy]) & np.all([(z in dims[
'depth']) for z in depth]) & np.all([(y in dims['lat']) for y in lat]
) & np.all([(x in dims['lon']) for x in lon]):
dn = np.nonzero([(d in doy) for d in dims['time']])[0]
zn = np.nonzero([(z in depth) for z in dims['depth']])[0]
yn = np.nonzero([(y in lat) for y in dims['lat']])[0]
xn = np.nonzero([(x in lon) for x in dims['lon']])[0]
output = {}
for v in subset:
output[v] = subset[v][:, :, :, xn][:, :, yn][:, zn][dn]
return output
points_out = []
for doyn in doy:
for depthn in depth:
for latn in lat:
for lonn in lon:
points_out.append([doyn, depthn, latn, lonn])
points_out = np.array(points_out)
output = {}
for v in var:
output[v] = ma.masked_all((doy.size, depth.size, lat.size, lon.size
), dtype=subset[v].dtype)
idx = np.nonzero(~ma.getmaskarray(subset[v]))
if idx[0].size > 0:
points = np.array([dims['time'][idx[0]], dims['depth'][idx[1]],
dims['lat'][idx[2]], dims['lon'][idx[3]]]).T
values = subset[v][idx]
ind = np.array([(np.unique(points[:, i]).size > 1) for i in
range(points.shape[1])])
assert ind.any()
values[ma.getmaskarray(values)] = np.nan
values_out = griddata(np.atleast_1d(np.squeeze(points[:, ind])),
values, np.atleast_1d(np.squeeze(points_out[:, ind])))
idx = np.isfinite(values_out)
for [t, z, y, x], out in zip(points_out[idx], values_out[idx]):
output[v][t == doy, z == depth, y == lat, x == lon] = out
output[v] = ma.fix_invalid(output[v])
return output
| interpolate | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, doy, depth, lat, lon, var):\n subset, dims = self.crop(doy, depth, lat, lon, var)\n if np.all([(d in dims['time']) for d in doy]) & np.all([(z in dims[\n 'depth']) for z in depth]) & np.all([(y in dims['lat']) for y in lat]\n ) & np.all([(x in dims['lon']) for x in lon]):\n dn = np.nonzero([(d in doy) for d in dims['time']])[0]\n zn = np.nonzero([(z in depth) for z in dims['depth']])[0]\n yn = np.nonzero([(y in lat) for y in dims['lat']])[0]\n xn = np.nonzero([(x in lon) for x in dims['lon']])[0]\n output = {}\n for v in subset:\n output[v] = subset[v][:, :, :, xn][:, :, yn][:, zn][dn]\n return output\n points_out = []\n for doyn in doy:\n for depthn in depth:\n for latn in lat:\n for lonn in lon:\n points_out.append([doyn, depthn, latn, lonn])\n points_out = np.array(points_out)\n output = {}\n for v in var:\n output[v] = ma.masked_all((doy.size, depth.size, lat.size, lon.size\n ), dtype=subset[v].dtype)\n idx = np.nonzero(~ma.getmaskarray(subset[v]))\n if idx[0].size > 0:\n points = np.array([dims['time'][idx[0]], dims['depth'][idx[1]],\n dims['lat'][idx[2]], dims['lon'][idx[3]]]).T\n values = subset[v][idx]\n ind = np.array([(np.unique(points[:, i]).size > 1) for i in\n range(points.shape[1])])\n assert ind.any()\n values[ma.getmaskarray(values)] = np.nan\n values_out = griddata(np.atleast_1d(np.squeeze(points[:, ind])),\n values, np.atleast_1d(np.squeeze(points_out[:, ind])))\n idx = np.isfinite(values_out)\n for [t, z, y, x], out in zip(points_out[idx], values_out[idx]):\n output[v][t == doy, z == depth, y == lat, x == lon] = out\n output[v] = ma.fix_invalid(output[v])\n return output\n",
"role": "user"
},
{
"content": "interpolate",
"role": "assistant"
}
] |
def x(self, room_name):
api_meth = self.set_user_items_url(room_name)
message_ids = self.unread_items(room_name).get('chat')
if message_ids:
return self.post(api_meth, data={'chat': message_ids})
else:
raise GitterItemsError(room_name)
| mark_as_read | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, room_name):\n api_meth = self.set_user_items_url(room_name)\n message_ids = self.unread_items(room_name).get('chat')\n if message_ids:\n return self.post(api_meth, data={'chat': message_ids})\n else:\n raise GitterItemsError(room_name)\n",
"role": "user"
},
{
"content": "mark_as_read",
"role": "assistant"
}
] |
def x(cls, site):
provider = cls.name
cls._render_config('wsgi.py', 'wsgi.py', site)
yaml_template_name = os.path.join(provider, cls.provider_yml_name)
cls._render_config(cls.provider_yml_name, yaml_template_name, site)
requirements_filename = 'requirements.txt'
if site['requirements'] != requirements_filename:
requirements_template_name = os.path.join(provider,
requirements_filename)
cls._render_config(requirements_filename,
requirements_template_name, site)
settings_template_name = os.path.join(provider, 'settings_%s.py' % provider
)
settings_path = site['django_settings'].replace('.', '/'
) + '_%s.py' % provider
cls._render_config(settings_path, settings_template_name, site)
| create_configs | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, site):\n provider = cls.name\n cls._render_config('wsgi.py', 'wsgi.py', site)\n yaml_template_name = os.path.join(provider, cls.provider_yml_name)\n cls._render_config(cls.provider_yml_name, yaml_template_name, site)\n requirements_filename = 'requirements.txt'\n if site['requirements'] != requirements_filename:\n requirements_template_name = os.path.join(provider,\n requirements_filename)\n cls._render_config(requirements_filename,\n requirements_template_name, site)\n settings_template_name = os.path.join(provider, 'settings_%s.py' % provider\n )\n settings_path = site['django_settings'].replace('.', '/'\n ) + '_%s.py' % provider\n cls._render_config(settings_path, settings_template_name, site)\n",
"role": "user"
},
{
"content": "create_configs",
"role": "assistant"
}
] |
def x(cls, dest, template_name, template_args):
template_args = template_args.copy()
pyversion = template_args['pyversion']
template_args['pyversion'] = cls.PYVERSIONS[pyversion]
template = template_env.get_template(template_name)
contents = template.render(**template_args)
_write_file(dest, contents)
| render_config | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, dest, template_name, template_args):\n template_args = template_args.copy()\n pyversion = template_args['pyversion']\n template_args['pyversion'] = cls.PYVERSIONS[pyversion]\n template = template_env.get_template(template_name)\n contents = template.render(**template_args)\n _write_file(dest, contents)\n",
"role": "user"
},
{
"content": "render_config",
"role": "assistant"
}
] |
def x(self):
commands = sorted(self.subcommands.items(), key=lambda i: i[0])
return '\n'.join('%s %s' % (subcommand.ljust(15), textwrap.shorten(
description, width=61)) for subcommand, (description, action, opts) in
commands)
| gen_subcommand_help | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n commands = sorted(self.subcommands.items(), key=lambda i: i[0])\n return '\\n'.join('%s %s' % (subcommand.ljust(15), textwrap.shorten(\n description, width=61)) for subcommand, (description, action, opts) in\n commands)\n",
"role": "user"
},
{
"content": "gen_subcommand_help",
"role": "assistant"
}
] |
def x(self, argv=None):
args = self.parse_args(argv)
action = self.subcommands[args.subcommand][1]
return action, args
| parse_args_to_action_args | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, argv=None):\n args = self.parse_args(argv)\n action = self.subcommands[args.subcommand][1]\n return action, args\n",
"role": "user"
},
{
"content": "parse_args_to_action_args",
"role": "assistant"
}
] |
def x(self, action, name, description='', arguments={}):
action = coerce_to_synchronous(action)
opts = []
for flags, kwargs in arguments.items():
if isinstance(flags, str):
flags = tuple([flags])
opts.append((flags, kwargs))
self.subcommands[name] = description, action, opts
| register_subparser | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, action, name, description='', arguments={}):\n action = coerce_to_synchronous(action)\n opts = []\n for flags, kwargs in arguments.items():\n if isinstance(flags, str):\n flags = tuple([flags])\n opts.append((flags, kwargs))\n self.subcommands[name] = description, action, opts\n",
"role": "user"
},
{
"content": "register_subparser",
"role": "assistant"
}
] |
def x(self, description='', arguments={}):
def x(func):
self.register_subparser(func, func.__name__.replace('_', '-'),
description=description, arguments=arguments)
return func
return decorator
| subcommand | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, description='', arguments={}):\n\n def x(func):\n self.register_subparser(func, func.__name__.replace('_', '-'),\n description=description, arguments=arguments)\n return func\n return decorator\n",
"role": "user"
},
{
"content": "subcommand",
"role": "assistant"
}
] |
def x(self, *args, **kwargs):
if self._last_args and self._last_args.verbose:
print(*args, **kwargs)
| print | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, *args, **kwargs):\n if self._last_args and self._last_args.verbose:\n print(*args, **kwargs)\n",
"role": "user"
},
{
"content": "print",
"role": "assistant"
}
] |
def x(fasta_pairs_lines, uc_lines):
alignments = get_next_two_fasta_records(fasta_pairs_lines)
for hit in get_next_record_type(uc_lines, 'H'):
matching_strand = hit[4]
if matching_strand == '-':
strand_id = '-'
target_rev_match = True
elif matching_strand == '+':
strand_id = '+'
target_rev_match = False
elif matching_strand == '.':
strand_id = ''
target_rev_match = False
else:
raise UclustParseError('Unknown strand type: %s' % matching_strand)
uc_query_id = hit[8]
uc_target_id = hit[9]
percent_id = float(hit[3])
fasta_pair = alignments.next()
fasta_query_id = fasta_pair[0][0]
aligned_query = fasta_pair[0][1]
if fasta_query_id != uc_query_id:
raise UclustParseError(
'Order of fasta and uc files do not match.' +
' Got query %s but expected %s.' % (fasta_query_id,
uc_query_id))
fasta_target_id = fasta_pair[1][0]
aligned_target = fasta_pair[1][1]
if fasta_target_id != uc_target_id + strand_id:
raise UclustParseError(
'Order of fasta and uc files do not match.' +
' Got target %s but expected %s.' % (fasta_target_id,
uc_target_id + strand_id))
if target_rev_match:
query_id = uc_query_id + ' RC'
aligned_query = DNA.rc(aligned_query)
target_id = uc_target_id
aligned_target = DNA.rc(aligned_target)
else:
query_id = uc_query_id
aligned_query = aligned_query
target_id = uc_target_id
aligned_target = aligned_target
yield query_id, target_id, aligned_query, aligned_target, percent_id
| process_uclust_pw_alignment_results | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fasta_pairs_lines, uc_lines):\n alignments = get_next_two_fasta_records(fasta_pairs_lines)\n for hit in get_next_record_type(uc_lines, 'H'):\n matching_strand = hit[4]\n if matching_strand == '-':\n strand_id = '-'\n target_rev_match = True\n elif matching_strand == '+':\n strand_id = '+'\n target_rev_match = False\n elif matching_strand == '.':\n strand_id = ''\n target_rev_match = False\n else:\n raise UclustParseError('Unknown strand type: %s' % matching_strand)\n uc_query_id = hit[8]\n uc_target_id = hit[9]\n percent_id = float(hit[3])\n fasta_pair = alignments.next()\n fasta_query_id = fasta_pair[0][0]\n aligned_query = fasta_pair[0][1]\n if fasta_query_id != uc_query_id:\n raise UclustParseError(\n 'Order of fasta and uc files do not match.' + \n ' Got query %s but expected %s.' % (fasta_query_id,\n uc_query_id))\n fasta_target_id = fasta_pair[1][0]\n aligned_target = fasta_pair[1][1]\n if fasta_target_id != uc_target_id + strand_id:\n raise UclustParseError(\n 'Order of fasta and uc files do not match.' + \n ' Got target %s but expected %s.' % (fasta_target_id, \n uc_target_id + strand_id))\n if target_rev_match:\n query_id = uc_query_id + ' RC'\n aligned_query = DNA.rc(aligned_query)\n target_id = uc_target_id\n aligned_target = DNA.rc(aligned_target)\n else:\n query_id = uc_query_id\n aligned_query = aligned_query\n target_id = uc_target_id\n aligned_target = aligned_target\n yield query_id, target_id, aligned_query, aligned_target, percent_id\n",
"role": "user"
},
{
"content": "process_uclust_pw_alignment_results",
"role": "assistant"
}
] |
def x(uc_lines, error_on_multiple_hits=True):
clusters = {}
failures = []
seeds = []
all_hits = set()
hit_types = {}.fromkeys(list('HSNL'))
for record in get_next_record_type(uc_lines, hit_types):
hit_type = record[0]
query_id = record[8].split()[0]
target_cluster = record[9].split()[0]
if hit_type == 'H':
if error_on_multiple_hits and query_id in all_hits:
raise UclustParseError('Query id ' + query_id +
" hit multiple seeds. This can happen if --allhits is enabled in the call to uclust, which isn't supported by default. Call clusters_from_uc_file(lines, error_on_multiple_hits=False) to allow a query to cluster to multiple seeds."
)
else:
clusters[target_cluster].append(query_id)
all_hits.add(query_id)
elif hit_type == 'S':
if query_id in clusters:
raise UclustParseError(
'A seq id was provided as a seed, but that seq id already represents a cluster. Are there overlapping seq ids in your reference and input files or repeated seq ids in either? Offending seq id is %s'
% query_id)
clusters[query_id] = [query_id]
seeds.append(query_id)
elif hit_type == 'L':
if query_id in clusters:
raise UclustParseError(
'A seq id was provided as a seed, but that seq id already represents a cluster. Are there overlapping seq ids in your reference and input files or repeated seq ids in either? Offending seq id is %s'
% query_id)
clusters[query_id] = []
elif hit_type == 'N':
failures.append(query_id)
else:
raise UclustParseError('Unexpected result parsing line:\n%s' %
'\t'.join(record))
return clusters, failures, seeds
| clusters_from_uc_file | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(uc_lines, error_on_multiple_hits=True):\n clusters = {}\n failures = []\n seeds = []\n all_hits = set()\n hit_types = {}.fromkeys(list('HSNL'))\n for record in get_next_record_type(uc_lines, hit_types):\n hit_type = record[0]\n query_id = record[8].split()[0]\n target_cluster = record[9].split()[0]\n if hit_type == 'H':\n if error_on_multiple_hits and query_id in all_hits:\n raise UclustParseError('Query id ' + query_id +\n \" hit multiple seeds. This can happen if --allhits is enabled in the call to uclust, which isn't supported by default. Call clusters_from_uc_file(lines, error_on_multiple_hits=False) to allow a query to cluster to multiple seeds.\"\n )\n else:\n clusters[target_cluster].append(query_id)\n all_hits.add(query_id)\n elif hit_type == 'S':\n if query_id in clusters:\n raise UclustParseError(\n 'A seq id was provided as a seed, but that seq id already represents a cluster. Are there overlapping seq ids in your reference and input files or repeated seq ids in either? Offending seq id is %s'\n % query_id)\n clusters[query_id] = [query_id]\n seeds.append(query_id)\n elif hit_type == 'L':\n if query_id in clusters:\n raise UclustParseError(\n 'A seq id was provided as a seed, but that seq id already represents a cluster. Are there overlapping seq ids in your reference and input files or repeated seq ids in either? Offending seq id is %s'\n % query_id)\n clusters[query_id] = []\n elif hit_type == 'N':\n failures.append(query_id)\n else:\n raise UclustParseError('Unexpected result parsing line:\\n%s' %\n '\\t'.join(record))\n return clusters, failures, seeds\n",
"role": "user"
},
{
"content": "clusters_from_uc_file",
"role": "assistant"
}
] |
def x(fasta_filepath, output_filepath=None, tmp_dir=gettempdir(), HALT_EXEC
=False):
if not output_filepath:
_, output_filepath = mkstemp(dir=tmp_dir, prefix=
'uclust_fasta_sort', suffix='.fasta')
app = Uclust(params={'--tmpdir': tmp_dir}, TmpDir=tmp_dir, HALT_EXEC=
HALT_EXEC)
app_result = app(data={'--mergesort': fasta_filepath, '--output':
output_filepath})
return app_result
| uclust_fasta_sort_from_filepath | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fasta_filepath, output_filepath=None, tmp_dir=gettempdir(), HALT_EXEC\n =False):\n if not output_filepath:\n _, output_filepath = mkstemp(dir=tmp_dir, prefix=\n 'uclust_fasta_sort', suffix='.fasta')\n app = Uclust(params={'--tmpdir': tmp_dir}, TmpDir=tmp_dir, HALT_EXEC=\n HALT_EXEC)\n app_result = app(data={'--mergesort': fasta_filepath, '--output':\n output_filepath})\n return app_result\n",
"role": "user"
},
{
"content": "uclust_fasta_sort_from_filepath",
"role": "assistant"
}
] |
def x(query_fasta_filepath, subject_fasta_filepath, percent_ID=0.75,
enable_rev_strand_matching=True, max_accepts=8, max_rejects=32, tmp_dir
=gettempdir(), HALT_EXEC=False):
params = {'--id': percent_ID, '--maxaccepts': max_accepts,
'--maxrejects': max_rejects, '--libonly': True, '--lib':
subject_fasta_filepath, '--tmpdir': tmp_dir}
if enable_rev_strand_matching:
params['--rev'] = True
app = Uclust(params, TmpDir=tmp_dir, HALT_EXEC=HALT_EXEC)
_, alignment_filepath = mkstemp(dir=tmp_dir, prefix='uclust_alignments',
suffix='.fasta')
_, uc_filepath = mkstemp(dir=tmp_dir, prefix='uclust_results', suffix='.uc'
)
input_data = {'--input': query_fasta_filepath, '--fastapairs':
alignment_filepath, '--uc': uc_filepath}
app_result = app(input_data)
for result in process_uclust_pw_alignment_results(app_result[
'PairwiseAlignments'], app_result['ClusterFile']):
try:
yield result
except GeneratorExit:
break
app_result.cleanUp()
return
| uclust_search_and_align_from_fasta_filepath | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(query_fasta_filepath, subject_fasta_filepath, percent_ID=0.75,\n enable_rev_strand_matching=True, max_accepts=8, max_rejects=32, tmp_dir\n =gettempdir(), HALT_EXEC=False):\n params = {'--id': percent_ID, '--maxaccepts': max_accepts,\n '--maxrejects': max_rejects, '--libonly': True, '--lib':\n subject_fasta_filepath, '--tmpdir': tmp_dir}\n if enable_rev_strand_matching:\n params['--rev'] = True\n app = Uclust(params, TmpDir=tmp_dir, HALT_EXEC=HALT_EXEC)\n _, alignment_filepath = mkstemp(dir=tmp_dir, prefix='uclust_alignments',\n suffix='.fasta')\n _, uc_filepath = mkstemp(dir=tmp_dir, prefix='uclust_results', suffix='.uc'\n )\n input_data = {'--input': query_fasta_filepath, '--fastapairs':\n alignment_filepath, '--uc': uc_filepath}\n app_result = app(input_data)\n for result in process_uclust_pw_alignment_results(app_result[\n 'PairwiseAlignments'], app_result['ClusterFile']):\n try:\n yield result\n except GeneratorExit:\n break\n app_result.cleanUp()\n return\n",
"role": "user"
},
{
"content": "uclust_search_and_align_from_fasta_filepath",
"role": "assistant"
}
] |
def x(fasta_filepath, uc_save_filepath=None, percent_ID=0.97, max_accepts=1,
max_rejects=8, stepwords=8, word_length=8, optimal=False, exact=False,
suppress_sort=False, enable_rev_strand_matching=False,
subject_fasta_filepath=None, suppress_new_clusters=False, stable_sort=
False, tmp_dir=gettempdir(), HALT_EXEC=False):
output_filepath = uc_save_filepath
if not output_filepath:
_, output_filepath = mkstemp(dir=tmp_dir, prefix='uclust_clusters',
suffix='.uc')
params = {'--id': percent_ID, '--maxaccepts': max_accepts,
'--maxrejects': max_rejects, '--stepwords': stepwords, '--w':
word_length, '--tmpdir': tmp_dir}
app = Uclust(params, TmpDir=tmp_dir, HALT_EXEC=HALT_EXEC)
if enable_rev_strand_matching:
app.Parameters['--rev'].on()
if optimal:
app.Parameters['--optimal'].on()
if exact:
app.Parameters['--exact'].on()
if suppress_sort:
app.Parameters['--usersort'].on()
if subject_fasta_filepath:
app.Parameters['--lib'].on(subject_fasta_filepath)
if suppress_new_clusters:
app.Parameters['--libonly'].on()
if stable_sort:
app.Parameters['--stable_sort'].on()
app_result = app({'--input': fasta_filepath, '--uc': output_filepath})
return app_result
| uclust_cluster_from_sorted_fasta_filepath | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fasta_filepath, uc_save_filepath=None, percent_ID=0.97, max_accepts=1,\n max_rejects=8, stepwords=8, word_length=8, optimal=False, exact=False,\n suppress_sort=False, enable_rev_strand_matching=False,\n subject_fasta_filepath=None, suppress_new_clusters=False, stable_sort=\n False, tmp_dir=gettempdir(), HALT_EXEC=False):\n output_filepath = uc_save_filepath\n if not output_filepath:\n _, output_filepath = mkstemp(dir=tmp_dir, prefix='uclust_clusters',\n suffix='.uc')\n params = {'--id': percent_ID, '--maxaccepts': max_accepts,\n '--maxrejects': max_rejects, '--stepwords': stepwords, '--w':\n word_length, '--tmpdir': tmp_dir}\n app = Uclust(params, TmpDir=tmp_dir, HALT_EXEC=HALT_EXEC)\n if enable_rev_strand_matching:\n app.Parameters['--rev'].on()\n if optimal:\n app.Parameters['--optimal'].on()\n if exact:\n app.Parameters['--exact'].on()\n if suppress_sort:\n app.Parameters['--usersort'].on()\n if subject_fasta_filepath:\n app.Parameters['--lib'].on(subject_fasta_filepath)\n if suppress_new_clusters:\n app.Parameters['--libonly'].on()\n if stable_sort:\n app.Parameters['--stable_sort'].on()\n app_result = app({'--input': fasta_filepath, '--uc': output_filepath})\n return app_result\n",
"role": "user"
},
{
"content": "uclust_cluster_from_sorted_fasta_filepath",
"role": "assistant"
}
] |
def x(fasta_filepath, original_fasta_path, percent_ID=0.97, max_accepts=1,
max_rejects=8, stepwords=8, word_length=8, optimal=False, exact=False,
suppress_sort=False, output_dir=None, enable_rev_strand_matching=False,
subject_fasta_filepath=None, suppress_new_clusters=False,
return_cluster_maps=False, stable_sort=False, tmp_dir=gettempdir(),
save_uc_files=True, HALT_EXEC=False):
fasta_output_filepath = None
uc_output_filepath = None
cd_hit_filepath = None
if output_dir and not output_dir.endswith('/'):
output_dir += '/'
if save_uc_files:
uc_save_filepath = get_output_filepaths(output_dir, original_fasta_path
)
else:
uc_save_filepath = None
sorted_fasta_filepath = ''
uc_filepath = ''
clstr_filepath = ''
files_to_remove = []
try:
if not suppress_sort:
sort_fasta = uclust_fasta_sort_from_filepath(fasta_filepath,
output_filepath=fasta_output_filepath)
sorted_fasta_filepath = sort_fasta['Output'].name
files_to_remove.append(sorted_fasta_filepath)
else:
sort_fasta = None
sorted_fasta_filepath = fasta_filepath
uclust_cluster = uclust_cluster_from_sorted_fasta_filepath(
sorted_fasta_filepath, uc_save_filepath, percent_ID=percent_ID,
max_accepts=max_accepts, max_rejects=max_rejects, stepwords=
stepwords, word_length=word_length, optimal=optimal, exact=
exact, suppress_sort=suppress_sort, enable_rev_strand_matching=
enable_rev_strand_matching, subject_fasta_filepath=
subject_fasta_filepath, suppress_new_clusters=
suppress_new_clusters, stable_sort=stable_sort, tmp_dir=tmp_dir,
HALT_EXEC=HALT_EXEC)
remove_files(files_to_remove)
except ApplicationError:
remove_files(files_to_remove)
raise ApplicationError(
'Error running uclust. Possible causes are unsupported version (current supported version is v1.2.22) is installed or improperly formatted input file was provided'
)
except ApplicationNotFoundError:
remove_files(files_to_remove)
raise ApplicationNotFoundError('uclust not found, is it properly ' +
'installed?')
clusters, failures, seeds = clusters_from_uc_file(uclust_cluster[
'ClusterFile'])
if not save_uc_files:
uclust_cluster.cleanUp()
if return_cluster_maps:
return clusters, failures, seeds
else:
return clusters.values(), failures, seeds
| get_clusters_from_fasta_filepath | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fasta_filepath, original_fasta_path, percent_ID=0.97, max_accepts=1,\n max_rejects=8, stepwords=8, word_length=8, optimal=False, exact=False,\n suppress_sort=False, output_dir=None, enable_rev_strand_matching=False,\n subject_fasta_filepath=None, suppress_new_clusters=False,\n return_cluster_maps=False, stable_sort=False, tmp_dir=gettempdir(),\n save_uc_files=True, HALT_EXEC=False):\n fasta_output_filepath = None\n uc_output_filepath = None\n cd_hit_filepath = None\n if output_dir and not output_dir.endswith('/'):\n output_dir += '/'\n if save_uc_files:\n uc_save_filepath = get_output_filepaths(output_dir, original_fasta_path\n )\n else:\n uc_save_filepath = None\n sorted_fasta_filepath = ''\n uc_filepath = ''\n clstr_filepath = ''\n files_to_remove = []\n try:\n if not suppress_sort:\n sort_fasta = uclust_fasta_sort_from_filepath(fasta_filepath,\n output_filepath=fasta_output_filepath)\n sorted_fasta_filepath = sort_fasta['Output'].name\n files_to_remove.append(sorted_fasta_filepath)\n else:\n sort_fasta = None\n sorted_fasta_filepath = fasta_filepath\n uclust_cluster = uclust_cluster_from_sorted_fasta_filepath(\n sorted_fasta_filepath, uc_save_filepath, percent_ID=percent_ID,\n max_accepts=max_accepts, max_rejects=max_rejects, stepwords=\n stepwords, word_length=word_length, optimal=optimal, exact=\n exact, suppress_sort=suppress_sort, enable_rev_strand_matching=\n enable_rev_strand_matching, subject_fasta_filepath=\n subject_fasta_filepath, suppress_new_clusters=\n suppress_new_clusters, stable_sort=stable_sort, tmp_dir=tmp_dir,\n HALT_EXEC=HALT_EXEC)\n remove_files(files_to_remove)\n except ApplicationError:\n remove_files(files_to_remove)\n raise ApplicationError(\n 'Error running uclust. Possible causes are unsupported version (current supported version is v1.2.22) is installed or improperly formatted input file was provided'\n )\n except ApplicationNotFoundError:\n remove_files(files_to_remove)\n raise ApplicationNotFoundError('uclust not found, is it properly ' +\n 'installed?')\n clusters, failures, seeds = clusters_from_uc_file(uclust_cluster[\n 'ClusterFile'])\n if not save_uc_files:\n uclust_cluster.cleanUp()\n if return_cluster_maps:\n return clusters, failures, seeds\n else:\n return clusters.values(), failures, seeds\n",
"role": "user"
},
{
"content": "get_clusters_from_fasta_filepath",
"role": "assistant"
}
] |
def x(self, data):
allowed_values = ['--input', '--uc', '--fastapairs', '--uc2clstr',
'--output', '--mergesort']
unsupported_parameters = set(data.keys()) - set(allowed_values)
if unsupported_parameters:
raise ApplicationError(
'Unsupported parameter(s) passed when calling uclust: %s' % ' '
.join(unsupported_parameters))
for v in allowed_values:
self.Parameters[v].off()
if v in data:
self.Parameters[v].on(data[v])
return ''
| input_as_parameters | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, data):\n allowed_values = ['--input', '--uc', '--fastapairs', '--uc2clstr',\n '--output', '--mergesort']\n unsupported_parameters = set(data.keys()) - set(allowed_values)\n if unsupported_parameters:\n raise ApplicationError(\n 'Unsupported parameter(s) passed when calling uclust: %s' % ' '\n .join(unsupported_parameters))\n for v in allowed_values:\n self.Parameters[v].off()\n if v in data:\n self.Parameters[v].on(data[v])\n return ''\n",
"role": "user"
},
{
"content": "input_as_parameters",
"role": "assistant"
}
] |
def x(self, data):
result = {}
result['Output'] = ResultPath(Path=self.Parameters['--output'].Value,
IsWritten=self.Parameters['--output'].isOn())
result['ClusterFile'] = ResultPath(Path=self.Parameters['--uc'].Value,
IsWritten=self.Parameters['--uc'].isOn())
result['PairwiseAlignments'] = ResultPath(Path=self.Parameters[
'--fastapairs'].Value, IsWritten=self.Parameters['--fastapairs'].isOn()
)
return result
| get_result_paths | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, data):\n result = {}\n result['Output'] = ResultPath(Path=self.Parameters['--output'].Value,\n IsWritten=self.Parameters['--output'].isOn())\n result['ClusterFile'] = ResultPath(Path=self.Parameters['--uc'].Value,\n IsWritten=self.Parameters['--uc'].isOn())\n result['PairwiseAlignments'] = ResultPath(Path=self.Parameters[\n '--fastapairs'].Value, IsWritten=self.Parameters['--fastapairs'].isOn()\n )\n return result\n",
"role": "user"
},
{
"content": "get_result_paths",
"role": "assistant"
}
] |
def x(status=None, title=None, detail=None, code=None):
error = {}
error.update({'title': title})
if status is not None:
error.update({'status': status})
if detail is not None:
error.update({'detail': detail})
if code is not None:
error.update({'code': code})
return error
| format_error | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(status=None, title=None, detail=None, code=None):\n error = {}\n error.update({'title': title})\n if status is not None:\n error.update({'status': status})\n if detail is not None:\n error.update({'detail': detail})\n if code is not None:\n error.update({'code': code})\n return error\n",
"role": "user"
},
{
"content": "format_error",
"role": "assistant"
}
] |
def x(*args):
list_errors = []
list_errors.extend(list(args))
errors = {'errors': list_errors}
return errors
| return_an_error | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(*args):\n list_errors = []\n list_errors.extend(list(args))\n errors = {'errors': list_errors}\n return errors\n",
"role": "user"
},
{
"content": "return_an_error",
"role": "assistant"
}
] |
def x(self, orm):
for doc in orm['document_library.Document'].objects.all():
for title in doc.documenttitle_set.all():
title.is_published = doc.is_published
title.save()
| forwards | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, orm):\n for doc in orm['document_library.Document'].objects.all():\n for title in doc.documenttitle_set.all():\n title.is_published = doc.is_published\n title.save()\n",
"role": "user"
},
{
"content": "forwards",
"role": "assistant"
}
] |
def x(self, time_doc, activities):
batch_updates = [time_doc]
td_start = time_doc['_start']
activities = filter(lambda act: act[0] < td_start and act[1] in
time_doc, activities)
creation_field = self.lconfig.get('cfield')
activities.sort(reverse=True, key=lambda o: o[0])
new_doc = {}
for when, field, removed, added in activities:
last_doc = batch_updates.pop()
if last_doc['_end'] == when:
new_doc = deepcopy(last_doc)
last_doc = batch_updates.pop()
else:
new_doc = deepcopy(last_doc)
new_doc['_start'] = when
new_doc['_end'] = when
last_doc['_start'] = when
last_val = last_doc[field]
new_val, inconsistent = self._activity_backwards(new_doc[field],
removed, added)
new_doc[field] = new_val
if inconsistent:
self._log_inconsistency(last_doc, last_val, field, removed,
added, when)
new_doc['_e'] = {} if not new_doc.get('_e') else new_doc['_e']
new_doc['_e'][field] = added
batch_updates.extend([last_doc, new_doc])
try:
last_doc = batch_updates[-1]
if creation_field:
creation_ts = dt2ts(last_doc[creation_field])
if creation_ts < last_doc['_start']:
last_doc['_start'] = creation_ts
elif len(batch_updates) == 1:
return []
else:
pass
except Exception as e:
logger.error('Error updating creation time; %s' % e)
return batch_updates
| activity_import_doc | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, time_doc, activities):\n batch_updates = [time_doc]\n td_start = time_doc['_start']\n activities = filter(lambda act: act[0] < td_start and act[1] in\n time_doc, activities)\n creation_field = self.lconfig.get('cfield')\n activities.sort(reverse=True, key=lambda o: o[0])\n new_doc = {}\n for when, field, removed, added in activities:\n last_doc = batch_updates.pop()\n if last_doc['_end'] == when:\n new_doc = deepcopy(last_doc)\n last_doc = batch_updates.pop()\n else:\n new_doc = deepcopy(last_doc)\n new_doc['_start'] = when\n new_doc['_end'] = when\n last_doc['_start'] = when\n last_val = last_doc[field]\n new_val, inconsistent = self._activity_backwards(new_doc[field],\n removed, added)\n new_doc[field] = new_val\n if inconsistent:\n self._log_inconsistency(last_doc, last_val, field, removed,\n added, when)\n new_doc['_e'] = {} if not new_doc.get('_e') else new_doc['_e']\n new_doc['_e'][field] = added\n batch_updates.extend([last_doc, new_doc])\n try:\n last_doc = batch_updates[-1]\n if creation_field:\n creation_ts = dt2ts(last_doc[creation_field])\n if creation_ts < last_doc['_start']:\n last_doc['_start'] = creation_ts\n elif len(batch_updates) == 1:\n return []\n else:\n pass\n except Exception as e:\n logger.error('Error updating creation time; %s' % e)\n return batch_updates\n",
"role": "user"
},
{
"content": "activity_import_doc",
"role": "assistant"
}
] |
def x(self, last_update=None):
mtime_columns = self.lconfig.get('delta_mtime', [])
if not (mtime_columns and last_update):
return []
mtime_columns = str2list(mtime_columns)
where = []
for _column in mtime_columns:
_sql = '%s >= %s' % (_column, last_update)
where.append(_sql)
return self.sql_get_oids(where)
| get_changed_oids | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, last_update=None):\n mtime_columns = self.lconfig.get('delta_mtime', [])\n if not (mtime_columns and last_update):\n return []\n mtime_columns = str2list(mtime_columns)\n where = []\n for _column in mtime_columns:\n _sql = '%s >= %s' % (_column, last_update)\n where.append(_sql)\n return self.sql_get_oids(where)\n",
"role": "user"
},
{
"content": "get_changed_oids",
"role": "assistant"
}
] |
def x(self):
if hasattr(self, '_sql_fieldmap') and self._sql_fieldmap:
fieldmap = self._sql_fieldmap
else:
fieldmap = defaultdict(str)
fields = copy(self.fields)
for field, opts in fields.iteritems():
field_id = opts.get('what')
if field_id is not None:
fieldmap[field_id] = field
self._sql_fieldmap = fieldmap
return fieldmap
| fieldmap | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if hasattr(self, '_sql_fieldmap') and self._sql_fieldmap:\n fieldmap = self._sql_fieldmap\n else:\n fieldmap = defaultdict(str)\n fields = copy(self.fields)\n for field, opts in fields.iteritems():\n field_id = opts.get('what')\n if field_id is not None:\n fieldmap[field_id] = field\n self._sql_fieldmap = fieldmap\n return fieldmap\n",
"role": "user"
},
{
"content": "fieldmap",
"role": "assistant"
}
] |
def x(self, force=None, last_update=None, flush=False):
return self._run_object_import(force=force, last_update=last_update,
flush=flush, full_history=False)
| get_objects | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, force=None, last_update=None, flush=False):\n return self._run_object_import(force=force, last_update=last_update,\n flush=flush, full_history=False)\n",
"role": "user"
},
{
"content": "get_objects",
"role": "assistant"
}
] |
def x(self):
table = self.lconfig.get('table')
_oid = self.lconfig.get('_oid')
if is_array(_oid):
_oid = _oid[0]
last_id = self.container.get_last_field(field='_oid')
ids = []
if last_id:
try:
last_id = float(last_id)
where = '%s.%s > %s' % (table, _oid, last_id)
except (TypeError, ValueError):
where = "%s.%s > '%s'" % (table, _oid, last_id)
ids = self.sql_get_oids(where)
return ids
| get_new_oids | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n table = self.lconfig.get('table')\n _oid = self.lconfig.get('_oid')\n if is_array(_oid):\n _oid = _oid[0]\n last_id = self.container.get_last_field(field='_oid')\n ids = []\n if last_id:\n try:\n last_id = float(last_id)\n where = '%s.%s > %s' % (table, _oid, last_id)\n except (TypeError, ValueError):\n where = \"%s.%s > '%s'\" % (table, _oid, last_id)\n ids = self.sql_get_oids(where)\n return ids\n",
"role": "user"
},
{
"content": "get_new_oids",
"role": "assistant"
}
] |
def x(self, force=None, last_update=None, flush=False):
return self._run_object_import(force=force, last_update=last_update,
flush=flush, full_history=True)
| get_full_history | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, force=None, last_update=None, flush=False):\n return self._run_object_import(force=force, last_update=last_update,\n flush=flush, full_history=True)\n",
"role": "user"
},
{
"content": "get_full_history",
"role": "assistant"
}
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.