source_code stringlengths 45 7.24k | function_name stringlengths 1 111 | messages list |
|---|---|---|
def x(self, klass=_x):
chain = super().copy()
new_chain = klass(chain._args[0])
new_chain._tokens = [[chain.compose, [], {}]]
return new_chain
| copy | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, klass=_x):\n chain = super().copy()\n new_chain = klass(chain._args[0])\n new_chain._tokens = [[chain.compose, [], {}]]\n return new_chain\n"... |
def x(filename):
return os.path.split(re.sub(_afni_suffix_regex, '', str(filename)))[1]
| prefix | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(filename):\n return os.path.split(re.sub(_afni_suffix_regex, '', str(filename)))[1]\n",
"role": "user"
},
{
"content": "prefix",
"role": "as... |
def x(filename, suffix):
return os.path.split(re.sub(_afni_suffix_regex, '%s\\g<1>' % suffix,
str(filename)))[1]
| suffix | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(filename, suffix):\n return os.path.split(re.sub(_afni_suffix_regex, '%s\\\\g<1>' % suffix,\n str(filename)))[1]\n",
"role": "user"
},
{
... |
def x(filename):
if nl.pkg_available('afni', True):
afni_filename = '%s+orig' % nl.prefix(filename)
if not os.path.exists(afni_filename + '.HEAD'):
nl.calc(filename, 'a', prefix=nl.prefix(filename))
return afni_filename
| afni_copy | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(filename):\n if nl.pkg_available('afni', True):\n afni_filename = '%s+orig' % nl.prefix(filename)\n if not os.path.exists(afni_filename + '.HE... |
def x(filename, prefix=None, gzip=True):
if prefix == None:
prefix = filename
nifti_filename = globals()['prefix'](prefix) + '.nii'
if gzip:
nifti_filename += '.gz'
if not os.path.exists(nifti_filename):
try:
subprocess.check_call(['3dAFNItoNIFTI', '-prefix',
nifti_filename, str(filename)])
except subprocess.CalledProcessError:
nl.notify('Error: could not convert "%s" to NIFTI dset!' %
filename, level=nl.level.error)
return None
return nifti_filename
| nifti_copy | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(filename, prefix=None, gzip=True):\n if prefix == None:\n prefix = filename\n nifti_filename = globals()['prefix'](prefix) + '.nii'\n if gzip:\... |
def x(dset, to_dir):
if nl.is_afni(dset):
dset_strip = re.sub('\\.(HEAD|BRIK)?(\\.(gz|bz))?', '', dset)
for dset_file in ([dset_strip + '.HEAD'] + glob.glob(dset_strip +
'.BRIK*')):
if os.path.exists(dset_file):
shutil.copy(dset_file, to_dir)
elif os.path.exists(dset):
shutil.copy(dset, to_dir)
else:
nl.notify("Warning: couldn't find file %s to copy to %s" % (dset,
to_dir), level=nl.level.warning)
| dset_copy | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(dset, to_dir):\n if nl.is_afni(dset):\n dset_strip = re.sub('\\\\.(HEAD|BRIK)?(\\\\.(gz|bz))?', '', dset)\n for dset_file in ([dset_strip + '.... |
def x(dset):
info = DsetInfo()
try:
raw_info = subprocess.check_output(['3dinfo', '-verb', str(dset)],
stderr=subprocess.STDOUT)
except:
return None
if raw_info == None:
return None
sub_pattern = (
"At sub-brick #(\\d+) \\'([^\\']+)\\' datum type is (\\w+)(:\\s+(.*)\\s+to\\s+(.*))?\\n(.*statcode = (\\w+); statpar = (.*)|)"
)
sub_info = re.findall(sub_pattern, raw_info)
for brick in sub_info:
brick_info = {'index': int(brick[0]), 'label': brick[1], 'datum':
brick[2]}
if brick[3] != '':
brick_info.update({'min': float(brick[4]), 'max': float(brick[5])})
if brick[6] != '':
brick_info.update({'stat': brick[7], 'params': brick[8].split()})
info.subbricks.append(brick_info)
info.reps = len(info.subbricks)
orient = re.search('\\[-orient ([A-Z]+)\\]', raw_info)
if orient:
info.orient = orient.group(1)
for axis in ['RL', 'AP', 'IS']:
m = re.search(
'%s-to-%s extent:\\s+([0-9-.]+) \\[.\\] -to-\\s+([0-9-.]+) \\[.\\] -step-\\s+([0-9-.]+) mm \\[\\s*([0-9]+) voxels\\]'
% (axis[0], axis[1]), raw_info)
if m:
info.spatial_from.append(float(m.group(1)))
info.spatial_to.append(float(m.group(2)))
info.voxel_size.append(float(m.group(3)))
info.voxel_dims.append(float(m.group(4)))
if len(info.voxel_size) == 3:
info.voxel_volume = reduce(mul, info.voxel_size)
slice_timing = re.findall('-time:[tz][tz] \\d+ \\d+ [0-9.]+ (.*?) ',
raw_info)
if len(slice_timing):
info.slice_timing = slice_timing[0]
TR = re.findall('Time step = ([0-9.]+)s', raw_info)
if len(TR):
info.TR = float(TR[0])
details_regex = {'identifier': 'Identifier Code:\\s+(.*)', 'filetype':
'Storage Mode:\\s+(.*)', 'space': 'Template Space:\\s+(.*)'}
for d in details_regex:
m = re.search(details_regex[d], raw_info)
if m:
setattr(info, d, m.group(1))
return info
| dset_info_afni | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(dset):\n info = DsetInfo()\n try:\n raw_info = subprocess.check_output(['3dinfo', '-verb', str(dset)],\n stderr=subprocess.STDOUT)\n ... |
def x(dset):
if nl.pkg_available('afni'):
return _dset_info_afni(dset)
nl.notify('Error: no packages available to get dset info', level=nl.
level.error)
return None
| dset_info | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(dset):\n if nl.pkg_available('afni'):\n return _dset_info_afni(dset)\n nl.notify('Error: no packages available to get dset info', level=nl.\n ... |
def x(dset, label, coef=False, tstat=False, fstat=False, rstat=False,
number_only=False):
if coef is not False:
if coef is True:
coef = 0
label += '#%d_Coef' % coef
elif tstat != False:
if tstat == True:
tstat = 0
label += '#%d_Tstat' % tstat
elif fstat:
label += '_Fstat'
elif rstat:
label += '_R^2'
info = nl.dset_info(dset)
if info == None:
nl.notify('Error: Couldn\'t get info from dset "%s"' % dset, level=
nl.level.error)
return None
i = info.subbrick_labeled(label)
if number_only:
return i
return '%s[%d]' % (dset, i)
| subbrick | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(dset, label, coef=False, tstat=False, fstat=False, rstat=False,\n number_only=False):\n if coef is not False:\n if coef is True:\n coef... |
def x(dsets):
infos = [dset_info(dset) for dset in dsets]
for i in xrange(3):
if len(set([x.voxel_size[i] for x in infos])) > 1 or len(set([x.
voxel_dims[i] for x in infos])) > 1:
return False
return True
| dset_grids_equal | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(dsets):\n infos = [dset_info(dset) for dset in dsets]\n for i in xrange(3):\n if len(set([x.voxel_size[i] for x in infos])) > 1 or len(set([x.\n ... |
def x(dset, template, prefix=None, resam='NN'):
if prefix == None:
prefix = nl.suffix(dset, '_resam')
nl.run(['3dresample', '-master', template, '-rmode', resam, '-prefix',
prefix, '-inset', dset])
| resample_dset | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(dset, template, prefix=None, resam='NN'):\n if prefix == None:\n prefix = nl.suffix(dset, '_resam')\n nl.run(['3dresample', '-master', template, '... |
def x(dset, ijk):
i = nl.dset_info(dset)
orient_codes = [int(x) for x in nl.run(['@AfniOrient2RAImap', i.orient]
).output.split()]
orient_is = [(abs(x) - 1) for x in orient_codes]
rai = []
for rai_i in xrange(3):
ijk_i = orient_is[rai_i]
if orient_codes[rai_i] > 0:
rai.append(ijk[ijk_i] * i.voxel_size[rai_i] + i.spatial_from[rai_i]
)
else:
rai.append(i.spatial_to[rai_i] - ijk[ijk_i] * i.voxel_size[rai_i])
return rai
| ijk_to_xyz | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(dset, ijk):\n i = nl.dset_info(dset)\n orient_codes = [int(x) for x in nl.run(['@AfniOrient2RAImap', i.orient]\n ).output.split()]\n orient_is ... |
def x(dset):
o = nl.run(['3dAutobox', '-input', dset])
ijk_coords = re.findall('[xyz]=(\\d+)\\.\\.(\\d+)', o.output)
from_rai = ijk_to_xyz(dset, [float(x[0]) for x in ijk_coords])
to_rai = ijk_to_xyz(dset, [float(x[1]) for x in ijk_coords])
return from_rai, to_rai
| bounding_box | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(dset):\n o = nl.run(['3dAutobox', '-input', dset])\n ijk_coords = re.findall('[xyz]=(\\\\d+)\\\\.\\\\.(\\\\d+)', o.output)\n from_rai = ijk_to_xyz(dse... |
def x(dset, coords):
return nl.numberize(nl.run(['3dmaskave', '-q', '-dbox'] + list(coords) +
[dset], stderr=None).output)
| value_at_coord | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(dset, coords):\n return nl.numberize(nl.run(['3dmaskave', '-q', '-dbox'] + list(coords) +\n [dset], stderr=None).output)\n",
"role": "user"
},
... |
def x(self, args):
parser = CommandArgumentParser('printActivity')
parser.add_argument(dest='index', type=int, help='refresh')
args = vars(parser.parse_args(args))
index = args['index']
activity = self.activities[index]
pprint(activity)
| do_printActivity | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, args):\n parser = CommandArgumentParser('printActivity')\n parser.add_argument(dest='index', type=int, help='refresh')\n args = vars(parser.pars... |
def x(self, args):
parser = CommandArgumentParser('printPolicy')
args = vars(parser.parse_args(args))
policy = self.client.describe_policies(AutoScalingGroupName=self.
scalingGroup)
pprint(policy)
| do_printPolicy | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, args):\n parser = CommandArgumentParser('printPolicy')\n args = vars(parser.parse_args(args))\n policy = self.client.describe_policies(AutoScali... |
def x(self, args):
parser = CommandArgumentParser('rebootInstance')
parser.add_argument(dest='instance', help='instance index or name')
args = vars(parser.parse_args(args))
instanceId = args['instance']
try:
index = int(instanceId)
instances = self.scalingGroupDescription['AutoScalingGroups'][0][
'Instances']
instanceId = instances[index]
except ValueError:
pass
client = AwsConnectionFactory.getEc2Client()
client.reboot_instances(InstanceIds=[instanceId['InstanceId']])
| do_rebootInstance | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, args):\n parser = CommandArgumentParser('rebootInstance')\n parser.add_argument(dest='instance', help='instance index or name')\n args = vars(pa... |
def x(self, args):
parser = CommandArgumentParser('run')
parser.add_argument('-R', '--replace-key', dest='replaceKey', default=
False, action='store_true', help=
"Replace the host's key. This is useful when AWS recycles an IP address you've seen before."
)
parser.add_argument('-Y', '--keyscan', dest='keyscan', default=False,
action='store_true', help=
"Perform a keyscan to avoid having to say 'yes' for a new host. Implies -R."
)
parser.add_argument('-ii', '--ignore-host-key', dest='ignore-host-key',
default=False, action='store_true', help='Ignore host key')
parser.add_argument('-ne', '--no-echo', dest='no-echo', default=False,
action='store_true', help='Do not echo command')
parser.add_argument(dest='command', nargs='+', help=
'Command to run on all hosts.')
parser.add_argument('-v', dest='verbosity', default=0, action=VAction,
nargs='?', help='Verbosity. The more instances, the more verbose')
parser.add_argument('-j', dest='jobs', type=int, default=1, help=
'Number of hosts to contact in parallel')
parser.add_argument('-s', dest='skip', type=int, default=0, help=
'Skip this many hosts')
parser.add_argument('-m', dest='macro', default=False, action=
'store_true', help=
'{command} is a series of macros to execute, not the actual command to run on the host'
)
args = vars(parser.parse_args(args))
replaceKey = args['replaceKey']
keyscan = args['keyscan']
verbosity = args['verbosity']
jobs = args['jobs']
skip = args['skip']
ignoreHostKey = args['ignore-host-key']
noEcho = args['no-echo']
instances = self.scalingGroupDescription['AutoScalingGroups'][0][
'Instances']
instances = instances[skip:]
if args['macro']:
if len(args['command']) > 1:
print('Only one macro may be specified with the -m switch.')
return
else:
macro = args['command'][0]
print('Macro:{}'.format(macro))
command = Config.config['ssh-macros'][macro]
else:
command = ' '.join(args['command'])
Parallel(n_jobs=jobs)(delayed(ssh)(instance['InstanceId'], 0, [],
replaceKey, keyscan, False, verbosity, command, ignoreHostKey=
ignoreHostKey, echoCommand=not noEcho, name='{}:{}: '.format(
instance['index'], instance['InstanceId'])) for instance in instances)
| do_run | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, args):\n parser = CommandArgumentParser('run')\n parser.add_argument('-R', '--replace-key', dest='replaceKey', default=\n False, action='sto... |
def x(self, args):
parser = CommandArgumentParser('ssh')
parser.add_argument(dest='instance', help='instance index or name')
parser.add_argument('-a', '--address-number', default='0', dest=
'interface-number', help='instance id of the instance to ssh to')
parser.add_argument('-ii', '--ignore-host-key', dest='ignore-host-key',
default=False, action='store_true', help='Ignore host key')
parser.add_argument('-ne', '--no-echo', dest='no-echo', default=False,
action='store_true', help='Do not echo command')
parser.add_argument('-L', dest='forwarding', nargs='*', help=
'port forwarding string of the form: {localport}:{host-visible-to-instance}:{remoteport} or {port}'
)
parser.add_argument('-R', '--replace-key', dest='replaceKey', default=
False, action='store_true', help=
"Replace the host's key. This is useful when AWS recycles an IP address you've seen before."
)
parser.add_argument('-Y', '--keyscan', dest='keyscan', default=False,
action='store_true', help=
"Perform a keyscan to avoid having to say 'yes' for a new host. Implies -R."
)
parser.add_argument('-B', '--background', dest='background', default=
False, action='store_true', help=
'Run in the background. (e.g., forward an ssh session and then do other stuff in aws-shell).'
)
parser.add_argument('-v', dest='verbosity', default=0, action=VAction,
nargs='?', help='Verbosity. The more instances, the more verbose')
parser.add_argument('-m', dest='macro', default=False, action=
'store_true', help=
'{command} is a series of macros to execute, not the actual command to run on the host'
)
parser.add_argument(dest='command', nargs='*', help=
'Command to run on all hosts.')
args = vars(parser.parse_args(args))
interfaceNumber = int(args['interface-number'])
forwarding = args['forwarding']
replaceKey = args['replaceKey']
keyscan = args['keyscan']
background = args['background']
verbosity = args['verbosity']
ignoreHostKey = args['ignore-host-key']
noEcho = args['no-echo']
target = args['instance']
try:
index = int(args['instance'])
instances = self.scalingGroupDescription['AutoScalingGroups'][0][
'Instances']
instance = instances[index]
target = instance['InstanceId']
except ValueError:
pass
if args['macro']:
if len(args['command']) > 1:
print('Only one macro may be specified with the -m switch.')
return
else:
macro = args['command'][0]
print('Macro:{}'.format(macro))
command = Config.config['ssh-macros'][macro]
else:
command = ' '.join(args['command'])
ssh(target, interfaceNumber, forwarding, replaceKey, keyscan,
background, verbosity, command, ignoreHostKey=ignoreHostKey,
echoCommand=not noEcho)
| do_ssh | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, args):\n parser = CommandArgumentParser('ssh')\n parser.add_argument(dest='instance', help='instance index or name')\n parser.add_argument('-a',... |
def x(self, args):
parser = CommandArgumentParser('startInstance')
parser.add_argument(dest='instance', help='instance index or name')
args = vars(parser.parse_args(args))
instanceId = args['instance']
force = args['force']
try:
index = int(instanceId)
instances = self.scalingGroupDescription['AutoScalingGroups'][0][
'Instances']
instanceId = instances[index]
except ValueError:
pass
client = AwsConnectionFactory.getEc2Client()
client.start_instances(InstanceIds=[instanceId['InstanceId']])
| do_startInstance | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, args):\n parser = CommandArgumentParser('startInstance')\n parser.add_argument(dest='instance', help='instance index or name')\n args = vars(par... |
def x(self, args):
parser = CommandArgumentParser('stopInstance')
parser.add_argument(dest='instance', help='instance index or name')
parser.add_argument('-f', '--force', action='store_true', dest='force',
help='instance index or name')
args = vars(parser.parse_args(args))
instanceId = args['instance']
force = args['force']
try:
index = int(instanceId)
instances = self.scalingGroupDescription['AutoScalingGroups'][0][
'Instances']
instanceId = instances[index]
except ValueError:
pass
client = AwsConnectionFactory.getEc2Client()
client.stop_instances(InstanceIds=[instanceId['InstanceId']], Force=force)
| do_stopInstance | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, args):\n parser = CommandArgumentParser('stopInstance')\n parser.add_argument(dest='instance', help='instance index or name')\n parser.add_argum... |
def x(self, args):
parser = CommandArgumentParser('terminateInstance')
parser.add_argument(dest='instance', help='instance index or name')
args = vars(parser.parse_args(args))
instanceId = args['instance']
try:
index = int(instanceId)
instances = self.scalingGroupDescription['AutoScalingGroups'][0][
'Instances']
instanceId = instances[index]
except ValueError:
pass
client = AwsConnectionFactory.getEc2Client()
client.terminate_instances(InstanceIds=[instanceId['InstanceId']])
self.do_printInstances('-r')
| do_terminateInstance | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, args):\n parser = CommandArgumentParser('terminateInstance')\n parser.add_argument(dest='instance', help='instance index or name')\n args = vars... |
def x(sequencepath, report, refseq_database, num_threads=12, start=time.time()
):
files = find_files(sequencepath)
file_dict = filer(files)
printtime('Using MASH to determine genera of samples', start)
genus_dict = find_genus(file_dict, refseq_database, threads=num_threads)
file_records = fasta_records(file_dict)
printtime('Collecting basic quality metrics', start)
contig_len_dict, gc_dict = fasta_stats(file_dict, file_records)
contig_dist_dict = find_contig_distribution(contig_len_dict)
longest_contig_dict = find_largest_contig(contig_len_dict)
genome_length_dict = find_genome_length(contig_len_dict)
num_contigs_dict = find_num_contigs(contig_len_dict)
n50_dict = find_n50(contig_len_dict, genome_length_dict)
n75_dict = find_n75(contig_len_dict, genome_length_dict)
n90_dict = find_n90(contig_len_dict, genome_length_dict)
l50_dict = find_l50(contig_len_dict, genome_length_dict)
l75_dict = find_l75(contig_len_dict, genome_length_dict)
l90_dict = find_l90(contig_len_dict, genome_length_dict)
printtime('Using prodigal to calculate number of ORFs in each sample',
start)
orf_file_dict = predict_orfs(file_dict, num_threads=num_threads)
orf_dist_dict = find_orf_distribution(orf_file_dict)
if report:
reporter(gc_dict, contig_dist_dict, longest_contig_dict,
genome_length_dict, num_contigs_dict, n50_dict, n75_dict,
n90_dict, l50_dict, l75_dict, l90_dict, orf_dist_dict,
genus_dict, sequencepath)
printtime('Features extracted!', start)
return (gc_dict, contig_dist_dict, longest_contig_dict,
genome_length_dict, num_contigs_dict, n50_dict, n75_dict, n90_dict,
l50_dict, l75_dict, l90_dict, orf_dist_dict)
| main | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(sequencepath, report, refseq_database, num_threads=12, start=time.time()\n ):\n files = find_files(sequencepath)\n file_dict = filer(files)\n print... |
def x(sequencepath):
files = sorted(glob(os.path.join(sequencepath, '*.fa*')))
return files
| find_files | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(sequencepath):\n files = sorted(glob(os.path.join(sequencepath, '*.fa*')))\n return files\n",
"role": "user"
},
{
"content": "find_files",
... |
def x(filelist):
filedict = dict()
for seqfile in filelist:
strainname = os.path.splitext(os.path.basename(seqfile))[0]
filedict[strainname] = seqfile
return filedict
| filer | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(filelist):\n filedict = dict()\n for seqfile in filelist:\n strainname = os.path.splitext(os.path.basename(seqfile))[0]\n filedict[strainna... |
def x(files):
file_records = dict()
for file_name, fasta in files.items():
record_dict = SeqIO.to_dict(SeqIO.parse(fasta, 'fasta'))
file_records[file_name] = record_dict
return file_records
| fasta_records | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(files):\n file_records = dict()\n for file_name, fasta in files.items():\n record_dict = SeqIO.to_dict(SeqIO.parse(fasta, 'fasta'))\n file_... |
def x(files, database, threads=12):
genus_dict = dict()
tmpdir = str(time.time()).split('.')[-1]
if not os.path.isdir(tmpdir):
os.makedirs(tmpdir)
for file_name, fasta in files.items():
mash.screen(database, fasta, threads=threads, w='', i=0.95,
output_file=os.path.join(tmpdir, 'screen.tab'))
screen_output = mash.read_mash_screen(os.path.join(tmpdir,
'screen.tab'))
try:
os.remove(os.path.join(tmpdir, 'screen.tab'))
except IOError:
pass
try:
genus = screen_output[0].query_id.split('/')[-3]
if genus == 'Shigella':
genus = 'Escherichia'
genus_dict[file_name] = genus
except IndexError:
genus_dict[file_name] = 'NA'
shutil.rmtree(tmpdir)
return genus_dict
| find_genus | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(files, database, threads=12):\n genus_dict = dict()\n tmpdir = str(time.time()).split('.')[-1]\n if not os.path.isdir(tmpdir):\n os.makedirs(tm... |
def x(files, records):
contig_len_dict = dict()
gc_dict = dict()
for file_name in files:
contig_lengths = list()
fasta_sequence = str()
for contig, record in records[file_name].items():
contig_lengths.append(len(record.seq))
fasta_sequence += record.seq
contig_len_dict[file_name] = sorted(contig_lengths, reverse=True)
gc_dict[file_name] = float('{:0.2f}'.format(GC(fasta_sequence)))
return contig_len_dict, gc_dict
| fasta_stats | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(files, records):\n contig_len_dict = dict()\n gc_dict = dict()\n for file_name in files:\n contig_lengths = list()\n fasta_sequence = st... |
def x(contig_lengths_dict):
contig_len_dist_dict = dict()
for file_name, contig_lengths in contig_lengths_dict.items():
over_1000000 = 0
over_500000 = 0
over_100000 = 0
over_50000 = 0
over_10000 = 0
over_5000 = 0
other = 0
for contig_length in contig_lengths:
if contig_length > 1000000:
over_1000000 += 1
elif contig_length > 500000:
over_500000 += 1
elif contig_length > 100000:
over_100000 += 1
elif contig_length > 50000:
over_50000 += 1
elif contig_length > 10000:
over_10000 += 1
elif contig_length > 5000:
over_5000 += 1
else:
other += 1
contig_len_dist_dict[file_name] = (over_1000000, over_500000,
over_100000, over_50000, over_10000, over_5000, other)
return contig_len_dist_dict
| find_contig_distribution | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(contig_lengths_dict):\n contig_len_dist_dict = dict()\n for file_name, contig_lengths in contig_lengths_dict.items():\n over_1000000 = 0\n ... |
def x(contig_lengths_dict):
longest_contig_dict = dict()
for file_name, contig_lengths in contig_lengths_dict.items():
longest_contig_dict[file_name] = contig_lengths[0]
return longest_contig_dict
| find_largest_contig | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(contig_lengths_dict):\n longest_contig_dict = dict()\n for file_name, contig_lengths in contig_lengths_dict.items():\n longest_contig_dict[file_na... |
def x(contig_lengths_dict):
genome_length_dict = dict()
for file_name, contig_lengths in contig_lengths_dict.items():
genome_length_dict[file_name] = sum(contig_lengths)
return genome_length_dict
| find_genome_length | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(contig_lengths_dict):\n genome_length_dict = dict()\n for file_name, contig_lengths in contig_lengths_dict.items():\n genome_length_dict[file_name... |
def x(contig_lengths_dict):
num_contigs_dict = dict()
for file_name, contig_lengths in contig_lengths_dict.items():
num_contigs_dict[file_name] = len(contig_lengths)
return num_contigs_dict
| find_num_contigs | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(contig_lengths_dict):\n num_contigs_dict = dict()\n for file_name, contig_lengths in contig_lengths_dict.items():\n num_contigs_dict[file_name] = ... |
def x(contig_lengths_dict, genome_length_dict):
n50_dict = dict()
for file_name, contig_lengths in contig_lengths_dict.items():
currentlength = 0
for contig_length in contig_lengths:
currentlength += contig_length
if currentlength >= genome_length_dict[file_name] * 0.5:
n50_dict[file_name] = contig_length
break
return n50_dict
| find_n50 | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(contig_lengths_dict, genome_length_dict):\n n50_dict = dict()\n for file_name, contig_lengths in contig_lengths_dict.items():\n currentlength = 0\... |
def x(contig_lengths_dict, genome_length_dict):
n75_dict = dict()
for file_name, contig_lengths in contig_lengths_dict.items():
currentlength = 0
for contig_length in contig_lengths:
currentlength += contig_length
if currentlength >= genome_length_dict[file_name] * 0.75:
n75_dict[file_name] = contig_length
break
return n75_dict
| find_n75 | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(contig_lengths_dict, genome_length_dict):\n n75_dict = dict()\n for file_name, contig_lengths in contig_lengths_dict.items():\n currentlength = 0\... |
def x(contig_lengths_dict, genome_length_dict):
n90_dict = dict()
for file_name, contig_lengths in contig_lengths_dict.items():
currentlength = 0
for contig_length in contig_lengths:
currentlength += contig_length
if currentlength >= genome_length_dict[file_name] * 0.95:
n90_dict[file_name] = contig_length
break
return n90_dict
| find_n90 | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(contig_lengths_dict, genome_length_dict):\n n90_dict = dict()\n for file_name, contig_lengths in contig_lengths_dict.items():\n currentlength = 0\... |
def x(contig_lengths_dict, genome_length_dict):
l50_dict = dict()
for file_name, contig_lengths in contig_lengths_dict.items():
currentlength = 0
currentcontig = 0
for contig_length in contig_lengths:
currentlength += contig_length
currentcontig += 1
if currentlength >= genome_length_dict[file_name] * 0.5:
l50_dict[file_name] = currentcontig
break
return l50_dict
| find_l50 | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(contig_lengths_dict, genome_length_dict):\n l50_dict = dict()\n for file_name, contig_lengths in contig_lengths_dict.items():\n currentlength = 0\... |
def x(contig_lengths_dict, genome_length_dict):
l75_dict = dict()
for file_name, contig_lengths in contig_lengths_dict.items():
currentlength = 0
currentcontig = 0
for contig_length in contig_lengths:
currentlength += contig_length
currentcontig += 1
if currentlength >= genome_length_dict[file_name] * 0.75:
l75_dict[file_name] = currentcontig
break
return l75_dict
| find_l75 | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(contig_lengths_dict, genome_length_dict):\n l75_dict = dict()\n for file_name, contig_lengths in contig_lengths_dict.items():\n currentlength = 0\... |
def x(contig_lengths_dict, genome_length_dict):
l90_dict = dict()
for file_name, contig_lengths in contig_lengths_dict.items():
currentlength = 0
currentcontig = 0
for contig_length in contig_lengths:
currentlength += contig_length
currentcontig += 1
if currentlength >= genome_length_dict[file_name] * 0.9:
l90_dict[file_name] = currentcontig
break
return l90_dict
| find_l90 | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(contig_lengths_dict, genome_length_dict):\n l90_dict = dict()\n for file_name, contig_lengths in contig_lengths_dict.items():\n currentlength = 0\... |
def x(file_dict, num_threads=1):
orf_file_dict = dict()
prodigallist = list()
for file_name, file_path in file_dict.items():
results = os.path.splitext(file_path)[0] + '.sco'
prodigal = ['prodigal', '-i', file_path, '-o', results, '-f', 'sco']
if not os.path.isfile(results):
prodigallist.append(prodigal)
orf_file_dict[file_name] = results
pool = multiprocessing.Pool(processes=num_threads)
pool.map(run_prodigal, prodigallist)
pool.close()
pool.join()
return orf_file_dict
| predict_orfs | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(file_dict, num_threads=1):\n orf_file_dict = dict()\n prodigallist = list()\n for file_name, file_path in file_dict.items():\n results = os.pat... |
def x(orf_file_dict):
orf_dist_dict = dict()
for file_name, orf_report in orf_file_dict.items():
total_orfs = 0
over_3000 = 0
over_1000 = 0
over_500 = 0
other = 0
with open(orf_report, 'r') as orfreport:
for line in orfreport:
if line.startswith('>'):
contig, start, stop, direction = line.split('_')
size = int(stop) - int(start)
total_orfs += 1
if size > 3000:
over_3000 += 1
elif size > 1000:
over_1000 += 1
elif size > 500:
over_500 += 1
else:
other += 1
orf_dist_dict[file_name
] = total_orfs, over_3000, over_1000, over_500, other
try:
os.remove(orf_report)
except IOError:
pass
return orf_dist_dict
| find_orf_distribution | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(orf_file_dict):\n orf_dist_dict = dict()\n for file_name, orf_report in orf_file_dict.items():\n total_orfs = 0\n over_3000 = 0\n ov... |
def x(gc_dict, contig_dist_dict, longest_contig_dict, genome_length_dict,
num_contigs_dict, n50_dict, n75_dict, n90_dict, l50_dict, l75_dict,
l90_dict, orf_dist_dict, genus_dict, sequencepath):
data = """SampleName,TotalLength,NumContigs,LongestContig,Contigs>1000000,Contigs>500000,Contigs>100000,Contigs>50000,Contigs>10000,Contigs>5000,Contigs<5000,TotalORFs,ORFs>3000,ORFs>1000,ORFs>500,ORFs<500,N50,N75,N90,L50,L75,L90,GC%,Genus
"""
with open(os.path.join(sequencepath, 'extracted_features.csv'), 'w'
) as feature_report:
for file_name in sorted(longest_contig_dict):
data += (
"""{name},{totlen},{numcontigs},{longestcontig},{over_106},{over_56},{over_105},{over_55},{over_104},{over_54},{under_54},{tORFS},{ORF33},{ORF13},{ORF52}, {ORF11},{n50},{n75},{n90},{l50},{l75},{l90},{gc},{genus}
"""
.format(name=file_name, totlen=genome_length_dict[file_name
], numcontigs=num_contigs_dict[file_name], longestcontig=
longest_contig_dict[file_name], over_106=contig_dist_dict[
file_name][0], over_56=contig_dist_dict[file_name][1],
over_105=contig_dist_dict[file_name][2], over_55=
contig_dist_dict[file_name][3], over_104=contig_dist_dict[
file_name][4], over_54=contig_dist_dict[file_name][5],
under_54=contig_dist_dict[file_name][6], tORFS=
orf_dist_dict[file_name][0], ORF33=orf_dist_dict[file_name]
[1], ORF13=orf_dist_dict[file_name][2], ORF52=orf_dist_dict
[file_name][3], ORF11=orf_dist_dict[file_name][4], n50=
n50_dict[file_name], n75=n75_dict[file_name], n90=n90_dict[
file_name], l50=l50_dict[file_name], l75=l75_dict[file_name
], l90=l90_dict[file_name], gc=gc_dict[file_name], genus=
genus_dict[file_name]))
feature_report.write(data)
| reporter | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(gc_dict, contig_dist_dict, longest_contig_dict, genome_length_dict,\n num_contigs_dict, n50_dict, n75_dict, n90_dict, l50_dict, l75_dict,\n l90_dict, orf... |
def x(sequencepath, report, refseq_database):
main(sequencepath, report, refseq_database, num_threads=multiprocessing
.cpu_count())
| cli | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(sequencepath, report, refseq_database):\n main(sequencepath, report, refseq_database, num_threads=multiprocessing\n .cpu_count())\n",
"role": "us... |
def x(data, mapping, nacode=None, nastate=False):
if nastate:
if nacode is None:
nacode = len(mapping)
if isinstance(mapping, list):
m = mapping
e = range(len(mapping))
elif isinstance(mapping, dict):
m = list(mapping.values())
e = list(mapping.keys())
else:
raise Exception("'data' must be list-of-list or dict.")
return grouplabelencode_loop(data, m, e, nacode=nacode)
| grouplabelencode | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(data, mapping, nacode=None, nastate=False):\n if nastate:\n if nacode is None:\n nacode = len(mapping)\n if isinstance(mapping, list):\... |
def x(rows, row_headers_count_value=0):
count = 0
if rows:
for row in rows:
if exclude_empty_values(row[:row_headers_count_value]):
break
count += 1
if len(rows) == count:
count = 1
return [r[row_headers_count_value:] for r in rows[:count]]
| get_csv_col_headers | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(rows, row_headers_count_value=0):\n count = 0\n if rows:\n for row in rows:\n if exclude_empty_values(row[:row_headers_count_value]):\n... |
def x(rows, partial_headers, column_headers_count=1):
result = [''] * (len(rows) - column_headers_count)
for i_index in range(0, len(partial_headers)):
for k_index in range(0, len(partial_headers[i_index])):
if not partial_headers[i_index][k_index] and i_index - 1 >= 0:
for t_index in range(i_index - 1, -1, -1):
partial_value = partial_headers[t_index][k_index]
if partial_value:
partial_headers[i_index][k_index] = partial_value
break
result[i_index] = ' '.join(map(str, partial_headers[i_index]))
return result
| populate_csv_headers | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(rows, partial_headers, column_headers_count=1):\n result = [''] * (len(rows) - column_headers_count)\n for i_index in range(0, len(partial_headers)):\n ... |
def x(rows, row_headers_count_value=0, column_headers_count=1):
partial_headers = []
if row_headers_count_value:
for k_index in range(0, len(rows) - column_headers_count):
header = rows[k_index + column_headers_count][:
row_headers_count_value]
partial_headers.append(remove_list_duplicates(force_list(header)))
populated_headers = populate_csv_headers(rows, partial_headers,
column_headers_count)
return populated_headers
| get_row_headers | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(rows, row_headers_count_value=0, column_headers_count=1):\n partial_headers = []\n if row_headers_count_value:\n for k_index in range(0, len(rows)... |
def x(rows, row_header=0, column_header=0, limit_column=0):
return [row[row_header:limit_column] for row in rows[column_header:]]
| retrieve_csv_data | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(rows, row_header=0, column_header=0, limit_column=0):\n return [row[row_header:limit_column] for row in rows[column_header:]]\n",
"role": "user"
},
... |
def x(path_to_file, **kwargs):
result = []
encoding = kwargs.get('encoding', 'utf-8')
delimiter = kwargs.get('delimiter', ',')
dialect = kwargs.get('dialect', csv.excel)
_, _ext = path_to_file.split('.', 1)
try:
file = codecs.open(path_to_file, 'r', encoding)
items_file = io.TextIOWrapper(file, encoding=encoding)
result = list(csv.reader(items_file, delimiter=delimiter, dialect=
dialect))
items_file.close()
file.close()
except Exception as ex:
result = []
logger.error('Fail parsing csv to list of rows - {}'.format(ex))
return result
| csv_tolist | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(path_to_file, **kwargs):\n result = []\n encoding = kwargs.get('encoding', 'utf-8')\n delimiter = kwargs.get('delimiter', ',')\n dialect = kwargs.g... |
def x(path_to_file, **kwargs):
result = collections.OrderedDict()
encoding = kwargs.get('encoding', 'utf-8')
formatting_info = '.xlsx' not in path_to_file
count = 0
with xlrd.open_workbook(path_to_file, encoding_override=encoding,
formatting_info=formatting_info) as _excelfile:
for sheet_name_raw in _excelfile.sheet_names():
sheet_name = sheet_name_raw or 'sheet{}'.format(count)
result[sheet_name] = []
xl_sheet = _excelfile.sheet_by_name(sheet_name_raw)
for row_idx in range(0, xl_sheet.nrows):
col_data = []
for col_idx in range(0, xl_sheet.ncols):
cell_obj = xl_sheet.cell(row_idx, col_idx)
merged_info = is_merged(xl_sheet, row_idx, col_idx)
if not cell_obj.value and merged_info:
cell_obj = search_mergedcell_value(xl_sheet,
merged_info[1])
col_data.append(cell_obj.value if cell_obj else '')
else:
col_data.append(cell_obj.value)
result[sheet_name].append(col_data)
count += 1
return result
| excel_todictlist | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(path_to_file, **kwargs):\n result = collections.OrderedDict()\n encoding = kwargs.get('encoding', 'utf-8')\n formatting_info = '.xlsx' not in path_to_... |
def x(xl_sheet, merged_range):
for search_row_idx in range(merged_range[0], merged_range[1]):
for search_col_idx in range(merged_range[2], merged_range[3]):
if xl_sheet.cell(search_row_idx, search_col_idx).value:
return xl_sheet.cell(search_row_idx, search_col_idx)
return False
| search_mergedcell_value | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(xl_sheet, merged_range):\n for search_row_idx in range(merged_range[0], merged_range[1]):\n for search_col_idx in range(merged_range[2], merged_range... |
def x(sheet, row, column):
for cell_range in sheet.merged_cells:
row_low, row_high, column_low, column_high = cell_range
if row in range(row_low, row_high) and column in range(column_low,
column_high):
if (column_high - column_low < sheet.ncols - 1 and row_high -
row_low < sheet.nrows - 1):
return True, cell_range
return False
| is_merged | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(sheet, row, column):\n for cell_range in sheet.merged_cells:\n row_low, row_high, column_low, column_high = cell_range\n if row in range(row_l... |
def x(headers):
result = [''] * len(headers[0])
values = [''] * len(headers)
for k_index in range(0, len(headers)):
for i_index in range(0, len(headers[k_index])):
if headers[k_index][i_index]:
values[k_index] = normalizer(str(headers[k_index][i_index]))
if len(exclude_empty_values(result)) > i_index:
result[i_index] += '-{}'.format(values[k_index])
else:
result[i_index] += str(values[k_index])
return result
| populate_headers | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(headers):\n result = [''] * len(headers[0])\n values = [''] * len(headers)\n for k_index in range(0, len(headers)):\n for i_index in range(0, l... |
def x(rows, limits=None):
limits = [None, None] if limits is None else limits
if len(exclude_empty_values(limits)) == 2:
upper_limit = limits[0]
lower_limit = limits[1]
elif len(exclude_empty_values(limits)) == 1:
upper_limit = limits[0]
lower_limit = row_iter_limiter(rows, 1, -1, 1)
else:
upper_limit = row_iter_limiter(rows, 0, 1, 0)
lower_limit = row_iter_limiter(rows, 1, -1, 1)
return rows[upper_limit:lower_limit]
| row_csv_limiter | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(rows, limits=None):\n limits = [None, None] if limits is None else limits\n if len(exclude_empty_values(limits)) == 2:\n upper_limit = limits[0]\n... |
def x(rows, begin_row, way, c_value):
limit = None
for index in range(begin_row, len(rows)):
if not len(exclude_empty_values(rows[way * index])) == 1:
limit = way * index + c_value if way * index + c_value not in [
way * len(rows), 0] else None
break
return limit
| row_iter_limiter | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(rows, begin_row, way, c_value):\n limit = None\n for index in range(begin_row, len(rows)):\n if not len(exclude_empty_values(rows[way * index])) =... |
def x(rows):
result = []
for row in rows:
check_empty = len(exclude_empty_values(row)) > 1
check_set = len(set(exclude_empty_values(row))) > 1
check_last_allready = result and result[-1] == row
if check_empty and check_set and not check_last_allready:
result.append(row)
return result
| csv_row_cleaner | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(rows):\n result = []\n for row in rows:\n check_empty = len(exclude_empty_values(row)) > 1\n check_set = len(set(exclude_empty_values(row))... |
def x(rows):
result = [[] for x in range(0, len(rows))]
for i_index in range(0, len(rows[0])):
partial_values = []
for x_row in rows:
partial_values.append(x_row[i_index] if len(x_row) > i_index else
'')
colum_rows = exclude_empty_values(partial_values)
if len(colum_rows) > len(rows) / 5:
for index in range(0, len(rows)):
result[index].append(rows[index][i_index] if len(rows[index
]) > i_index else '')
return result
| csv_column_cleaner | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(rows):\n result = [[] for x in range(0, len(rows))]\n for i_index in range(0, len(rows[0])):\n partial_values = []\n for x_row in rows:\n ... |
def x(csv_data, c_headers=None, r_headers=None):
if r_headers:
result = {}
for k_index in range(0, len(csv_data)):
if r_headers[k_index]:
result[r_headers[k_index]] = collections.OrderedDict(zip(
c_headers, csv_data[k_index]))
else:
result = []
for k_index in range(0, len(csv_data)):
result.append(collections.OrderedDict(zip(c_headers, csv_data[
k_index])))
result = [result]
return result
| csv_dict_format | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(csv_data, c_headers=None, r_headers=None):\n if r_headers:\n result = {}\n for k_index in range(0, len(csv_data)):\n if r_headers[k... |
def x(csv_data, c_headers=None, r_headers=None):
result = []
real_num_header = len(force_list(r_headers[0])) if r_headers else 0
result.append([''] * real_num_header + c_headers)
for k_index in range(0, len(csv_data)):
if r_headers:
result.append(list(itertools.chain([r_headers[k_index]],
csv_data[k_index])))
else:
result.append(csv_data[k_index])
return result
| csv_array_clean_format | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(csv_data, c_headers=None, r_headers=None):\n result = []\n real_num_header = len(force_list(r_headers[0])) if r_headers else 0\n result.append([''] * ... |
def x(csv_data, c_headers=None, r_headers=None, rows=None, **kwargs):
result = None
c_headers = [] if c_headers is None else c_headers
r_headers = [] if r_headers is None else r_headers
rows = [] if rows is None else rows
result_format = kwargs.get('result_format', ARRAY_RAW_FORMAT)
if result_format == DICT_FORMAT:
result = csv_dict_format(csv_data, c_headers, r_headers)
elif result_format == ARRAY_RAW_FORMAT:
result = rows
elif result_format == ARRAY_CLEAN_FORMAT:
result = csv_array_clean_format(csv_data, c_headers, r_headers)
else:
result = None
if result and result_format < DICT_FORMAT:
result = [result]
return result
| csv_format | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(csv_data, c_headers=None, r_headers=None, rows=None, **kwargs):\n result = None\n c_headers = [] if c_headers is None else c_headers\n r_headers = [] ... |
def x(csv_filepath, **kwargs):
callbacks = {'to_list': csv_tolist, 'row_csv_limiter': row_csv_limiter,
'csv_row_cleaner': csv_row_cleaner, 'row_headers_count':
row_headers_count, 'get_col_header': get_csv_col_headers,
'get_row_headers': get_row_headers, 'populate_headers':
populate_headers, 'csv_column_header_cleaner':
csv_column_header_cleaner, 'csv_column_cleaner': csv_column_cleaner,
'retrieve_csv_data': retrieve_csv_data}
callbacks.update(kwargs.get('alt_callbacks', {}))
rows = kwargs.get('rows', [])
if not rows:
rows = callbacks.get('to_list')(csv_filepath, **kwargs)
if not rows:
msg = 'Empty rows obtained from {}'.format(csv_filepath)
logger.warning(msg)
raise ValueError(msg)
rows = callbacks.get('row_csv_limiter')(rows, kwargs.get('limits', [
None, None]))
rows = callbacks.get('csv_row_cleaner')(rows)
rows = callbacks.get('csv_column_cleaner')(rows)
num_row_headers = callbacks.get('row_headers_count')(rows)
c_headers_raw = callbacks.get('get_col_header')(rows, num_row_headers)
r_headers = callbacks.get('get_row_headers')(rows, num_row_headers, len
(c_headers_raw))
c_headers_dirty = callbacks.get('populate_headers')(c_headers_raw) if len(
c_headers_raw) > 1 else c_headers_raw[0]
c_headers = callbacks.get('csv_column_header_cleaner')(c_headers_dirty)
csv_data = callbacks.get('retrieve_csv_data')(rows, column_header=len(
c_headers_raw), row_header=num_row_headers, limit_column=len(
c_headers) - len(c_headers_dirty) or None)
if csv_data:
assert len(c_headers) == len(csv_data[0])
if r_headers:
assert len(r_headers) == len(csv_data)
kwargs.pop('rows', None)
result = csv_format(csv_data, c_headers, r_headers, rows, **kwargs)
return result
| csv_to_dict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(csv_filepath, **kwargs):\n callbacks = {'to_list': csv_tolist, 'row_csv_limiter': row_csv_limiter,\n 'csv_row_cleaner': csv_row_cleaner, 'row_headers... |
def x(excel_filepath, encapsulate_filepath=False, **kwargs):
result = {}
try:
callbacks = {'to_dictlist': excel_todictlist}
callbacks.update(kwargs.get('alt_callbacks', {}))
excel_data = callbacks.get('to_dictlist')(excel_filepath, **kwargs)
for sheet in excel_data.keys():
try:
kwargs['rows'] = excel_data.get(sheet, [])
result[sheet] = csv_to_dict(excel_filepath, **kwargs)
except Exception as ex:
logger.error('Fail to parse sheet {} - {}'.format(sheet, ex))
result[sheet] = []
continue
if encapsulate_filepath:
result = {excel_filepath: result}
except Exception as ex:
msg = 'Fail transform excel to dict - {}'.format(ex)
logger.error(msg, excel_filepath=excel_filepath)
return result
| excel_to_dict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(excel_filepath, encapsulate_filepath=False, **kwargs):\n result = {}\n try:\n callbacks = {'to_dictlist': excel_todictlist}\n callbacks.upd... |
def x(self, force_notify=None, use_email=None, use_sms=None,
email_body_template=None, **kwargs):
email_sent = None
sms_sent = None
use_email = use_email or getattr(settings, 'EMAIL_ENABLED', False)
use_sms = use_sms or getattr(settings, 'TWILIO_ENABLED', False)
if force_notify or self._notify_on_condition(**kwargs):
if use_email:
email_body_template = (email_body_template or self.
email_body_template) + self.email_footer_template
email_sent = self.send_email(email_body_template=
email_body_template, **kwargs)
if use_sms:
sms_sent = self.send_sms(**kwargs)
self.post_notification_actions(email_sent=email_sent, sms_sent=
sms_sent, **kwargs)
return True if email_sent or sms_sent else False
| notify | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, force_notify=None, use_email=None, use_sms=None,\n email_body_template=None, **kwargs):\n email_sent = None\n sms_sent = None\n use_email = u... |
def x(self, test_message=None, **kwargs):
if test_message:
return True
else:
return self.enabled and self.notify_on_condition(**kwargs)
| notify_on_condition | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, test_message=None, **kwargs):\n if test_message:\n return True\n else:\n return self.enabled and self.notify_on_condition(**kwargs)\n... |
def x(self):
if not self._notification_enabled:
self._notification_enabled = self.notification_model.enabled
return self._notification_enabled
| enabled | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if not self._notification_enabled:\n self._notification_enabled = self.notification_model.enabled\n return self._notification_enabled\n",... |
def x(self):
NotificationModel = django_apps.get_model('edc_notification.notification')
site_notifications.get(self.name)
try:
notification_model = NotificationModel.objects.get(name=self.name)
except ObjectDoesNotExist:
site_notifications.update_notification_list()
notification_model = NotificationModel.objects.get(name=self.name)
return notification_model
| notification_model | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n NotificationModel = django_apps.get_model('edc_notification.notification')\n site_notifications.get(self.name)\n try:\n notification_m... |
def x(self, instance=None, test_message=None, **kwargs):
protocol_name = django_apps.get_app_config('edc_protocol').protocol_name
test_message = test_message or self.test_message
template_options = dict(name=self.name, protocol_name=protocol_name,
display_name=self.display_name, email_from=self.email_from,
test_subject_line=(self.email_test_subject_line if test_message else
'').strip(), test_body_line=self.email_test_body_line if
test_message else '', test_line=self.sms_test_line if test_message else
'', message_datetime=get_utcnow(), message_reference='')
if 'subject_identifier' not in template_options:
try:
template_options.update(subject_identifier=instance.
subject_identifier)
except AttributeError:
pass
if 'site_name' not in template_options:
try:
template_options.update(site_name=instance.site.name.title())
except AttributeError:
pass
return template_options
| get_template_options | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, instance=None, test_message=None, **kwargs):\n protocol_name = django_apps.get_app_config('edc_protocol').protocol_name\n test_message = test_messa... |
def x(self):
sms_recipients = []
UserProfile = django_apps.get_model('edc_auth.UserProfile')
for user_profile in UserProfile.objects.filter(user__is_active=True,
user__is_staff=True):
try:
user_profile.sms_notifications.get(name=self.name)
except ObjectDoesNotExist:
pass
else:
if user_profile.mobile:
sms_recipients.append(user_profile.mobile)
return sms_recipients
| sms_recipients | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n sms_recipients = []\n UserProfile = django_apps.get_model('edc_auth.UserProfile')\n for user_profile in UserProfile.objects.filter(user__is_a... |
def x(permissions):
permissions = {key.lower(): value for key, value in permissions.iteritems()
}
return signing.dumps(permissions, compress=True)
| sign_filter_permissions | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(permissions):\n permissions = {key.lower(): value for key, value in permissions.iteritems()\n }\n return signing.dumps(permissions, compress=True)... |
def x(sign, dotted_model_name):
permissions = signing.loads(sign)
return permissions.get(dotted_model_name, [])
| unsign_filters_and_actions | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(sign, dotted_model_name):\n permissions = signing.loads(sign)\n return permissions.get(dotted_model_name, [])\n",
"role": "user"
},
{
"conten... |
def x(obj1, obj2):
Comparable.log(obj1, obj2, '==')
equality = obj1.equality(obj2)
Comparable.log(obj1, obj2, '==', result=equality)
return equality
| equal | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(obj1, obj2):\n Comparable.log(obj1, obj2, '==')\n equality = obj1.equality(obj2)\n Comparable.log(obj1, obj2, '==', result=equality)\n return equal... |
def x(obj1, obj2):
Comparable.log(obj1, obj2, '%')
similarity = obj1.similarity(obj2)
Comparable.log(obj1, obj2, '%', result=similarity)
return similarity
| similar | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(obj1, obj2):\n Comparable.log(obj1, obj2, '%')\n similarity = obj1.similarity(obj2)\n Comparable.log(obj1, obj2, '%', result=similarity)\n return s... |
def x(self, *args, **kwargs):
kwargs = {k: v for k, v in kwargs.items() if v is not None}
kwargs = OrderedDict(sorted(kwargs.items()))
args_repr = ', '.join(repr(arg) for arg in args)
kwargs_repr = ', '.join(k + '=' + repr(v) for k, v in kwargs.items())
if args_repr and kwargs_repr:
kwargs_repr = ', ' + kwargs_repr
name = self.__class__.__name__
return '{}({}{})'.format(name, args_repr, kwargs_repr)
| repr | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, *args, **kwargs):\n kwargs = {k: v for k, v in kwargs.items() if v is not None}\n kwargs = OrderedDict(sorted(kwargs.items()))\n args_repr = ', ... |
def x(self, other):
cname = self.__class__.__name__
for aname in self.attributes:
try:
attr1 = getattr(self, aname)
attr2 = getattr(other, aname)
except AttributeError as error:
logging.debug('%s.%s: %s', cname, aname, error)
return False
self.log(attr1, attr2, '==', cname=cname, aname=aname)
eql = attr1 == attr2
self.log(attr1, attr2, '==', cname=cname, aname=aname, result=eql)
if not eql:
return False
return True
| equality | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, other):\n cname = self.__class__.__name__\n for aname in self.attributes:\n try:\n attr1 = getattr(self, aname)\n attr... |
def x(self, other):
sim = self.Similarity()
total = 0.0
cname = self.__class__.__name__
for aname, weight in self.attributes.items():
attr1 = getattr(self, aname, None)
attr2 = getattr(other, aname, None)
self.log(attr1, attr2, '%', cname=cname, aname=aname)
if attr1 is None and attr2 is None:
self.log(attr1, attr2, '%', cname=cname, aname=aname, result=
'attributes are both None')
continue
if not all((isinstance(attr1, Comparable), isinstance(attr2,
Comparable))):
self.log(attr1, attr2, '%', cname=cname, aname=aname, result=
'attributes not Comparable')
total += weight
continue
attr_sim = attr1 % attr2
self.log(attr1, attr2, '%', cname=cname, aname=aname, result=attr_sim)
sim += attr_sim * weight
total += weight
if total:
sim *= 1.0 / total
return sim
| similarity | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, other):\n sim = self.Similarity()\n total = 0.0\n cname = self.__class__.__name__\n for aname, weight in self.attributes.items():\n at... |
def x(self, value=None):
if value is None:
value = 0.0
return Similarity(value, threshold=self.threshold)
| Similarity | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, value=None):\n if value is None:\n value = 0.0\n return Similarity(value, threshold=self.threshold)\n",
"role": "user"
},
{
"con... |
def x(obj1, obj2, sym, cname=None, aname=None, result=None):
fmt = '{o1} {sym} {o2} : {r}'
if cname or aname:
assert cname and aname
fmt = '{c}.{a}: ' + fmt
if result is None:
result = '...'
fmt = _Indent.indent(fmt)
_Indent.more()
else:
_Indent.less()
fmt = _Indent.indent(fmt)
msg = fmt.format(o1=repr(obj1), o2=repr(obj2), c=cname, a=aname, sym=
sym, r=result)
logging.info(msg)
| log | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(obj1, obj2, sym, cname=None, aname=None, result=None):\n fmt = '{o1} {sym} {o2} : {r}'\n if cname or aname:\n assert cname and aname\n fmt ... |
def x(self, dep_file, dep_rule):
dst_base = dep_file.split(os.path.join(dep_rule.address.repo, dep_rule.
address.path), 1)[-1]
if self.params['strip_prefix']:
dst_base = dep_file.split(self.params['strip_prefix'], 1)[-1]
return os.path.join(self.address.repo, self.address.path, self.params[
'prefix'].lstrip('/'), dst_base.lstrip('/'))
| translate_path | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dep_file, dep_rule):\n dst_base = dep_file.split(os.path.join(dep_rule.address.repo, dep_rule.\n address.path), 1)[-1]\n if self.params['str... |
def x(self):
for dep in self.subgraph.successors(self.address):
dep_rule = self.subgraph.node[dep]['target_obj']
for dep_file in dep_rule.output_files:
yield self.translate_path(dep_file, dep_rule).lstrip('/')
| output_files | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n for dep in self.subgraph.successors(self.address):\n dep_rule = self.subgraph.node[dep]['target_obj']\n for dep_file in dep_rule.outp... |
def x(ruletype, **kwargs):
try:
ruleclass = TYPE_MAP[ruletype]
except KeyError:
raise error.InvalidRule('Unrecognized rule type: %s' % ruletype)
try:
return ruleclass(**kwargs)
except TypeError:
log.error('BADNESS. ruletype: %s, data: %s', ruletype, kwargs)
raise
| new | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(ruletype, **kwargs):\n try:\n ruleclass = TYPE_MAP[ruletype]\n except KeyError:\n raise error.InvalidRule('Unrecognized rule type: %s' % ru... |
def x(human):
if not isinstance(human, list) or len(human) == 0:
return None
if human[0].lower() == 'yes':
return True
return False
| human_to_boolean | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(human):\n if not isinstance(human, list) or len(human) == 0:\n return None\n if human[0].lower() == 'yes':\n return True\n return False\... |
def x(self, data, *args, **kwargs):
with tempfile.TemporaryDirectory() as tmpdirname:
self.options.setdefault('package_build_dir', tmpdirname)
crawler = CrawlerScript(self.options)
crawler.crawl(data, self.spider, *args, **kwargs)
output_file = os.path.join(self.options['package_build_dir'],
'source', 'moear.mobi')
with open(output_file, 'rb') as fh:
content = fh.read()
return content
| generate | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, data, *args, **kwargs):\n with tempfile.TemporaryDirectory() as tmpdirname:\n self.options.setdefault('package_build_dir', tmpdirname)\n ... |
def x(seconds):
for i in range(seconds, 0, -1):
sys.stdout.write('%02d' % i)
time.sleep(1)
sys.stdout.write('\x08\x08')
sys.stdout.flush()
sys.stdout.flush()
| countdown | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(seconds):\n for i in range(seconds, 0, -1):\n sys.stdout.write('%02d' % i)\n time.sleep(1)\n sys.stdout.write('\\x08\\x08')\n sy... |
def x(self, location):
temp_dir = tempfile.mkdtemp('-export', 'pip-')
self.unpack(temp_dir)
if os.path.exists(location):
rmtree(location)
try:
call_subprocess([self.cmd, 'export', location], cwd=temp_dir,
filter_stdout=self._filter, show_stdout=False)
finally:
rmtree(temp_dir)
| export | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, location):\n temp_dir = tempfile.mkdtemp('-export', 'pip-')\n self.unpack(temp_dir)\n if os.path.exists(location):\n rmtree(location)\n ... |
def x(self, email):
resp = self._call(endpoint='single', data={'email': email})
return VerifiedEmail(email, resp['result'])
| verify | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, email):\n resp = self._call(endpoint='single', data={'email': email})\n return VerifiedEmail(email, resp['result'])\n",
"role": "user"
},
{... |
def x(self, emails):
resp = self._call(endpoint='bulk', data={'input_location': '1', 'input':
'\n'.join(emails)})
return Job(resp['job_id'])
| create_job | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, emails):\n resp = self._call(endpoint='bulk', data={'input_location': '1', 'input':\n '\\n'.join(emails)})\n return Job(resp['job_id'])\n",
... |
def x(self, job_id):
resp = self._call(endpoint='status', data={'job_id': job_id})
map = {'id': 'job_id', 'status': 'status_code', 'type': 'type_code'}
job_status_args = {map.get(k, k): v for k, v in resp.items()}
return JobStatus(**job_status_args)
| check_job | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, job_id):\n resp = self._call(endpoint='status', data={'job_id': job_id})\n map = {'id': 'job_id', 'status': 'status_code', 'type': 'type_code'}\n ... |
def x(self, job_id):
resp = self._call(endpoint='download', data={'job_id': job_id})
Row = namedtuple('Row', ['email', 'result_text_code'])
for line in resp:
row = Row(*line.decode('utf-8').split(','))
yield VerifiedEmail.from_text_code(row.email, row.result_text_code)
| results | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, job_id):\n resp = self._call(endpoint='download', data={'job_id': job_id})\n Row = namedtuple('Row', ['email', 'result_text_code'])\n for line i... |
def x(self, job_id):
warnings.warn(
'Use results generator method instead of retrieve_job which returns a list'
, UserWarning)
return list(self.results(job_id))
| retrieve_job | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, job_id):\n warnings.warn(\n 'Use results generator method instead of retrieve_job which returns a list'\n , UserWarning)\n return lis... |
def x(self):
resp = self._call(endpoint='account')
return Account(resp['credits'], resp['jobs_completed'], resp[
'jobs_processing'])
| account | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n resp = self._call(endpoint='account')\n return Account(resp['credits'], resp['jobs_completed'], resp[\n 'jobs_processing'])\n",
"role... |
def x(self):
if self._cached_access_token is not None:
return self._cached_access_token
resp = self._request(endpoint='access_token', data={'grant_type':
'client_credentials', 'scope': 'basic user'}, auth=(self.
api_username, self.api_key))
self._cached_access_token = resp['access_token']
return self._cached_access_token
| access_token | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self._cached_access_token is not None:\n return self._cached_access_token\n resp = self._request(endpoint='access_token', data={'grant... |
def x(self, endpoint, data=None):
data = {} if data is None else data
try:
data['access_token'] = self.access_token()
return self._request(endpoint, data)
except AccessTokenExpired:
self._cached_access_token = None
data['access_token'] = self.access_token()
return self._request(endpoint, data)
| call | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, endpoint, data=None):\n data = {} if data is None else data\n try:\n data['access_token'] = self.access_token()\n return self._reques... |
def x(self, endpoint, data, auth=None):
url = '{}/{}'.format(self.base_url, endpoint)
response = requests.post(url, data, auth=auth)
return self._handle_response(response)
| request | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, endpoint, data, auth=None):\n url = '{}/{}'.format(self.base_url, endpoint)\n response = requests.post(url, data, auth=auth)\n return self._hand... |
def x(response):
if not response.ok:
raise NeverBounceAPIError(response)
if response.headers.get('Content-Type') == 'application/octet-stream':
return response.iter_lines()
try:
resp = response.json()
except ValueError:
raise InvalidResponseError(
'Failed to handle the response content-type {}.'.format(
response.headers.get('Content-Type')))
if 'success' in resp and not resp['success']:
if 'msg' in resp and resp['msg'] == 'Authentication failed':
raise AccessTokenExpired
else:
raise NeverBounceAPIError(response)
return resp
| handle_response | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(response):\n if not response.ok:\n raise NeverBounceAPIError(response)\n if response.headers.get('Content-Type') == 'application/octet-stream':\n ... |
def x(table, post_processors):
table_result = table
for processor in post_processors:
table_result = processor(table_result)
return table_result
| post_process | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(table, post_processors):\n table_result = table\n for processor in post_processors:\n table_result = processor(table_result)\n return table_res... |
def x(cls, full=False):
divider_double = '=' * 80
divider_single = '-' * 80
description = cls.__doc__
message = []
message.append(divider_double)
message.append(cls.__name__ + ':')
message.append(description)
if full and cls.post_processors(cls):
message.append(divider_single)
message.append('Post processors:')
message.append(divider_single)
for processor in cls.post_processors(cls):
message.append('>' + ' ' * 3 + processor.__name__ + ':')
message.append(' ' * 4 + processor.__doc__)
message.append('')
message.append(divider_double)
message.append('')
for line in message:
print(line)
| describe | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, full=False):\n divider_double = '=' * 80\n divider_single = '-' * 80\n description = cls.__doc__\n message = []\n message.append(divider_do... |
def x(cls):
for processor in cls.post_processors(cls):
yield {'name': processor.__name__, 'description': processor.__doc__,
'processor': processor}
| describe_processors | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls):\n for processor in cls.post_processors(cls):\n yield {'name': processor.__name__, 'description': processor.__doc__,\n 'processor': p... |
def x(cls):
dependencies = []
try:
dependencies += cls.source.dependencies
except AttributeError:
pass
for processor in cls.post_processors(cls):
try:
assert isinstance(processor.dependencies, list
), '{}.dependencies must be a list'.format(processor.__name__)
dependencies += processor.dependencies
except AttributeError:
pass
return dependencies
| dependencies | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls):\n dependencies = []\n try:\n dependencies += cls.source.dependencies\n except AttributeError:\n pass\n for processor in cls.pos... |
def x(self):
return [self.source, self.output, self.kwargs, self.post_processors]
| get_settings_list | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n return [self.source, self.output, self.kwargs, self.post_processors]\n",
"role": "user"
},
{
"content": "get_settings_list",
"role"... |
def x(self):
depencency_hashes = [dep.get_hash() for dep in self.dep()]
sl = inspect.getsourcelines
hash_sources = [sl(self.__class__), self.args, self.kwargs, *
depencency_hashes]
hash_input = pickle.dumps(hash_sources)
return hashlib.md5(hash_input).hexdigest()
| get_hash | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n depencency_hashes = [dep.get_hash() for dep in self.dep()]\n sl = inspect.getsourcelines\n hash_sources = [sl(self.__class__), self.args, sel... |
def x(self, filename, extention, settings_list=None):
cached_name = '_'.join([filename, self.get_hash()])
return '.'.join([cached_name, extention])
| get_cached_filename | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, filename, extention, settings_list=None):\n cached_name = '_'.join([filename, self.get_hash()])\n return '.'.join([cached_name, extention])\n",
... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.