source_code stringlengths 45 7.24k | function_name stringlengths 1 111 | messages list |
|---|---|---|
def x(connection, without_connection_test):
database.set_connection(connection=connection)
if not without_connection_test:
test_connection(connection)
| connection | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(connection, without_connection_test):\n database.set_connection(connection=connection)\n if not without_connection_test:\n test_connection(connection)\n",
"role": "user"
},
{
"content": "connection",
"role": "assistant"
}
] |
def x(initialParams, laterParams):
err = []
for param in initialParams:
if param in laterParams:
if param not in VARIABLE_PARAMS and initialParams[param
] != laterParams[param]:
err.append(
'\tParam %r initial value %r differs from later value %r' %
(param, initialParams[param], laterParams[param]))
else:
err.append(
'\t%r found in initial parameters, not found in later parameters'
% param)
for param in laterParams:
if param not in initialParams:
err.append(
'\t%r found in later parameters, not seen in initial parameters'
% param)
return 'Summary of differences:\n%s' % '\n'.join(err) if err else None
| checkCompatibleParams | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(initialParams, laterParams):\n err = []\n for param in initialParams:\n if param in laterParams:\n if param not in VARIABLE_PARAMS and initialParams[param\n ] != laterParams[param]:\n err.append(\n '\\tParam %r initial value %r differs from later value %r' %\n (param, initialParams[param], laterParams[param]))\n else:\n err.append(\n '\\t%r found in initial parameters, not found in later parameters'\n % param)\n for param in laterParams:\n if param not in initialParams:\n err.append(\n '\\t%r found in later parameters, not seen in initial parameters'\n % param)\n return 'Summary of differences:\\n%s' % '\\n'.join(err) if err else None\n",
"role": "user"
},
{
"content": "checkCompatibleParams",
"role": "assistant"
}
] |
def x(self):
src = self._source or ''
coll = self._collection or ''
desc = self._description or ''
l = [self._id, src, coll, self._name, ','.join(sorted(self._genes)), desc]
return l
| to_list | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n src = self._source or ''\n coll = self._collection or ''\n desc = self._description or ''\n l = [self._id, src, coll, self._name, ','.join(sorted(self._genes)), desc]\n return l\n",
"role": "user"
},
{
"content": "to_list",
"role": "assistant"
}
] |
def x(cls, l):
id_ = l[0]
name = l[3]
genes = l[4].split(',')
src = l[1] or None
coll = l[2] or None
desc = l[5] or None
return cls(id_, name, genes, src, coll, desc)
| from_list | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, l):\n id_ = l[0]\n name = l[3]\n genes = l[4].split(',')\n src = l[1] or None\n coll = l[2] or None\n desc = l[5] or None\n return cls(id_, name, genes, src, coll, desc)\n",
"role": "user"
},
{
"content": "from_list",
"role": "assistant"
}
] |
def x(primer, seq):
offsets = []
seq = seq.upper()
primer = primer.upper()
primerLen = len(primer)
discarded = 0
offset = seq.find(primer)
while offset > -1:
offsets.append(discarded + offset)
seq = seq[offset + primerLen:]
discarded += offset + primerLen
offset = seq.find(primer)
return offsets
| findPrimer | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(primer, seq):\n offsets = []\n seq = seq.upper()\n primer = primer.upper()\n primerLen = len(primer)\n discarded = 0\n offset = seq.find(primer)\n while offset > -1:\n offsets.append(discarded + offset)\n seq = seq[offset + primerLen:]\n discarded += offset + primerLen\n offset = seq.find(primer)\n return offsets\n",
"role": "user"
},
{
"content": "findPrimer",
"role": "assistant"
}
] |
def x(primer, seq):
forward = findPrimer(primer, seq)
reverse = findPrimer(reverse_complement(primer), seq)
return forward, reverse
| findPrimerBidi | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(primer, seq):\n forward = findPrimer(primer, seq)\n reverse = findPrimer(reverse_complement(primer), seq)\n return forward, reverse\n",
"role": "user"
},
{
"content": "findPrimerBidi",
"role": "assistant"
}
] |
def x(primer, seq):
forward, reverse = findPrimerBidi(primer, seq)
if forward:
start = forward[-1] + len(primer)
end = len(seq)
for offset in reverse:
if offset >= start:
end = offset
break
else:
start = 0
end = reverse[0] if reverse else len(seq)
return start, end
| findPrimerBidiLimits | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(primer, seq):\n forward, reverse = findPrimerBidi(primer, seq)\n if forward:\n start = forward[-1] + len(primer)\n end = len(seq)\n for offset in reverse:\n if offset >= start:\n end = offset\n break\n else:\n start = 0\n end = reverse[0] if reverse else len(seq)\n return start, end\n",
"role": "user"
},
{
"content": "findPrimerBidiLimits",
"role": "assistant"
}
] |
def x(names):
def x(name):
count = 0
for ch in name:
if ch in string.digits:
count += 1
else:
break
return 0 if count == 0 else int(name[0:count])
return sorted(names, key=lambda name: numericPrefix(basename(name)))
| numericallySortFilenames | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(names):\n\n def x(name):\n count = 0\n for ch in name:\n if ch in string.digits:\n count += 1\n else:\n break\n return 0 if count == 0 else int(name[0:count])\n return sorted(names, key=lambda name: numericPrefix(basename(name)))\n",
"role": "user"
},
{
"content": "numericallySortFilenames",
"role": "assistant"
}
] |
def x(fileNameOrHandle, mode='r'):
if isinstance(fileNameOrHandle, six.string_types):
if fileNameOrHandle.endswith('.gz'):
if six.PY3:
yield gzip.open(fileNameOrHandle, mode='rt', encoding='UTF-8')
else:
yield gzip.GzipFile(fileNameOrHandle)
elif fileNameOrHandle.endswith('.bz2'):
if six.PY3:
yield bz2.open(fileNameOrHandle, mode='rt', encoding='UTF-8')
else:
yield bz2.BZ2File(fileNameOrHandle)
else:
with open(fileNameOrHandle) as fp:
yield fp
else:
yield fileNameOrHandle
| asHandle | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fileNameOrHandle, mode='r'):\n if isinstance(fileNameOrHandle, six.string_types):\n if fileNameOrHandle.endswith('.gz'):\n if six.PY3:\n yield gzip.open(fileNameOrHandle, mode='rt', encoding='UTF-8')\n else:\n yield gzip.GzipFile(fileNameOrHandle)\n elif fileNameOrHandle.endswith('.bz2'):\n if six.PY3:\n yield bz2.open(fileNameOrHandle, mode='rt', encoding='UTF-8')\n else:\n yield bz2.BZ2File(fileNameOrHandle)\n else:\n with open(fileNameOrHandle) as fp:\n yield fp\n else:\n yield fileNameOrHandle\n",
"role": "user"
},
{
"content": "asHandle",
"role": "assistant"
}
] |
def x(s, convertToZeroBased=False):
result = set()
for _range in s.split(','):
match = _rangeRegex.match(_range)
if match:
start, end = match.groups()
start = int(start)
if end is None:
end = start
else:
end = int(end)
if start > end:
start, end = end, start
if convertToZeroBased:
result.update(range(start - 1, end))
else:
result.update(range(start, end + 1))
else:
raise ValueError(
'Illegal range %r. Ranges must single numbers or number-number.'
% _range)
return result
| parseRangeString | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(s, convertToZeroBased=False):\n result = set()\n for _range in s.split(','):\n match = _rangeRegex.match(_range)\n if match:\n start, end = match.groups()\n start = int(start)\n if end is None:\n end = start\n else:\n end = int(end)\n if start > end:\n start, end = end, start\n if convertToZeroBased:\n result.update(range(start - 1, end))\n else:\n result.update(range(start, end + 1))\n else:\n raise ValueError(\n 'Illegal range %r. Ranges must single numbers or number-number.'\n % _range)\n return result\n",
"role": "user"
},
{
"content": "parseRangeString",
"role": "assistant"
}
] |
def x(nucleotides, prefix=''):
result = []
for offset in sorted(nucleotides):
result.append('%s%d: %s' % (prefix, offset, baseCountsToStr(
nucleotides[offset])))
return '\n'.join(result)
| nucleotidesToStr | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(nucleotides, prefix=''):\n result = []\n for offset in sorted(nucleotides):\n result.append('%s%d: %s' % (prefix, offset, baseCountsToStr(\n nucleotides[offset])))\n return '\\n'.join(result)\n",
"role": "user"
},
{
"content": "nucleotidesToStr",
"role": "assistant"
}
] |
def x(mesg, count, len1, len2=None):
def x(a, b):
return 100.0 * a / b if b else 0.0
if count == 0:
return '%s: %d' % (mesg, count)
else:
len2 = len2 or len1
if len1 == len2:
return '%s: %d/%d (%.2f%%)' % (mesg, count, len1, percentage(
count, len1))
else:
return (
'%s: %d/%d (%.2f%%) of sequence 1, %d/%d (%.2f%%) of sequence 2'
% (mesg, count, len1, percentage(count, len1), count, len2,
percentage(count, len2)))
| countPrint | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(mesg, count, len1, len2=None):\n\n def x(a, b):\n return 100.0 * a / b if b else 0.0\n if count == 0:\n return '%s: %d' % (mesg, count)\n else:\n len2 = len2 or len1\n if len1 == len2:\n return '%s: %d/%d (%.2f%%)' % (mesg, count, len1, percentage(\n count, len1))\n else:\n return (\n '%s: %d/%d (%.2f%%) of sequence 1, %d/%d (%.2f%%) of sequence 2'\n % (mesg, count, len1, percentage(count, len1), count, len2,\n percentage(count, len2)))\n",
"role": "user"
},
{
"content": "countPrint",
"role": "assistant"
}
] |
def x(client, script_dir, overwrite=True):
local_dir = os.path.join(genometools._root, 'data', 'gcloud', 'scripts')
match = _BUCKET_PAT.match(script_dir)
script_bucket = match.group(1)
script_prefix = match.group(2)
depth = len(local_dir.split(os.sep))
for root, dirs, files in os.walk(local_dir):
rel_path = '/'.join(root.split(os.sep)[depth:])
for f in files:
local_path = os.path.join(root, f)
if rel_path:
remote_path = '/'.join([script_prefix, rel_path, f])
else:
remote_path = '/'.join([script_prefix, f])
_LOGGER.info('Uploading "%s"...', remote_path)
storage.upload_file(client, script_bucket, local_path,
remote_path, overwrite=overwrite)
| upload_scripts | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(client, script_dir, overwrite=True):\n local_dir = os.path.join(genometools._root, 'data', 'gcloud', 'scripts')\n match = _BUCKET_PAT.match(script_dir)\n script_bucket = match.group(1)\n script_prefix = match.group(2)\n depth = len(local_dir.split(os.sep))\n for root, dirs, files in os.walk(local_dir):\n rel_path = '/'.join(root.split(os.sep)[depth:])\n for f in files:\n local_path = os.path.join(root, f)\n if rel_path:\n remote_path = '/'.join([script_prefix, rel_path, f])\n else:\n remote_path = '/'.join([script_prefix, f])\n _LOGGER.info('Uploading \"%s\"...', remote_path)\n storage.upload_file(client, script_bucket, local_path,\n remote_path, overwrite=overwrite)\n",
"role": "user"
},
{
"content": "upload_scripts",
"role": "assistant"
}
] |
def x(self, filename, scoreClass):
if filename.endswith('.json') or filename.endswith('.json.bz2'):
return JSONRecordsReader(filename, scoreClass)
else:
raise ValueError('Unknown DIAMOND record file suffix for file %r.' %
filename)
| getReader | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, filename, scoreClass):\n if filename.endswith('.json') or filename.endswith('.json.bz2'):\n return JSONRecordsReader(filename, scoreClass)\n else:\n raise ValueError('Unknown DIAMOND record file suffix for file %r.' %\n filename)\n",
"role": "user"
},
{
"content": "getReader",
"role": "assistant"
}
] |
def x(self):
reads = iter(self.reads)
first = True
for filename in self.filenames:
if first:
first = False
reader = self._reader
else:
reader = self._getReader(filename, self.scoreClass)
for readAlignments in reader.readAlignments(reads):
yield readAlignments
for read in reads:
yield ReadAlignments(read, [])
| iter | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n reads = iter(self.reads)\n first = True\n for filename in self.filenames:\n if first:\n first = False\n reader = self._reader\n else:\n reader = self._getReader(filename, self.scoreClass)\n for readAlignments in reader.readAlignments(reads):\n yield readAlignments\n for read in reads:\n yield ReadAlignments(read, [])\n",
"role": "user"
},
{
"content": "iter",
"role": "assistant"
}
] |
def x(self, title):
if self._subjectTitleToSubject is None:
if self._databaseFilename is None:
self._subjectTitleToSubject = SqliteIndex(self.
_sqliteDatabaseFilename, fastaDirectory=self.
_databaseDirectory, readClass=AAReadWithX)
else:
titles = {}
for read in FastaReads(self._databaseFilename, readClass=
AAReadWithX):
titles[read.id] = read
self._subjectTitleToSubject = titles
return self._subjectTitleToSubject[title]
| getSubjectSequence | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, title):\n if self._subjectTitleToSubject is None:\n if self._databaseFilename is None:\n self._subjectTitleToSubject = SqliteIndex(self.\n _sqliteDatabaseFilename, fastaDirectory=self.\n _databaseDirectory, readClass=AAReadWithX)\n else:\n titles = {}\n for read in FastaReads(self._databaseFilename, readClass=\n AAReadWithX):\n titles[read.id] = read\n self._subjectTitleToSubject = titles\n return self._subjectTitleToSubject[title]\n",
"role": "user"
},
{
"content": "getSubjectSequence",
"role": "assistant"
}
] |
def x(project):
urls = []
with ftputil.FTPHost(sra_host, sra_user, sra_password) as ftp_host:
download_paths = []
exp_dir = '/sra/sra-instant/reads/ByStudy/sra/SRP/%s/%s/' % (project
[:6], project)
ftp_host.chdir(exp_dir)
run_folders = ftp_host.listdir(ftp_host.curdir)
for folder in run_folders:
files = ftp_host.listdir(folder)
assert len(files) == 1
for f in files:
path = exp_dir + folder + '/' + f
urls.append(path)
return urls
| get_project_urls | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(project):\n urls = []\n with ftputil.FTPHost(sra_host, sra_user, sra_password) as ftp_host:\n download_paths = []\n exp_dir = '/sra/sra-instant/reads/ByStudy/sra/SRP/%s/%s/' % (project\n [:6], project)\n ftp_host.chdir(exp_dir)\n run_folders = ftp_host.listdir(ftp_host.curdir)\n for folder in run_folders:\n files = ftp_host.listdir(folder)\n assert len(files) == 1\n for f in files:\n path = exp_dir + folder + '/' + f\n urls.append(path)\n return urls\n",
"role": "user"
},
{
"content": "get_project_urls",
"role": "assistant"
}
] |
def x(rst, clk, rx_rdy, rx_vld, rx_dat, tx_rdy, tx_vld, tx_dat):
DATA_WIDTH = len(rx_dat)
NUM_STAGES = 3
stage_en = Signal(intbv(0)[NUM_STAGES:])
pipe_ctrl = pipeline_control(rst=rst, clk=clk, rx_vld=rx_vld, rx_rdy=
rx_rdy, tx_vld=tx_vld, tx_rdy=tx_rdy, stage_enable=stage_en)
s0_dat = Signal(intbv(0)[DATA_WIDTH:])
@always_seq(clk.posedge, reset=rst)
def x():
if stage_en[0]:
s0_dat.next = rx_dat
s1_dat = Signal(intbv(0)[DATA_WIDTH:])
@always_seq(clk.posedge, reset=rst)
def x():
if stage_en[1]:
s1_dat.next = ~s0_dat
s2_dat = Signal(intbv(0)[DATA_WIDTH:])
@always_seq(clk.posedge, reset=rst)
def x():
if stage_en[2]:
s2_dat.next = s1_dat + 1
@always_comb
def x():
tx_dat.next = s2_dat.signed()
return instances()
| twos_complement | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(rst, clk, rx_rdy, rx_vld, rx_dat, tx_rdy, tx_vld, tx_dat):\n DATA_WIDTH = len(rx_dat)\n NUM_STAGES = 3\n stage_en = Signal(intbv(0)[NUM_STAGES:])\n pipe_ctrl = pipeline_control(rst=rst, clk=clk, rx_vld=rx_vld, rx_rdy=\n rx_rdy, tx_vld=tx_vld, tx_rdy=tx_rdy, stage_enable=stage_en)\n s0_dat = Signal(intbv(0)[DATA_WIDTH:])\n\n @always_seq(clk.posedge, reset=rst)\n def x():\n if stage_en[0]:\n s0_dat.next = rx_dat\n s1_dat = Signal(intbv(0)[DATA_WIDTH:])\n\n @always_seq(clk.posedge, reset=rst)\n def x():\n if stage_en[1]:\n s1_dat.next = ~s0_dat\n s2_dat = Signal(intbv(0)[DATA_WIDTH:])\n\n @always_seq(clk.posedge, reset=rst)\n def x():\n if stage_en[2]:\n s2_dat.next = s1_dat + 1\n\n @always_comb\n def x():\n tx_dat.next = s2_dat.signed()\n return instances()\n",
"role": "user"
},
{
"content": "twos_complement",
"role": "assistant"
}
] |
def x():
desc = 'Generate a mapping of Entrez IDs to gene symbols.'
parser = cli.get_argument_parser(desc=desc)
parser.add_argument('-f', '--gene2acc-file', type=str, required=True,
help=textwrap.dedent(
""" Path of gene2accession.gz file (from
ftp://ftp.ncbi.nlm.nih.gov/gene/DATA), or a filtered version
thereof."""
))
parser.add_argument('-o', '--output-file', type=str, required=True,
help=textwrap.dedent(
""" Path of output file. If set to ``-``, print to ``stdout``,
and redirect logging messages to ``stderr``."""
))
parser.add_argument('-l', '--log-file', type=str, default=None, help=
'Path of log file. If not specified, print to stdout.')
parser.add_argument('-q', '--quiet', action='store_true', help=
'Suppress all output except warnings and errors.')
parser.add_argument('-v', '--verbose', action='store_true', help=
'Enable verbose output. Ignored if ``--quiet`` is specified.')
return parser
| get_argument_parser | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n desc = 'Generate a mapping of Entrez IDs to gene symbols.'\n parser = cli.get_argument_parser(desc=desc)\n parser.add_argument('-f', '--gene2acc-file', type=str, required=True,\n help=textwrap.dedent(\n \"\"\" Path of gene2accession.gz file (from\n ftp://ftp.ncbi.nlm.nih.gov/gene/DATA), or a filtered version\n thereof.\"\"\"\n ))\n parser.add_argument('-o', '--output-file', type=str, required=True,\n help=textwrap.dedent(\n \"\"\" Path of output file. If set to ``-``, print to ``stdout``,\n and redirect logging messages to ``stderr``.\"\"\"\n ))\n parser.add_argument('-l', '--log-file', type=str, default=None, help=\n 'Path of log file. If not specified, print to stdout.')\n parser.add_argument('-q', '--quiet', action='store_true', help=\n 'Suppress all output except warnings and errors.')\n parser.add_argument('-v', '--verbose', action='store_true', help=\n 'Enable verbose output. Ignored if ``--quiet`` is specified.')\n return parser\n",
"role": "user"
},
{
"content": "get_argument_parser",
"role": "assistant"
}
] |
def x(file_path, logger):
gene2acc = {}
with misc.smart_open_read(file_path, mode='rb', try_gzip=True) as fh:
reader = csv.reader(fh, dialect='excel-tab')
next(reader)
for i, l in enumerate(reader):
id_ = int(l[1])
symbol = l[15]
try:
gene2acc[id_].append(symbol)
except KeyError:
gene2acc[id_] = [symbol]
n = len(gene2acc)
for k, v in gene2acc.items():
symbols = sorted(set(v))
assert len(symbols) == 1
gene2acc[k] = symbols[0]
all_symbols = sorted(set(gene2acc.values()))
m = len(all_symbols)
logger.info('Found %d Entrez Gene IDs associated with %d gene symbols.',
n, m)
return gene2acc
| read_gene2acc | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(file_path, logger):\n gene2acc = {}\n with misc.smart_open_read(file_path, mode='rb', try_gzip=True) as fh:\n reader = csv.reader(fh, dialect='excel-tab')\n next(reader)\n for i, l in enumerate(reader):\n id_ = int(l[1])\n symbol = l[15]\n try:\n gene2acc[id_].append(symbol)\n except KeyError:\n gene2acc[id_] = [symbol]\n n = len(gene2acc)\n for k, v in gene2acc.items():\n symbols = sorted(set(v))\n assert len(symbols) == 1\n gene2acc[k] = symbols[0]\n all_symbols = sorted(set(gene2acc.values()))\n m = len(all_symbols)\n logger.info('Found %d Entrez Gene IDs associated with %d gene symbols.',\n n, m)\n return gene2acc\n",
"role": "user"
},
{
"content": "read_gene2acc",
"role": "assistant"
}
] |
def x(file_path, entrez2gene, logger):
with misc.smart_open_write(file_path, mode='wb') as ofh:
writer = csv.writer(ofh, dialect='excel-tab', lineterminator=os.linesep
)
for k in sorted(entrez2gene.keys(), key=lambda x: int(x)):
writer.writerow([k, entrez2gene[k]])
logger.info('Output written to file "%s".', file_path)
| write_entrez2gene | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(file_path, entrez2gene, logger):\n with misc.smart_open_write(file_path, mode='wb') as ofh:\n writer = csv.writer(ofh, dialect='excel-tab', lineterminator=os.linesep\n )\n for k in sorted(entrez2gene.keys(), key=lambda x: int(x)):\n writer.writerow([k, entrez2gene[k]])\n logger.info('Output written to file \"%s\".', file_path)\n",
"role": "user"
},
{
"content": "write_entrez2gene",
"role": "assistant"
}
] |
def x(args=None):
vinfo = sys.version_info
if not vinfo >= (2, 7):
raise SystemError(
'Python interpreter version >= 2.7 required, found %d.%d instead.'
% (vinfo.major, vinfo.minor))
if args is None:
parser = get_argument_parser()
args = parser.parse_args()
gene2acc_file = args.gene2acc_file
output_file = args.output_file
log_file = args.log_file
quiet = args.quiet
verbose = args.verbose
log_stream = sys.stdout
if output_file == '-':
log_stream = sys.stderr
logger = misc.get_logger(log_stream=log_stream, log_file=log_file,
quiet=quiet, verbose=verbose)
entrez2gene = read_gene2acc(gene2acc_file, logger)
write_entrez2gene(output_file, entrez2gene, logger)
return 0
| main | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(args=None):\n vinfo = sys.version_info\n if not vinfo >= (2, 7):\n raise SystemError(\n 'Python interpreter version >= 2.7 required, found %d.%d instead.'\n % (vinfo.major, vinfo.minor))\n if args is None:\n parser = get_argument_parser()\n args = parser.parse_args()\n gene2acc_file = args.gene2acc_file\n output_file = args.output_file\n log_file = args.log_file\n quiet = args.quiet\n verbose = args.verbose\n log_stream = sys.stdout\n if output_file == '-':\n log_stream = sys.stderr\n logger = misc.get_logger(log_stream=log_stream, log_file=log_file,\n quiet=quiet, verbose=verbose)\n entrez2gene = read_gene2acc(gene2acc_file, logger)\n write_entrez2gene(output_file, entrez2gene, logger)\n return 0\n",
"role": "user"
},
{
"content": "main",
"role": "assistant"
}
] |
def x(s):
abbrev1 = None
origS = s
if ' ' in s:
first, rest = s.split(' ', 1)
s = first.title() + ' ' + rest.lower()
else:
s = s.title()
if s in NAMES:
abbrev1 = s
elif s in ABBREV3_TO_ABBREV1:
abbrev1 = ABBREV3_TO_ABBREV1[s]
elif s in NAMES_TO_ABBREV1:
abbrev1 = NAMES_TO_ABBREV1[s]
else:
def x(target):
for abbrev1, codons in CODONS.items():
for codon in codons:
if codon == target:
return abbrev1
abbrev1 = findCodon(origS.upper())
if abbrev1:
return AminoAcid(NAMES[abbrev1], ABBREV3[abbrev1], abbrev1, CODONS[
abbrev1], PROPERTIES[abbrev1], PROPERTY_DETAILS[abbrev1],
PROPERTY_CLUSTERS[abbrev1])
| find | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(s):\n abbrev1 = None\n origS = s\n if ' ' in s:\n first, rest = s.split(' ', 1)\n s = first.title() + ' ' + rest.lower()\n else:\n s = s.title()\n if s in NAMES:\n abbrev1 = s\n elif s in ABBREV3_TO_ABBREV1:\n abbrev1 = ABBREV3_TO_ABBREV1[s]\n elif s in NAMES_TO_ABBREV1:\n abbrev1 = NAMES_TO_ABBREV1[s]\n else:\n\n def x(target):\n for abbrev1, codons in CODONS.items():\n for codon in codons:\n if codon == target:\n return abbrev1\n abbrev1 = findCodon(origS.upper())\n if abbrev1:\n return AminoAcid(NAMES[abbrev1], ABBREV3[abbrev1], abbrev1, CODONS[\n abbrev1], PROPERTIES[abbrev1], PROPERTY_DETAILS[abbrev1],\n PROPERTY_CLUSTERS[abbrev1])\n",
"role": "user"
},
{
"content": "find",
"role": "assistant"
}
] |
def x(sequence, propertyNames, propertyValues, missingAAValue):
propertyNames = sorted(map(str.lower, set(propertyNames)))
knownProperties = set()
for names in propertyValues.values():
knownProperties.update(names)
unknown = set(propertyNames) - knownProperties
if unknown:
raise ValueError('Unknown propert%s: %s.' % ('y' if len(unknown) ==
1 else 'ies', ', '.join(unknown)))
aas = sequence.sequence.upper()
result = {}
for propertyName in propertyNames:
result[propertyName] = values = []
append = values.append
for aa in aas:
try:
properties = propertyValues[aa]
except KeyError:
append(missingAAValue)
else:
append(properties[propertyName])
return result
| propertiesOrClustersForSequence | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(sequence, propertyNames, propertyValues, missingAAValue):\n propertyNames = sorted(map(str.lower, set(propertyNames)))\n knownProperties = set()\n for names in propertyValues.values():\n knownProperties.update(names)\n unknown = set(propertyNames) - knownProperties\n if unknown:\n raise ValueError('Unknown propert%s: %s.' % ('y' if len(unknown) ==\n 1 else 'ies', ', '.join(unknown)))\n aas = sequence.sequence.upper()\n result = {}\n for propertyName in propertyNames:\n result[propertyName] = values = []\n append = values.append\n for aa in aas:\n try:\n properties = propertyValues[aa]\n except KeyError:\n append(missingAAValue)\n else:\n append(properties[propertyName])\n return result\n",
"role": "user"
},
{
"content": "propertiesOrClustersForSequence",
"role": "assistant"
}
] |
def x(aaMatch, read1, read2, indent='', offsets=None):
match = aaMatch['match']
matchCount = match['matchCount']
gapMismatchCount = match['gapMismatchCount']
gapGapMismatchCount = match['gapGapMismatchCount']
nonGapMismatchCount = match['nonGapMismatchCount']
if offsets:
len1 = len2 = len(offsets)
else:
len1, len2 = map(len, (read1, read2))
result = []
append = result.append
append(countPrint('%sMatches' % indent, matchCount, len1, len2))
mismatchCount = (gapMismatchCount + gapGapMismatchCount +
nonGapMismatchCount)
append(countPrint('%sMismatches' % indent, mismatchCount, len1, len2))
append(countPrint('%s Not involving gaps (i.e., conflicts)' % indent,
nonGapMismatchCount, len1, len2))
append(countPrint('%s Involving a gap in one sequence' % indent,
gapMismatchCount, len1, len2))
append(countPrint('%s Involving a gap in both sequences' % indent,
gapGapMismatchCount, len1, len2))
for read, key in zip((read1, read2), ('read1', 'read2')):
append('%s Id: %s' % (indent, read.id))
length = len(read)
append('%s Length: %d' % (indent, length))
gapCount = len(aaMatch[key]['gapOffsets'])
append(countPrint('%s Gaps' % indent, gapCount, length))
if gapCount:
append('%s Gap locations (1-based): %s' % (indent, ', '.join
(map(lambda offset: str(offset + 1), sorted(aaMatch[key][
'gapOffsets'])))))
extraCount = aaMatch[key]['extraCount']
if extraCount:
append(countPrint('%s Extra nucleotides at end' % indent,
extraCount, length))
return '\n'.join(result)
| matchToString | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(aaMatch, read1, read2, indent='', offsets=None):\n match = aaMatch['match']\n matchCount = match['matchCount']\n gapMismatchCount = match['gapMismatchCount']\n gapGapMismatchCount = match['gapGapMismatchCount']\n nonGapMismatchCount = match['nonGapMismatchCount']\n if offsets:\n len1 = len2 = len(offsets)\n else:\n len1, len2 = map(len, (read1, read2))\n result = []\n append = result.append\n append(countPrint('%sMatches' % indent, matchCount, len1, len2))\n mismatchCount = (gapMismatchCount + gapGapMismatchCount +\n nonGapMismatchCount)\n append(countPrint('%sMismatches' % indent, mismatchCount, len1, len2))\n append(countPrint('%s Not involving gaps (i.e., conflicts)' % indent,\n nonGapMismatchCount, len1, len2))\n append(countPrint('%s Involving a gap in one sequence' % indent,\n gapMismatchCount, len1, len2))\n append(countPrint('%s Involving a gap in both sequences' % indent,\n gapGapMismatchCount, len1, len2))\n for read, key in zip((read1, read2), ('read1', 'read2')):\n append('%s Id: %s' % (indent, read.id))\n length = len(read)\n append('%s Length: %d' % (indent, length))\n gapCount = len(aaMatch[key]['gapOffsets'])\n append(countPrint('%s Gaps' % indent, gapCount, length))\n if gapCount:\n append('%s Gap locations (1-based): %s' % (indent, ', '.join\n (map(lambda offset: str(offset + 1), sorted(aaMatch[key][\n 'gapOffsets'])))))\n extraCount = aaMatch[key]['extraCount']\n if extraCount:\n append(countPrint('%s Extra nucleotides at end' % indent,\n extraCount, length))\n return '\\n'.join(result)\n",
"role": "user"
},
{
"content": "matchToString",
"role": "assistant"
}
] |
def x(read1, read2, gapChars='-', offsets=None):
matchCount = 0
gapMismatchCount = nonGapMismatchCount = gapGapMismatchCount = 0
read1ExtraCount = read2ExtraCount = 0
read1GapOffsets = []
read2GapOffsets = []
for offset, (a, b) in enumerate(zip_longest(read1.sequence.upper(),
read2.sequence.upper())):
if offsets is not None and offset not in offsets:
continue
if a is None:
assert b is not None
read2ExtraCount += 1
if b in gapChars:
read2GapOffsets.append(offset)
elif b is None:
read1ExtraCount += 1
if a in gapChars:
read1GapOffsets.append(offset)
elif a in gapChars:
read1GapOffsets.append(offset)
if b in gapChars:
gapGapMismatchCount += 1
read2GapOffsets.append(offset)
else:
gapMismatchCount += 1
elif b in gapChars:
gapMismatchCount += 1
read2GapOffsets.append(offset)
elif a == b:
matchCount += 1
else:
nonGapMismatchCount += 1
return {'match': {'matchCount': matchCount, 'gapMismatchCount':
gapMismatchCount, 'gapGapMismatchCount': gapGapMismatchCount,
'nonGapMismatchCount': nonGapMismatchCount}, 'read1': {'extraCount':
read1ExtraCount, 'gapOffsets': read1GapOffsets}, 'read2': {
'extraCount': read2ExtraCount, 'gapOffsets': read2GapOffsets}}
| compareAaReads | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(read1, read2, gapChars='-', offsets=None):\n matchCount = 0\n gapMismatchCount = nonGapMismatchCount = gapGapMismatchCount = 0\n read1ExtraCount = read2ExtraCount = 0\n read1GapOffsets = []\n read2GapOffsets = []\n for offset, (a, b) in enumerate(zip_longest(read1.sequence.upper(),\n read2.sequence.upper())):\n if offsets is not None and offset not in offsets:\n continue\n if a is None:\n assert b is not None\n read2ExtraCount += 1\n if b in gapChars:\n read2GapOffsets.append(offset)\n elif b is None:\n read1ExtraCount += 1\n if a in gapChars:\n read1GapOffsets.append(offset)\n elif a in gapChars:\n read1GapOffsets.append(offset)\n if b in gapChars:\n gapGapMismatchCount += 1\n read2GapOffsets.append(offset)\n else:\n gapMismatchCount += 1\n elif b in gapChars:\n gapMismatchCount += 1\n read2GapOffsets.append(offset)\n elif a == b:\n matchCount += 1\n else:\n nonGapMismatchCount += 1\n return {'match': {'matchCount': matchCount, 'gapMismatchCount':\n gapMismatchCount, 'gapGapMismatchCount': gapGapMismatchCount,\n 'nonGapMismatchCount': nonGapMismatchCount}, 'read1': {'extraCount':\n read1ExtraCount, 'gapOffsets': read1GapOffsets}, 'read2': {\n 'extraCount': read2ExtraCount, 'gapOffsets': read2GapOffsets}}\n",
"role": "user"
},
{
"content": "compareAaReads",
"role": "assistant"
}
] |
def x(colors, args):
result = defaultdict(set)
for colorInfo in colors:
readIds = colorInfo.split()
color = readIds.pop(0)
for readId in readIds:
if os.path.isfile(readId):
filename = readId
if args.fasta:
reads = FastaReads(filename)
else:
reads = FastqReads(filename)
for read in reads:
result[color].add(read.id)
else:
result[color].add(readId)
return result
| parseColors | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(colors, args):\n result = defaultdict(set)\n for colorInfo in colors:\n readIds = colorInfo.split()\n color = readIds.pop(0)\n for readId in readIds:\n if os.path.isfile(readId):\n filename = readId\n if args.fasta:\n reads = FastaReads(filename)\n else:\n reads = FastqReads(filename)\n for read in reads:\n result[color].add(read.id)\n else:\n result[color].add(readId)\n return result\n",
"role": "user"
},
{
"content": "parseColors",
"role": "assistant"
}
] |
def x(download_file, release=None):
if release is None:
release = get_latest_release()
url = (
'http://viewvc.geneontology.org/viewvc/GO-SVN/ontology-releases/%s/go-basic.obo'
% release)
misc.http_download(url, download_file)
| download_release | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(download_file, release=None):\n if release is None:\n release = get_latest_release()\n url = (\n 'http://viewvc.geneontology.org/viewvc/GO-SVN/ontology-releases/%s/go-basic.obo'\n % release)\n misc.http_download(url, download_file)\n",
"role": "user"
},
{
"content": "download_release",
"role": "assistant"
}
] |
def x():
with closing(requests.get(
'http://geneontology.org/ontology/go-basic.obo', stream=True)) as r:
for i, l in enumerate(r.iter_lines(decode_unicode=True)):
if i == 1:
assert l.split(':')[0] == 'data-version'
date = l.split('/')[-1]
break
return date
| get_current_ontology_date | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n with closing(requests.get(\n 'http://geneontology.org/ontology/go-basic.obo', stream=True)) as r:\n for i, l in enumerate(r.iter_lines(decode_unicode=True)):\n if i == 1:\n assert l.split(':')[0] == 'data-version'\n date = l.split('/')[-1]\n break\n return date\n",
"role": "user"
},
{
"content": "get_current_ontology_date",
"role": "assistant"
}
] |
def x(function, name):
try:
return TaskResult(name, True, None, function())
except Exception as exc:
return TaskResult(name, False, exc, None)
| execute | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(function, name):\n try:\n return TaskResult(name, True, None, function())\n except Exception as exc:\n return TaskResult(name, False, exc, None)\n",
"role": "user"
},
{
"content": "execute",
"role": "assistant"
}
] |
def x(value, arg):
try:
length = int(arg)
except ValueError:
return value
if len(value) > length:
return value[:length] + '...'
return value
| truncatechars | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(value, arg):\n try:\n length = int(arg)\n except ValueError:\n return value\n if len(value) > length:\n return value[:length] + '...'\n return value\n",
"role": "user"
},
{
"content": "truncatechars",
"role": "assistant"
}
] |
def x(desc, default_field_name='gene'):
parser = cli.get_argument_parser(desc=desc)
parser.add_argument('-a', '--annotation-file', default='-', type=str,
help=textwrap.dedent(
""" Path of Ensembl gene annotation file (in GTF format). The file
may be gzip'ed. If set to ``-``, read from ``stdin``."""
))
parser.add_argument('-o', '--output-file', required=True, type=str,
help=textwrap.dedent(
""" Path of output file. If set to ``-``, print to ``stdout``,
and redirect logging messages to ``stderr``."""
))
parser.add_argument('-c', '--chromosome-pattern', type=str, required=
False, default=None, help=textwrap.dedent(
""" Regular expression that chromosome names have to match. [None]
"""
))
cli.add_reporting_args(parser)
return parser
| get_gtf_argument_parser | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(desc, default_field_name='gene'):\n parser = cli.get_argument_parser(desc=desc)\n parser.add_argument('-a', '--annotation-file', default='-', type=str,\n help=textwrap.dedent(\n \"\"\" Path of Ensembl gene annotation file (in GTF format). The file\n may be gzip'ed. If set to ``-``, read from ``stdin``.\"\"\"\n ))\n parser.add_argument('-o', '--output-file', required=True, type=str,\n help=textwrap.dedent(\n \"\"\" Path of output file. If set to ``-``, print to ``stdout``,\n and redirect logging messages to ``stderr``.\"\"\"\n ))\n parser.add_argument('-c', '--chromosome-pattern', type=str, required=\n False, default=None, help=textwrap.dedent(\n \"\"\" Regular expression that chromosome names have to match. [None]\n \"\"\"\n ))\n cli.add_reporting_args(parser)\n return parser\n",
"role": "user"
},
{
"content": "get_gtf_argument_parser",
"role": "assistant"
}
] |
def x(url_map):
from flask import current_app
from . import current_jsonschemas
url_map.add(Rule('{0}/<path:path>'.format(current_app.config[
'JSONSCHEMAS_ENDPOINT']), endpoint=current_jsonschemas.get_schema,
host=current_app.config['JSONSCHEMAS_HOST']))
| jsonresolver_loader | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(url_map):\n from flask import current_app\n from . import current_jsonschemas\n url_map.add(Rule('{0}/<path:path>'.format(current_app.config[\n 'JSONSCHEMAS_ENDPOINT']), endpoint=current_jsonschemas.get_schema,\n host=current_app.config['JSONSCHEMAS_HOST']))\n",
"role": "user"
},
{
"content": "jsonresolver_loader",
"role": "assistant"
}
] |
def x(layouts):
layout = layouts[0].clone()
for l in layouts[1:]:
layout.files.update(l.files)
layout.domains.update(l.domains)
for k, v in l.entities.items():
if k not in layout.entities:
layout.entities[k] = v
else:
layout.entities[k].files.update(v.files)
return layout
| merge_layouts | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(layouts):\n layout = layouts[0].clone()\n for l in layouts[1:]:\n layout.files.update(l.files)\n layout.domains.update(l.domains)\n for k, v in l.entities.items():\n if k not in layout.entities:\n layout.entities[k] = v\n else:\n layout.entities[k].files.update(v.files)\n return layout\n",
"role": "user"
},
{
"content": "merge_layouts",
"role": "assistant"
}
] |
def x(self, entities=None, extensions=None, domains=None, regex_search=False):
if extensions is not None:
if isinstance(extensions, six.string_types):
extensions = [extensions]
extensions = '(' + '|'.join(extensions) + ')$'
if re.search(extensions, self.filename) is None:
return False
if domains is not None:
domains = listify(domains)
if not set(self.domains) & set(domains):
return False
if entities is not None:
for name, val in entities.items():
if (name not in self.tags) ^ (val is None):
return False
if val is None:
continue
def x(x):
patt = '%s' % x
if isinstance(x, (int, float)):
patt = '0*' + patt
if not regex_search:
patt = '^%s$' % patt
return patt
ent_patts = [make_patt(x) for x in listify(val)]
patt = '|'.join(ent_patts)
if re.search(patt, str(self.tags[name].value)) is None:
return False
return True
| matches | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, entities=None, extensions=None, domains=None, regex_search=False):\n if extensions is not None:\n if isinstance(extensions, six.string_types):\n extensions = [extensions]\n extensions = '(' + '|'.join(extensions) + ')$'\n if re.search(extensions, self.filename) is None:\n return False\n if domains is not None:\n domains = listify(domains)\n if not set(self.domains) & set(domains):\n return False\n if entities is not None:\n for name, val in entities.items():\n if (name not in self.tags) ^ (val is None):\n return False\n if val is None:\n continue\n\n def x(x):\n patt = '%s' % x\n if isinstance(x, (int, float)):\n patt = '0*' + patt\n if not regex_search:\n patt = '^%s$' % patt\n return patt\n ent_patts = [make_patt(x) for x in listify(val)]\n patt = '|'.join(ent_patts)\n if re.search(patt, str(self.tags[name].value)) is None:\n return False\n return True\n",
"role": "user"
},
{
"content": "matches",
"role": "assistant"
}
] |
def x(self):
keys = list(self.entities.keys())
replaced = []
for i, k in enumerate(keys):
if iskeyword(k):
replaced.append(k)
keys[i] = '%s_' % k
if replaced:
safe = [('%s_' % k) for k in replaced]
warnings.warn(
'Entity names cannot be reserved keywords when representing a File as a namedtuple. Replacing entities %s with safe versions %s.'
% (keys, safe))
entities = dict(zip(keys, self.entities.values()))
_File = namedtuple('File', 'filename ' + ' '.join(entities.keys()))
return _File(filename=self.path, **entities)
| as_named_tuple | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n keys = list(self.entities.keys())\n replaced = []\n for i, k in enumerate(keys):\n if iskeyword(k):\n replaced.append(k)\n keys[i] = '%s_' % k\n if replaced:\n safe = [('%s_' % k) for k in replaced]\n warnings.warn(\n 'Entity names cannot be reserved keywords when representing a File as a namedtuple. Replacing entities %s with safe versions %s.'\n % (keys, safe))\n entities = dict(zip(keys, self.entities.values()))\n _File = namedtuple('File', 'filename ' + ' '.join(entities.keys()))\n return _File(filename=self.path, **entities)\n",
"role": "user"
},
{
"content": "as_named_tuple",
"role": "assistant"
}
] |
def x(self, path_patterns, symbolic_link=False, root=None, conflicts='fail'):
new_filename = build_path(self.entities, path_patterns)
if not new_filename:
return None
if new_filename[-1] == os.sep:
new_filename += self.filename
if isabs(self.path) or root is None:
path = self.path
else:
path = join(root, self.path)
if not exists(path):
raise ValueError(
"Target filename to copy/symlink (%s) doesn't exist." % path)
if symbolic_link:
contents = None
link_to = path
else:
with open(path, 'r') as f:
contents = f.read()
link_to = None
write_contents_to_file(new_filename, contents=contents, link_to=link_to,
content_mode='text', root=root, conflicts=conflicts)
| copy | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, path_patterns, symbolic_link=False, root=None, conflicts='fail'):\n new_filename = build_path(self.entities, path_patterns)\n if not new_filename:\n return None\n if new_filename[-1] == os.sep:\n new_filename += self.filename\n if isabs(self.path) or root is None:\n path = self.path\n else:\n path = join(root, self.path)\n if not exists(path):\n raise ValueError(\n \"Target filename to copy/symlink (%s) doesn't exist.\" % path)\n if symbolic_link:\n contents = None\n link_to = path\n else:\n with open(path, 'r') as f:\n contents = f.read()\n link_to = None\n write_contents_to_file(new_filename, contents=contents, link_to=link_to,\n content_mode='text', root=root, conflicts=conflicts)\n",
"role": "user"
},
{
"content": "copy",
"role": "assistant"
}
] |
def x(self, f, update_file=False):
if self.map_func is not None:
val = self.map_func(f)
else:
m = self.regex.search(f.path)
val = m.group(1) if m is not None else None
return self._astype(val)
| match_file | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, f, update_file=False):\n if self.map_func is not None:\n val = self.map_func(f)\n else:\n m = self.regex.search(f.path)\n val = m.group(1) if m is not None else None\n return self._astype(val)\n",
"role": "user"
},
{
"content": "match_file",
"role": "assistant"
}
] |
def x(self, domain):
if isinstance(domain, six.string_types):
if domain in self.domains:
return self.domains[domain]
elif exists(domain):
with open(domain, 'r') as fobj:
domain = json.load(fobj)
else:
raise ValueError(
"No domain could be found/loaded from input '{}'; value must be either the name of an existing Domain, or a valid path to a configuration file."
.format(domain))
name = domain['name']
if name in self.domains:
msg = (
"Domain with name '{}' already exists; returning existing Domain configuration."
.format(name))
warnings.warn(msg)
return self.domains[name]
entities = domain.get('entities', [])
domain = Domain(domain)
for e in entities:
self.add_entity(domain=domain, **e)
self.domains[name] = domain
return self.domains[name]
| get_or_load_domain | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, domain):\n if isinstance(domain, six.string_types):\n if domain in self.domains:\n return self.domains[domain]\n elif exists(domain):\n with open(domain, 'r') as fobj:\n domain = json.load(fobj)\n else:\n raise ValueError(\n \"No domain could be found/loaded from input '{}'; value must be either the name of an existing Domain, or a valid path to a configuration file.\"\n .format(domain))\n name = domain['name']\n if name in self.domains:\n msg = (\n \"Domain with name '{}' already exists; returning existing Domain configuration.\"\n .format(name))\n warnings.warn(msg)\n return self.domains[name]\n entities = domain.get('entities', [])\n domain = Domain(domain)\n for e in entities:\n self.add_entity(domain=domain, **e)\n self.domains[name] = domain\n return self.domains[name]\n",
"role": "user"
},
{
"content": "get_or_load_domain",
"role": "assistant"
}
] |
def x(self, f, domains=None):
filename = f if isinstance(f, six.string_types) else f.path
if domains is None:
domains = list(self.domains.values())
domains = list(domains)
domains.insert(0, self)
for dom in domains:
if dom.include:
for regex in dom.include:
if re.search(regex, filename):
return True
return False
else:
for regex in dom.exclude:
if re.search(regex, filename, flags=re.UNICODE):
return False
return True
| check_inclusions | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, f, domains=None):\n filename = f if isinstance(f, six.string_types) else f.path\n if domains is None:\n domains = list(self.domains.values())\n domains = list(domains)\n domains.insert(0, self)\n for dom in domains:\n if dom.include:\n for regex in dom.include:\n if re.search(regex, filename):\n return True\n return False\n else:\n for regex in dom.exclude:\n if re.search(regex, filename, flags=re.UNICODE):\n return False\n return True\n",
"role": "user"
},
{
"content": "check_inclusions",
"role": "assistant"
}
] |
def x(self, entity):
if entity in self.entities:
return self.entities[entity]
_ent = [e for e in self.entities.values() if e.name == entity]
if len(_ent) > 1:
raise ValueError(
"Entity name '%s' matches %d entities. To avoid ambiguity, please prefix the entity name with its domain (e.g., 'bids.%s'."
% (entity, len(_ent), entity))
if _ent:
return _ent[0]
raise ValueError("No entity '%s' found." % entity)
| find_entity | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, entity):\n if entity in self.entities:\n return self.entities[entity]\n _ent = [e for e in self.entities.values() if e.name == entity]\n if len(_ent) > 1:\n raise ValueError(\n \"Entity name '%s' matches %d entities. To avoid ambiguity, please prefix the entity name with its domain (e.g., 'bids.%s'.\"\n % (entity, len(_ent), entity))\n if _ent:\n return _ent[0]\n raise ValueError(\"No entity '%s' found.\" % entity)\n",
"role": "user"
},
{
"content": "find_entity",
"role": "assistant"
}
] |
def x(self, filename):
data = {}
for f in self.files.values():
entities = {v.entity.id: v.value for k, v in f.tags.items()}
data[f.path] = {'domains': f.domains, 'entities': entities}
with open(filename, 'w') as outfile:
json.dump(data, outfile)
| save_index | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, filename):\n data = {}\n for f in self.files.values():\n entities = {v.entity.id: v.value for k, v in f.tags.items()}\n data[f.path] = {'domains': f.domains, 'entities': entities}\n with open(filename, 'w') as outfile:\n json.dump(data, outfile)\n",
"role": "user"
},
{
"content": "save_index",
"role": "assistant"
}
] |
def x(self, filename, reindex=False):
self._reset_index()
with open(filename, 'r') as fobj:
data = json.load(fobj)
for path, file in data.items():
ents, domains = file['entities'], file['domains']
root, f = dirname(path), basename(path)
if reindex:
self._index_file(root, f, domains)
else:
f = self._make_file_object(root, f)
tags = {k: Tag(self.entities[k], v) for k, v in ents.items()}
f.tags = tags
self.files[f.path] = f
for ent, val in f.entities.items():
self.entities[ent].add_file(f.path, val)
| load_index | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, filename, reindex=False):\n self._reset_index()\n with open(filename, 'r') as fobj:\n data = json.load(fobj)\n for path, file in data.items():\n ents, domains = file['entities'], file['domains']\n root, f = dirname(path), basename(path)\n if reindex:\n self._index_file(root, f, domains)\n else:\n f = self._make_file_object(root, f)\n tags = {k: Tag(self.entities[k], v) for k, v in ents.items()}\n f.tags = tags\n self.files[f.path] = f\n for ent, val in f.entities.items():\n self.entities[ent].add_file(f.path, val)\n",
"role": "user"
},
{
"content": "load_index",
"role": "assistant"
}
] |
def x(self, domain, **kwargs):
map_func = kwargs.get('map_func', None)
if map_func is not None and not callable(kwargs['map_func']):
if self.entity_mapper is None:
raise ValueError(
"Mapping function '%s' specified for Entity '%s', but no entity mapper was passed when initializing the current Layout. Please make sure the 'entity_mapper' argument is set."
% (map_func, kwargs['name']))
map_func = getattr(self.entity_mapper, kwargs['map_func'])
kwargs['map_func'] = map_func
ent = Entity(domain=domain, **kwargs)
domain.add_entity(ent)
if ent.mandatory:
self.mandatory.add(ent.id)
if ent.directory is not None:
ent.directory = ent.directory.replace('{{root}}', self.root)
self.entities[ent.id] = ent
for alias in ent.aliases:
self.entities[alias] = ent
if self.dynamic_getters:
func = partial(getattr(self, 'get'), target=ent.name, return_type='id')
func_name = inflect.engine().plural(ent.name)
setattr(self, 'get_%s' % func_name, func)
| add_entity | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, domain, **kwargs):\n map_func = kwargs.get('map_func', None)\n if map_func is not None and not callable(kwargs['map_func']):\n if self.entity_mapper is None:\n raise ValueError(\n \"Mapping function '%s' specified for Entity '%s', but no entity mapper was passed when initializing the current Layout. Please make sure the 'entity_mapper' argument is set.\"\n % (map_func, kwargs['name']))\n map_func = getattr(self.entity_mapper, kwargs['map_func'])\n kwargs['map_func'] = map_func\n ent = Entity(domain=domain, **kwargs)\n domain.add_entity(ent)\n if ent.mandatory:\n self.mandatory.add(ent.id)\n if ent.directory is not None:\n ent.directory = ent.directory.replace('{{root}}', self.root)\n self.entities[ent.id] = ent\n for alias in ent.aliases:\n self.entities[alias] = ent\n if self.dynamic_getters:\n func = partial(getattr(self, 'get'), target=ent.name, return_type='id')\n func_name = inflect.engine().plural(ent.name)\n setattr(self, 'get_%s' % func_name, func)\n",
"role": "user"
},
{
"content": "add_entity",
"role": "assistant"
}
] |
def x(self, return_type='tuple', target=None, extensions=None, domains=None,
regex_search=None, **kwargs):
if regex_search is None:
regex_search = self.regex_search
result = []
filters = {}
filters.update(kwargs)
for filename, file in self.files.items():
if not file._matches(filters, extensions, domains, regex_search):
continue
result.append(file)
if not self.absolute_paths:
for i, f in enumerate(result):
f = copy(f)
f.path = relpath(f.path, self.root)
result[i] = f
if return_type == 'file':
return natural_sort([f.path for f in result])
if return_type == 'tuple':
result = [r.as_named_tuple() for r in result]
return natural_sort(result, field='filename')
if return_type.startswith('obj'):
return result
else:
valid_entities = self.get_domain_entities(domains)
if target is None:
raise ValueError(
'If return_type is "id" or "dir", a valid target entity must also be specified.'
)
result = [x for x in result if target in x.entities]
if return_type == 'id':
result = list(set([x.entities[target] for x in result]))
return natural_sort(result)
elif return_type == 'dir':
template = valid_entities[target].directory
if template is None:
raise ValueError(
'Return type set to directory, but no directory template is defined for the target entity ("%s").'
% target)
to_rep = re.findall('\\{(.*?)\\}', template)
for ent in to_rep:
patt = valid_entities[ent].pattern
template = template.replace('{%s}' % ent, patt)
template += '[^\\%s]*$' % os.path.sep
matches = [f.dirname for f in result if re.search(template, f.
dirname)]
return natural_sort(list(set(matches)))
else:
raise ValueError(
"Invalid return_type specified (must be one of 'tuple', 'file', 'id', or 'dir'."
)
| get | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, return_type='tuple', target=None, extensions=None, domains=None,\n regex_search=None, **kwargs):\n if regex_search is None:\n regex_search = self.regex_search\n result = []\n filters = {}\n filters.update(kwargs)\n for filename, file in self.files.items():\n if not file._matches(filters, extensions, domains, regex_search):\n continue\n result.append(file)\n if not self.absolute_paths:\n for i, f in enumerate(result):\n f = copy(f)\n f.path = relpath(f.path, self.root)\n result[i] = f\n if return_type == 'file':\n return natural_sort([f.path for f in result])\n if return_type == 'tuple':\n result = [r.as_named_tuple() for r in result]\n return natural_sort(result, field='filename')\n if return_type.startswith('obj'):\n return result\n else:\n valid_entities = self.get_domain_entities(domains)\n if target is None:\n raise ValueError(\n 'If return_type is \"id\" or \"dir\", a valid target entity must also be specified.'\n )\n result = [x for x in result if target in x.entities]\n if return_type == 'id':\n result = list(set([x.entities[target] for x in result]))\n return natural_sort(result)\n elif return_type == 'dir':\n template = valid_entities[target].directory\n if template is None:\n raise ValueError(\n 'Return type set to directory, but no directory template is defined for the target entity (\"%s\").'\n % target)\n to_rep = re.findall('\\\\{(.*?)\\\\}', template)\n for ent in to_rep:\n patt = valid_entities[ent].pattern\n template = template.replace('{%s}' % ent, patt)\n template += '[^\\\\%s]*$' % os.path.sep\n matches = [f.dirname for f in result if re.search(template, f.\n dirname)]\n return natural_sort(list(set(matches)))\n else:\n raise ValueError(\n \"Invalid return_type specified (must be one of 'tuple', 'file', 'id', or 'dir'.\"\n )\n",
"role": "user"
},
{
"content": "get",
"role": "assistant"
}
] |
def x(self, entity, files=False):
return self._find_entity(entity).count(files)
| count | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, entity, files=False):\n return self._find_entity(entity).count(files)\n",
"role": "user"
},
{
"content": "count",
"role": "assistant"
}
] |
def x(self, **kwargs):
try:
import pandas as pd
except ImportError:
raise ImportError(
"What are you doing trying to export a Layout as a pandas DataFrame when you don't have pandas installed? Eh? Eh?"
)
if kwargs:
files = self.get(return_type='obj', **kwargs)
else:
files = self.files.values()
data = pd.DataFrame.from_records([f.entities for f in files])
data.insert(0, 'path', [f.path for f in files])
return data
| as_data_frame | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, **kwargs):\n try:\n import pandas as pd\n except ImportError:\n raise ImportError(\n \"What are you doing trying to export a Layout as a pandas DataFrame when you don't have pandas installed? Eh? Eh?\"\n )\n if kwargs:\n files = self.get(return_type='obj', **kwargs)\n else:\n files = self.files.values()\n data = pd.DataFrame.from_records([f.entities for f in files])\n data.insert(0, 'path', [f.path for f in files])\n return data\n",
"role": "user"
},
{
"content": "as_data_frame",
"role": "assistant"
}
] |
def x(self, path, return_type='file', strict=True, all_=False,
ignore_strict_entities=None, full_search=False, **kwargs):
entities = {}
for ent in self.entities.values():
m = ent.regex.search(path)
if m:
entities[ent.name] = ent._astype(m.group(1))
if strict and ignore_strict_entities is not None:
for k in ignore_strict_entities:
entities.pop(k, None)
results = self.get(return_type='file', **kwargs)
folders = defaultdict(list)
for filename in results:
f = self.get_file(filename)
folders[f.dirname].append(f)
def x(f):
f_ents = f.entities
keys = set(entities.keys()) & set(f_ents.keys())
shared = len(keys)
return [shared, sum([(entities[k] == f_ents[k]) for k in keys])]
matches = []
search_paths = []
while True:
if path in folders and folders[path]:
search_paths.append(path)
parent = dirname(path)
if parent == path:
break
path = parent
if full_search:
unchecked = set(folders.keys()) - set(search_paths)
search_paths.extend(path for path in unchecked if folders[path])
for path in search_paths:
num_ents = [([f] + count_matches(f)) for f in folders[path]]
if strict:
num_ents = [f for f in num_ents if f[1] == f[2]]
num_ents.sort(key=lambda x: x[2], reverse=True)
if num_ents:
matches.append(num_ents[0][0])
if not all_:
break
matches = [(m.path if return_type == 'file' else m.as_named_tuple()) for
m in matches]
return matches if all_ else matches[0] if matches else None
| get_nearest | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, path, return_type='file', strict=True, all_=False,\n ignore_strict_entities=None, full_search=False, **kwargs):\n entities = {}\n for ent in self.entities.values():\n m = ent.regex.search(path)\n if m:\n entities[ent.name] = ent._astype(m.group(1))\n if strict and ignore_strict_entities is not None:\n for k in ignore_strict_entities:\n entities.pop(k, None)\n results = self.get(return_type='file', **kwargs)\n folders = defaultdict(list)\n for filename in results:\n f = self.get_file(filename)\n folders[f.dirname].append(f)\n\n def x(f):\n f_ents = f.entities\n keys = set(entities.keys()) & set(f_ents.keys())\n shared = len(keys)\n return [shared, sum([(entities[k] == f_ents[k]) for k in keys])]\n matches = []\n search_paths = []\n while True:\n if path in folders and folders[path]:\n search_paths.append(path)\n parent = dirname(path)\n if parent == path:\n break\n path = parent\n if full_search:\n unchecked = set(folders.keys()) - set(search_paths)\n search_paths.extend(path for path in unchecked if folders[path])\n for path in search_paths:\n num_ents = [([f] + count_matches(f)) for f in folders[path]]\n if strict:\n num_ents = [f for f in num_ents if f[1] == f[2]]\n num_ents.sort(key=lambda x: x[2], reverse=True)\n if num_ents:\n matches.append(num_ents[0][0])\n if not all_:\n break\n matches = [(m.path if return_type == 'file' else m.as_named_tuple()) for\n m in matches]\n return matches if all_ else matches[0] if matches else None\n",
"role": "user"
},
{
"content": "get_nearest",
"role": "assistant"
}
] |
def x(self, source, path_patterns=None, strict=False, domains=None):
if isinstance(source, six.string_types):
if source not in self.files:
source = join(self.root, source)
source = self.get_file(source)
if isinstance(source, File):
source = source.entities
if path_patterns is None:
if domains is None:
domains = list(self.domains.keys())
path_patterns = []
for dom in listify(domains):
path_patterns.extend(self.domains[dom].path_patterns)
return build_path(source, path_patterns, strict)
| build_path | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, source, path_patterns=None, strict=False, domains=None):\n if isinstance(source, six.string_types):\n if source not in self.files:\n source = join(self.root, source)\n source = self.get_file(source)\n if isinstance(source, File):\n source = source.entities\n if path_patterns is None:\n if domains is None:\n domains = list(self.domains.keys())\n path_patterns = []\n for dom in listify(domains):\n path_patterns.extend(self.domains[dom].path_patterns)\n return build_path(source, path_patterns, strict)\n",
"role": "user"
},
{
"content": "build_path",
"role": "assistant"
}
] |
def x(self, entities, path_patterns=None, contents=None, link_to=None,
content_mode='text', conflicts='fail', strict=False, domains=None,
index=False, index_domains=None):
path = self.build_path(entities, path_patterns, strict, domains)
if path is None:
raise ValueError(
'Cannot construct any valid filename for the passed entities given available path patterns.'
)
write_contents_to_file(path, contents=contents, link_to=link_to,
content_mode=content_mode, conflicts=conflicts, root=self.root)
if index:
if index_domains is None:
index_domains = list(self.domains.keys())
self._index_file(self.root, path, index_domains)
| write_contents_to_file | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, entities, path_patterns=None, contents=None, link_to=None,\n content_mode='text', conflicts='fail', strict=False, domains=None,\n index=False, index_domains=None):\n path = self.build_path(entities, path_patterns, strict, domains)\n if path is None:\n raise ValueError(\n 'Cannot construct any valid filename for the passed entities given available path patterns.'\n )\n write_contents_to_file(path, contents=contents, link_to=link_to,\n content_mode=content_mode, conflicts=conflicts, root=self.root)\n if index:\n if index_domains is None:\n index_domains = list(self.domains.keys())\n self._index_file(self.root, path, index_domains)\n",
"role": "user"
},
{
"content": "write_contents_to_file",
"role": "assistant"
}
] |
def x(name, log_stream=sys.stdout, log_file=None, log_level=logging.INFO,
keep_old_handlers=False, propagate=False):
logger = logging.getLogger(name)
logger.setLevel(log_level)
logger.propagate = propagate
if not keep_old_handlers:
logger.handlers = []
log_fmt = '[%(asctime)s] %(levelname)s: %(message)s'
log_datefmt = '%Y-%m-%d %H:%M:%S'
formatter = logging.Formatter(log_fmt, log_datefmt)
if log_stream is not None:
stream_handler = logging.StreamHandler(log_stream)
stream_handler.setFormatter(formatter)
logger.addHandler(stream_handler)
if log_file is not None:
file_handler = logging.FileHandler(log_file)
file_handler.setFormatter(formatter)
logger.addHandler(file_handler)
if log_stream is None and log_file is None:
logger.addHandler(logging.NullHandler())
return logger
| configure_logger | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(name, log_stream=sys.stdout, log_file=None, log_level=logging.INFO,\n keep_old_handlers=False, propagate=False):\n logger = logging.getLogger(name)\n logger.setLevel(log_level)\n logger.propagate = propagate\n if not keep_old_handlers:\n logger.handlers = []\n log_fmt = '[%(asctime)s] %(levelname)s: %(message)s'\n log_datefmt = '%Y-%m-%d %H:%M:%S'\n formatter = logging.Formatter(log_fmt, log_datefmt)\n if log_stream is not None:\n stream_handler = logging.StreamHandler(log_stream)\n stream_handler.setFormatter(formatter)\n logger.addHandler(stream_handler)\n if log_file is not None:\n file_handler = logging.FileHandler(log_file)\n file_handler.setFormatter(formatter)\n logger.addHandler(file_handler)\n if log_stream is None and log_file is None:\n logger.addHandler(logging.NullHandler())\n return logger\n",
"role": "user"
},
{
"content": "configure_logger",
"role": "assistant"
}
] |
def x(name='', log_stream=None, log_file=None, quiet=False, verbose=False):
log_level = logging.INFO
if quiet:
log_level = logging.WARNING
elif verbose:
log_level = logging.DEBUG
if log_stream is None:
log_stream = sys.stdout
new_logger = configure_logger(name, log_stream=log_stream, log_file=
log_file, log_level=log_level)
return new_logger
| get_logger | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(name='', log_stream=None, log_file=None, quiet=False, verbose=False):\n log_level = logging.INFO\n if quiet:\n log_level = logging.WARNING\n elif verbose:\n log_level = logging.DEBUG\n if log_stream is None:\n log_stream = sys.stdout\n new_logger = configure_logger(name, log_stream=log_stream, log_file=\n log_file, log_level=log_level)\n return new_logger\n",
"role": "user"
},
{
"content": "get_logger",
"role": "assistant"
}
] |
def x(milliseconds, func, *args, **kwargs):
stopper = threading.Event()
def x(seconds, func, *args, **kwargs):
def x():
if stopper.isSet():
return
interval(seconds, func, *args, **kwargs)
try:
func(*args, **kwargs)
except:
logging.error('Error during interval')
logging.error(traceback.format_exc())
thread = threading.Timer(seconds, wrapper)
thread.daemon = True
thread.start()
interval(milliseconds / 1000, func, *args, **kwargs)
return stopper
| start | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(milliseconds, func, *args, **kwargs):\n stopper = threading.Event()\n\n def x(seconds, func, *args, **kwargs):\n\n def x():\n if stopper.isSet():\n return\n interval(seconds, func, *args, **kwargs)\n try:\n func(*args, **kwargs)\n except:\n logging.error('Error during interval')\n logging.error(traceback.format_exc())\n thread = threading.Timer(seconds, wrapper)\n thread.daemon = True\n thread.start()\n interval(milliseconds / 1000, func, *args, **kwargs)\n return stopper\n",
"role": "user"
},
{
"content": "start",
"role": "assistant"
}
] |
def x(codons1, codons2):
result = defaultdict(list)
for c1 in codons1:
for c2 in codons2:
distance = findDistance(c1, c2)
result[distance].append([c1, c2])
return result
| codonInformation | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(codons1, codons2):\n result = defaultdict(list)\n for c1 in codons1:\n for c2 in codons2:\n distance = findDistance(c1, c2)\n result[distance].append([c1, c2])\n return result\n",
"role": "user"
},
{
"content": "codonInformation",
"role": "assistant"
}
] |
def x(api_client):
try:
pprint((yield from api_client.echo()))
except errors.RequestError as exc:
log.exception('Exception occurred: %s', exc)
yield gen.Task(lambda *args, **kwargs: ioloop.IOLoop.current().stop())
| example_async_client | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(api_client):\n try:\n pprint((yield from api_client.echo()))\n except errors.RequestError as exc:\n log.exception('Exception occurred: %s', exc)\n yield gen.Task(lambda *args, **kwargs: ioloop.IOLoop.current().stop())\n",
"role": "user"
},
{
"content": "example_async_client",
"role": "assistant"
}
] |
def x(api_client):
try:
pprint(api_client.echo())
except errors.RequestError as exc:
log.exception('Exception occurred: %s', exc)
| example_sync_client | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(api_client):\n try:\n pprint(api_client.echo())\n except errors.RequestError as exc:\n log.exception('Exception occurred: %s', exc)\n",
"role": "user"
},
{
"content": "example_sync_client",
"role": "assistant"
}
] |
def x():
logging.basicConfig(level=logging.INFO)
example_sync_client(SyncAPIClient())
example_async_client(AsyncAPIClient())
io_loop = ioloop.IOLoop.current()
io_loop.start()
| main | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n logging.basicConfig(level=logging.INFO)\n example_sync_client(SyncAPIClient())\n example_async_client(AsyncAPIClient())\n io_loop = ioloop.IOLoop.current()\n io_loop.start()\n",
"role": "user"
},
{
"content": "main",
"role": "assistant"
}
] |
def x(args=None):
vinfo = sys.version_info
if not vinfo >= (2, 7):
raise SystemError(
'Python interpreter version >= 2.7 required, found %d.%d instead.'
% (vinfo.major, vinfo.minor))
if args is None:
parser = get_argument_parser()
args = parser.parse_args()
input_file = args.annotation_file
output_file = args.output_file
chrom_pat = args.chromosome_pattern
log_file = args.log_file
quiet = args.quiet
verbose = args.verbose
log_stream = sys.stdout
if output_file == '-':
log_stream = sys.stderr
logger = misc.get_logger(log_stream=log_stream, log_file=log_file,
quiet=quiet, verbose=verbose)
if chrom_pat is not None:
logger.info(
'Regular expression used for filtering chromosome names: "%s"',
chrom_pat)
if input_file == '-':
input_file = sys.stdin
if output_file == '-':
output_file = sys.stdout
genes = ensembl.get_protein_coding_genes(input_file, chromosome_pattern
=chrom_pat)
genes.to_csv(output_file, sep='\t', index=False)
return 0
| main | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(args=None):\n vinfo = sys.version_info\n if not vinfo >= (2, 7):\n raise SystemError(\n 'Python interpreter version >= 2.7 required, found %d.%d instead.'\n % (vinfo.major, vinfo.minor))\n if args is None:\n parser = get_argument_parser()\n args = parser.parse_args()\n input_file = args.annotation_file\n output_file = args.output_file\n chrom_pat = args.chromosome_pattern\n log_file = args.log_file\n quiet = args.quiet\n verbose = args.verbose\n log_stream = sys.stdout\n if output_file == '-':\n log_stream = sys.stderr\n logger = misc.get_logger(log_stream=log_stream, log_file=log_file,\n quiet=quiet, verbose=verbose)\n if chrom_pat is not None:\n logger.info(\n 'Regular expression used for filtering chromosome names: \"%s\"',\n chrom_pat)\n if input_file == '-':\n input_file = sys.stdin\n if output_file == '-':\n output_file = sys.stdout\n genes = ensembl.get_protein_coding_genes(input_file, chromosome_pattern\n =chrom_pat)\n genes.to_csv(output_file, sep='\\t', index=False)\n return 0\n",
"role": "user"
},
{
"content": "main",
"role": "assistant"
}
] |
def x(self, *stats):
self.statistics = OrderedDict()
for port in self.ports:
port_stats = IxeStatTotal(port).get_attributes(FLAG_RDONLY, *stats)
port_stats.update({(c + '_rate'): v for c, v in IxeStatRate(port).
get_attributes(FLAG_RDONLY, *stats).items()})
self.statistics[str(port)] = port_stats
return self.statistics
| read_stats | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, *stats):\n self.statistics = OrderedDict()\n for port in self.ports:\n port_stats = IxeStatTotal(port).get_attributes(FLAG_RDONLY, *stats)\n port_stats.update({(c + '_rate'): v for c, v in IxeStatRate(port).\n get_attributes(FLAG_RDONLY, *stats).items()})\n self.statistics[str(port)] = port_stats\n return self.statistics\n",
"role": "user"
},
{
"content": "read_stats",
"role": "assistant"
}
] |
def x(self, *stats):
from ixexplorer.ixe_stream import IxePacketGroupStream
sleep_time = 0.1
if not stats:
stats = [m.attrname for m in IxePgStats.__tcl_members__ if m.flags &
FLAG_RDONLY]
sleep_time = 1
for port in self.tx_ports_streams:
port.api.call_rc('streamTransmitStats get {} 1 4096'.format(port.uri))
for rx_port in self.rx_ports:
rx_port.api.call_rc('packetGroupStats get {} 0 65536'.format(
rx_port.uri))
time.sleep(sleep_time)
self.statistics = OrderedDict()
for tx_port, streams in self.tx_ports_streams.items():
for stream in streams:
stream_stats = OrderedDict()
tx_port.api.call_rc('streamTransmitStats get {} 1 4096'.format(
tx_port.uri))
stream_tx_stats = IxeStreamTxStats(tx_port, stream.index)
stream_stats_tx = {c: v for c, v in stream_tx_stats.
get_attributes(FLAG_RDONLY).items()}
stream_stats['tx'] = stream_stats_tx
stream_stat_pgid = IxePacketGroupStream(stream).groupId
stream_stats_pg = pg_stats_dict()
for port in self.session.ports.values():
stream_stats_pg[str(port)] = OrderedDict(zip(stats, [-1] *
len(stats)))
for rx_port in self.rx_ports:
if not stream.rx_ports or rx_port in stream.rx_ports:
rx_port.api.call_rc('packetGroupStats get {} 0 65536'.
format(rx_port.uri))
pg_stats = IxePgStats(rx_port, stream_stat_pgid)
stream_stats_pg[str(rx_port)] = pg_stats.read_stats(*stats)
stream_stats['rx'] = stream_stats_pg
self.statistics[str(stream)] = stream_stats
return self.statistics
| read_stats | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, *stats):\n from ixexplorer.ixe_stream import IxePacketGroupStream\n sleep_time = 0.1\n if not stats:\n stats = [m.attrname for m in IxePgStats.__tcl_members__ if m.flags &\n FLAG_RDONLY]\n sleep_time = 1\n for port in self.tx_ports_streams:\n port.api.call_rc('streamTransmitStats get {} 1 4096'.format(port.uri))\n for rx_port in self.rx_ports:\n rx_port.api.call_rc('packetGroupStats get {} 0 65536'.format(\n rx_port.uri))\n time.sleep(sleep_time)\n self.statistics = OrderedDict()\n for tx_port, streams in self.tx_ports_streams.items():\n for stream in streams:\n stream_stats = OrderedDict()\n tx_port.api.call_rc('streamTransmitStats get {} 1 4096'.format(\n tx_port.uri))\n stream_tx_stats = IxeStreamTxStats(tx_port, stream.index)\n stream_stats_tx = {c: v for c, v in stream_tx_stats.\n get_attributes(FLAG_RDONLY).items()}\n stream_stats['tx'] = stream_stats_tx\n stream_stat_pgid = IxePacketGroupStream(stream).groupId\n stream_stats_pg = pg_stats_dict()\n for port in self.session.ports.values():\n stream_stats_pg[str(port)] = OrderedDict(zip(stats, [-1] *\n len(stats)))\n for rx_port in self.rx_ports:\n if not stream.rx_ports or rx_port in stream.rx_ports:\n rx_port.api.call_rc('packetGroupStats get {} 0 65536'.\n format(rx_port.uri))\n pg_stats = IxePgStats(rx_port, stream_stat_pgid)\n stream_stats_pg[str(rx_port)] = pg_stats.read_stats(*stats)\n stream_stats['rx'] = stream_stats_pg\n self.statistics[str(stream)] = stream_stats\n return self.statistics\n",
"role": "user"
},
{
"content": "read_stats",
"role": "assistant"
}
] |
def x(rst, clk, req_vec, gnt_vec=None, gnt_idx=None, gnt_vld=None, gnt_rdy=
None, ARBITER_TYPE='priority'):
if ARBITER_TYPE == 'priority':
_arb = arbiter_priority(req_vec, gnt_vec, gnt_idx, gnt_vld)
elif ARBITER_TYPE == 'roundrobin':
_arb = arbiter_roundrobin(rst, clk, req_vec, gnt_vec, gnt_idx,
gnt_vld, gnt_rdy)
else:
assert 'Arbiter: Unknown arbiter type: {}'.format(ARBITER_TYPE)
return _arb
| arbiter | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(rst, clk, req_vec, gnt_vec=None, gnt_idx=None, gnt_vld=None, gnt_rdy=\n None, ARBITER_TYPE='priority'):\n if ARBITER_TYPE == 'priority':\n _arb = arbiter_priority(req_vec, gnt_vec, gnt_idx, gnt_vld)\n elif ARBITER_TYPE == 'roundrobin':\n _arb = arbiter_roundrobin(rst, clk, req_vec, gnt_vec, gnt_idx,\n gnt_vld, gnt_rdy)\n else:\n assert 'Arbiter: Unknown arbiter type: {}'.format(ARBITER_TYPE)\n return _arb\n",
"role": "user"
},
{
"content": "arbiter",
"role": "assistant"
}
] |
def x(req_vec, gnt_vec=None, gnt_idx=None, gnt_vld=None):
REQ_NUM = len(req_vec)
gnt_vec_s = Signal(intbv(0)[REQ_NUM:])
gnt_idx_s = Signal(intbv(0, min=0, max=REQ_NUM))
gnt_vld_s = Signal(bool(0))
@always_comb
def x():
gnt_vec_s.next = 0
gnt_idx_s.next = 0
gnt_vld_s.next = 0
for i in range(REQ_NUM):
if req_vec[i] == 1:
gnt_vec_s.next[i] = 1
gnt_idx_s.next = i
gnt_vld_s.next = 1
break
if gnt_vec != None:
_vec = assign(gnt_vec, gnt_vec_s)
if gnt_idx != None:
_idx = assign(gnt_idx, gnt_idx_s)
if gnt_vld != None:
_vld = assign(gnt_vld, gnt_vld_s)
return instances()
| arbiter_priority | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(req_vec, gnt_vec=None, gnt_idx=None, gnt_vld=None):\n REQ_NUM = len(req_vec)\n gnt_vec_s = Signal(intbv(0)[REQ_NUM:])\n gnt_idx_s = Signal(intbv(0, min=0, max=REQ_NUM))\n gnt_vld_s = Signal(bool(0))\n\n @always_comb\n def x():\n gnt_vec_s.next = 0\n gnt_idx_s.next = 0\n gnt_vld_s.next = 0\n for i in range(REQ_NUM):\n if req_vec[i] == 1:\n gnt_vec_s.next[i] = 1\n gnt_idx_s.next = i\n gnt_vld_s.next = 1\n break\n if gnt_vec != None:\n _vec = assign(gnt_vec, gnt_vec_s)\n if gnt_idx != None:\n _idx = assign(gnt_idx, gnt_idx_s)\n if gnt_vld != None:\n _vld = assign(gnt_vld, gnt_vld_s)\n return instances()\n",
"role": "user"
},
{
"content": "arbiter_priority",
"role": "assistant"
}
] |
def x(rst, clk, req_vec, gnt_vec=None, gnt_idx=None, gnt_vld=None, gnt_rdy=None
):
REQ_NUM = len(req_vec)
ptr = Signal(intbv(0, min=0, max=REQ_NUM))
gnt_vec_s = Signal(intbv(0)[REQ_NUM:])
gnt_idx_s = Signal(intbv(0, min=0, max=REQ_NUM))
gnt_vld_s = Signal(bool(0))
@always(clk.posedge)
def x():
if rst:
ptr.next = REQ_NUM - 1
elif gnt_rdy and gnt_vld_s:
ptr.next = gnt_idx_s
@always_comb
def x():
gnt_vec_s.next = 0
gnt_idx_s.next = 0
gnt_vld_s.next = 0
for i in range(REQ_NUM):
if i > ptr:
if req_vec[i] == 1:
gnt_vec_s.next[i] = 1
gnt_idx_s.next = i
gnt_vld_s.next = 1
return
for i in range(REQ_NUM):
if req_vec[i] == 1:
gnt_vec_s.next[i] = 1
gnt_idx_s.next = i
gnt_vld_s.next = 1
return
if gnt_vec != None:
_vec = assign(gnt_vec, gnt_vec_s)
if gnt_idx != None:
_idx = assign(gnt_idx, gnt_idx_s)
if gnt_vld != None:
_vld = assign(gnt_vld, gnt_vld_s)
return instances()
| arbiter_roundrobin | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(rst, clk, req_vec, gnt_vec=None, gnt_idx=None, gnt_vld=None, gnt_rdy=None\n ):\n REQ_NUM = len(req_vec)\n ptr = Signal(intbv(0, min=0, max=REQ_NUM))\n gnt_vec_s = Signal(intbv(0)[REQ_NUM:])\n gnt_idx_s = Signal(intbv(0, min=0, max=REQ_NUM))\n gnt_vld_s = Signal(bool(0))\n\n @always(clk.posedge)\n def x():\n if rst:\n ptr.next = REQ_NUM - 1\n elif gnt_rdy and gnt_vld_s:\n ptr.next = gnt_idx_s\n\n @always_comb\n def x():\n gnt_vec_s.next = 0\n gnt_idx_s.next = 0\n gnt_vld_s.next = 0\n for i in range(REQ_NUM):\n if i > ptr:\n if req_vec[i] == 1:\n gnt_vec_s.next[i] = 1\n gnt_idx_s.next = i\n gnt_vld_s.next = 1\n return\n for i in range(REQ_NUM):\n if req_vec[i] == 1:\n gnt_vec_s.next[i] = 1\n gnt_idx_s.next = i\n gnt_vld_s.next = 1\n return\n if gnt_vec != None:\n _vec = assign(gnt_vec, gnt_vec_s)\n if gnt_idx != None:\n _idx = assign(gnt_idx, gnt_idx_s)\n if gnt_vld != None:\n _vld = assign(gnt_vld, gnt_vld_s)\n return instances()\n",
"role": "user"
},
{
"content": "arbiter_roundrobin",
"role": "assistant"
}
] |
def x(seq, normalize=True):
seq = normalize_sequence(seq) if normalize else seq
return str(vmc_digest(seq, digest_size=24))
| seq_seqhash | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(seq, normalize=True):\n seq = normalize_sequence(seq) if normalize else seq\n return str(vmc_digest(seq, digest_size=24))\n",
"role": "user"
},
{
"content": "seq_seqhash",
"role": "assistant"
}
] |
def x(seq, normalize=True):
seq = normalize_sequence(seq) if normalize else seq
bseq = seq.encode('ascii')
return base64.b64encode(hashlib.sha1(bseq).digest()).decode('ascii'
).rstrip('=')
| seq_seguid | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(seq, normalize=True):\n seq = normalize_sequence(seq) if normalize else seq\n bseq = seq.encode('ascii')\n return base64.b64encode(hashlib.sha1(bseq).digest()).decode('ascii'\n ).rstrip('=')\n",
"role": "user"
},
{
"content": "seq_seguid",
"role": "assistant"
}
] |
def x(seq, normalize=True):
seq = normalize_sequence(seq) if normalize else seq
bseq = seq.encode('ascii')
return hashlib.md5(bseq).hexdigest()
| seq_md5 | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(seq, normalize=True):\n seq = normalize_sequence(seq) if normalize else seq\n bseq = seq.encode('ascii')\n return hashlib.md5(bseq).hexdigest()\n",
"role": "user"
},
{
"content": "seq_md5",
"role": "assistant"
}
] |
def x(seq, normalize=True):
seq = normalize_sequence(seq) if normalize else seq
bseq = seq.encode('ascii')
return hashlib.sha1(bseq).hexdigest()
| seq_sha1 | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(seq, normalize=True):\n seq = normalize_sequence(seq) if normalize else seq\n bseq = seq.encode('ascii')\n return hashlib.sha1(bseq).hexdigest()\n",
"role": "user"
},
{
"content": "seq_sha1",
"role": "assistant"
}
] |
def x(seq, normalize=True):
seq = normalize_sequence(seq) if normalize else seq
bseq = seq.encode('ascii')
return hashlib.sha512(bseq).hexdigest()
| seq_sha512 | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(seq, normalize=True):\n seq = normalize_sequence(seq) if normalize else seq\n bseq = seq.encode('ascii')\n return hashlib.sha512(bseq).hexdigest()\n",
"role": "user"
},
{
"content": "seq_sha512",
"role": "assistant"
}
] |
def x(seq, normalize=True):
seq = normalize_sequence(seq) if normalize else seq
return {'namespace': 'VMC', 'accession': 'GS_' + str(vmc_digest(seq))}
| seq_vmc_identifier | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(seq, normalize=True):\n seq = normalize_sequence(seq) if normalize else seq\n return {'namespace': 'VMC', 'accession': 'GS_' + str(vmc_digest(seq))}\n",
"role": "user"
},
{
"content": "seq_vmc_identifier",
"role": "assistant"
}
] |
def x(credentials, instance_config, instance_name, script_dir, index_dir,
fastq_file, output_dir, num_threads=None, seed_start_lmax=None,
mismatch_nmax=None, multimap_nmax=None, splice_min_overhang=None,
out_mult_nmax=None, sort_bam=True, keep_unmapped=False, self_destruct=
True, compressed=True, **kwargs):
if sort_bam:
out_sam_type = 'BAM SortedByCoordinate'
else:
out_sam_type = 'BAM Unsorted'
fastq_files = fastq_file
if isinstance(fastq_files, (str, _oldstr)):
fastq_files = [fastq_file]
template = _TEMPLATE_ENV.get_template(os.path.join('map_single-end.sh'))
startup_script = template.render(script_dir=script_dir, index_dir=
index_dir, fastq_files=fastq_files, output_dir=output_dir,
num_threads=num_threads, seed_start_lmax=seed_start_lmax,
self_destruct=self_destruct, mismatch_nmax=mismatch_nmax,
multimap_nmax=multimap_nmax, splice_min_overhang=
splice_min_overhang, out_mult_nmax=out_mult_nmax, keep_unmapped=
keep_unmapped, compressed=compressed, out_sam_type=out_sam_type)
if len(startup_script) > 32768:
raise ValueError('Startup script larger than 32,768 bytes!')
op_name = instance_config.create_instance(credentials, instance_name,
startup_script=startup_script, **kwargs)
return op_name
| map_single_end | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(credentials, instance_config, instance_name, script_dir, index_dir,\n fastq_file, output_dir, num_threads=None, seed_start_lmax=None,\n mismatch_nmax=None, multimap_nmax=None, splice_min_overhang=None,\n out_mult_nmax=None, sort_bam=True, keep_unmapped=False, self_destruct=\n True, compressed=True, **kwargs):\n if sort_bam:\n out_sam_type = 'BAM SortedByCoordinate'\n else:\n out_sam_type = 'BAM Unsorted'\n fastq_files = fastq_file\n if isinstance(fastq_files, (str, _oldstr)):\n fastq_files = [fastq_file]\n template = _TEMPLATE_ENV.get_template(os.path.join('map_single-end.sh'))\n startup_script = template.render(script_dir=script_dir, index_dir=\n index_dir, fastq_files=fastq_files, output_dir=output_dir,\n num_threads=num_threads, seed_start_lmax=seed_start_lmax,\n self_destruct=self_destruct, mismatch_nmax=mismatch_nmax,\n multimap_nmax=multimap_nmax, splice_min_overhang=\n splice_min_overhang, out_mult_nmax=out_mult_nmax, keep_unmapped=\n keep_unmapped, compressed=compressed, out_sam_type=out_sam_type)\n if len(startup_script) > 32768:\n raise ValueError('Startup script larger than 32,768 bytes!')\n op_name = instance_config.create_instance(credentials, instance_name,\n startup_script=startup_script, **kwargs)\n return op_name\n",
"role": "user"
},
{
"content": "map_single_end",
"role": "assistant"
}
] |
def x(credentials, instance_config, instance_name, script_dir, genome_file,
output_dir, annotation_file=None, splice_overhang=100, num_threads=8,
chromosome_bin_bits=18, genome_memory_limit=31000000000, self_destruct=
True, **kwargs):
template = _TEMPLATE_ENV.get_template(os.path.join('generate_index.sh'))
startup_script = template.render(script_dir=script_dir, genome_file=
genome_file, annotation_file=annotation_file, splice_overhang=
splice_overhang, output_dir=output_dir, num_threads=num_threads,
chromosome_bin_bits=chromosome_bin_bits, genome_memory_limit=
genome_memory_limit, self_destruct=self_destruct)
op_name = instance_config.create_instance(credentials, instance_name,
startup_script=startup_script, **kwargs)
return op_name
| generate_index | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(credentials, instance_config, instance_name, script_dir, genome_file,\n output_dir, annotation_file=None, splice_overhang=100, num_threads=8,\n chromosome_bin_bits=18, genome_memory_limit=31000000000, self_destruct=\n True, **kwargs):\n template = _TEMPLATE_ENV.get_template(os.path.join('generate_index.sh'))\n startup_script = template.render(script_dir=script_dir, genome_file=\n genome_file, annotation_file=annotation_file, splice_overhang=\n splice_overhang, output_dir=output_dir, num_threads=num_threads,\n chromosome_bin_bits=chromosome_bin_bits, genome_memory_limit=\n genome_memory_limit, self_destruct=self_destruct)\n op_name = instance_config.create_instance(credentials, instance_name,\n startup_script=startup_script, **kwargs)\n return op_name\n",
"role": "user"
},
{
"content": "generate_index",
"role": "assistant"
}
] |
def x(rst, clk, full, we, din, empty, re, dout, afull=None, aempty=None,
afull_th=None, aempty_th=None, ovf=None, udf=None, count=None,
count_max=None, depth=None, width=None):
if width == None:
width = 0
if din is not None:
width = len(din)
if depth == None:
depth = 2
full_flg = Signal(bool(1))
empty_flg = Signal(bool(1))
we_safe = Signal(bool(0))
re_safe = Signal(bool(0))
rd_ptr = Signal(intbv(0, min=0, max=depth))
rd_ptr_new = Signal(intbv(0, min=0, max=depth))
wr_ptr = Signal(intbv(0, min=0, max=depth))
wr_ptr_new = Signal(intbv(0, min=0, max=depth))
@always_comb
def x():
full.next = full_flg
empty.next = empty_flg
we_safe.next = we and not full_flg
re_safe.next = re and not empty_flg
@always_comb
def x():
rd_ptr_new.next = (rd_ptr + 1) % depth
wr_ptr_new.next = (wr_ptr + 1) % depth
@always(clk.posedge)
def x():
if rst:
wr_ptr.next = 0
rd_ptr.next = 0
full_flg.next = 0
empty_flg.next = 1
else:
if we_safe:
wr_ptr.next = wr_ptr_new
if re_safe:
rd_ptr.next = rd_ptr_new
if we_safe:
empty_flg.next = 0
elif re_safe and rd_ptr_new == wr_ptr:
empty_flg.next = 1
if re_safe:
full_flg.next = 0
elif we_safe and wr_ptr_new == rd_ptr:
full_flg.next = 1
""" Count """
if count != None or count_max != None or afull != None or aempty != None:
count_r = Signal(intbv(0, min=0, max=depth + 1))
count_new = Signal(intbv(0, min=-1, max=depth + 2))
if count != None:
assert count.max > depth
@always_comb
def x():
count.next = count_r
@always_comb
def x():
if we_safe and not re_safe:
count_new.next = count_r + 1
elif not we_safe and re_safe:
count_new.next = count_r - 1
else:
count_new.next = count_r
@always(clk.posedge)
def x():
if rst:
count_r.next = 0
else:
count_r.next = count_new
""" Count max """
if count_max != None:
assert count_max.max > depth
count_max_r = Signal(intbv(0, min=0, max=count_max.max))
@always(clk.posedge)
def x():
if rst:
count_max_r.next = 0
elif count_max_r < count_new:
count_max_r.next = count_new
@always_comb
def x():
count_max.next = count_max_r
""" AlmostFull flag """
if afull != None:
if afull_th == None:
afull_th = depth // 2
@always(clk.posedge)
def x():
if rst:
afull.next = 0
else:
afull.next = count_new >= depth - afull_th
""" AlmostEmpty flag """
if aempty != None:
if aempty_th == None:
aempty_th = depth // 2
@always(clk.posedge)
def x():
if rst:
aempty.next = 1
else:
aempty.next = count_new <= aempty_th
""" Overflow flag """
if ovf != None:
@always(clk.posedge)
def x():
if rst:
ovf.next = 0
elif we and full_flg:
ovf.next = 1
""" Underflow flag """
if udf != None:
@always(clk.posedge)
def x():
if rst:
udf.next = 0
elif re and empty_flg:
udf.next = 1
if width > 0:
mem_we = Signal(bool(0))
mem_addrw = Signal(intbv(0, min=0, max=depth))
mem_addrr = Signal(intbv(0, min=0, max=depth))
mem_di = Signal(intbv(0)[width:0])
mem_do = Signal(intbv(0)[width:0])
mem = ram_sdp_ar(clk=clk, we=mem_we, addrw=mem_addrw, addrr=
mem_addrr, di=mem_di, do=mem_do)
@always_comb
def x():
mem_we.next = we_safe
mem_addrw.next = wr_ptr
mem_addrr.next = rd_ptr
mem_di.next = din
dout.next = mem_do
return instances()
| fifo | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(rst, clk, full, we, din, empty, re, dout, afull=None, aempty=None,\n afull_th=None, aempty_th=None, ovf=None, udf=None, count=None,\n count_max=None, depth=None, width=None):\n if width == None:\n width = 0\n if din is not None:\n width = len(din)\n if depth == None:\n depth = 2\n full_flg = Signal(bool(1))\n empty_flg = Signal(bool(1))\n we_safe = Signal(bool(0))\n re_safe = Signal(bool(0))\n rd_ptr = Signal(intbv(0, min=0, max=depth))\n rd_ptr_new = Signal(intbv(0, min=0, max=depth))\n wr_ptr = Signal(intbv(0, min=0, max=depth))\n wr_ptr_new = Signal(intbv(0, min=0, max=depth))\n\n @always_comb\n def x():\n full.next = full_flg\n empty.next = empty_flg\n we_safe.next = we and not full_flg\n re_safe.next = re and not empty_flg\n\n @always_comb\n def x():\n rd_ptr_new.next = (rd_ptr + 1) % depth\n wr_ptr_new.next = (wr_ptr + 1) % depth\n\n @always(clk.posedge)\n def x():\n if rst:\n wr_ptr.next = 0\n rd_ptr.next = 0\n full_flg.next = 0\n empty_flg.next = 1\n else:\n if we_safe:\n wr_ptr.next = wr_ptr_new\n if re_safe:\n rd_ptr.next = rd_ptr_new\n if we_safe:\n empty_flg.next = 0\n elif re_safe and rd_ptr_new == wr_ptr:\n empty_flg.next = 1\n if re_safe:\n full_flg.next = 0\n elif we_safe and wr_ptr_new == rd_ptr:\n full_flg.next = 1\n \"\"\" Count \"\"\"\n if count != None or count_max != None or afull != None or aempty != None:\n count_r = Signal(intbv(0, min=0, max=depth + 1))\n count_new = Signal(intbv(0, min=-1, max=depth + 2))\n if count != None:\n assert count.max > depth\n\n @always_comb\n def x():\n count.next = count_r\n\n @always_comb\n def x():\n if we_safe and not re_safe:\n count_new.next = count_r + 1\n elif not we_safe and re_safe:\n count_new.next = count_r - 1\n else:\n count_new.next = count_r\n\n @always(clk.posedge)\n def x():\n if rst:\n count_r.next = 0\n else:\n count_r.next = count_new\n \"\"\" Count max \"\"\"\n if count_max != None:\n assert count_max.max > depth\n count_max_r = Signal(intbv(0, min=0, max=count_max.max))\n\n @always(clk.posedge)\n def x():\n if rst:\n count_max_r.next = 0\n elif count_max_r < count_new:\n count_max_r.next = count_new\n\n @always_comb\n def x():\n count_max.next = count_max_r\n \"\"\" AlmostFull flag \"\"\"\n if afull != None:\n if afull_th == None:\n afull_th = depth // 2\n\n @always(clk.posedge)\n def x():\n if rst:\n afull.next = 0\n else:\n afull.next = count_new >= depth - afull_th\n \"\"\" AlmostEmpty flag \"\"\"\n if aempty != None:\n if aempty_th == None:\n aempty_th = depth // 2\n\n @always(clk.posedge)\n def x():\n if rst:\n aempty.next = 1\n else:\n aempty.next = count_new <= aempty_th\n \"\"\" Overflow flag \"\"\"\n if ovf != None:\n\n @always(clk.posedge)\n def x():\n if rst:\n ovf.next = 0\n elif we and full_flg:\n ovf.next = 1\n \"\"\" Underflow flag \"\"\"\n if udf != None:\n\n @always(clk.posedge)\n def x():\n if rst:\n udf.next = 0\n elif re and empty_flg:\n udf.next = 1\n if width > 0:\n mem_we = Signal(bool(0))\n mem_addrw = Signal(intbv(0, min=0, max=depth))\n mem_addrr = Signal(intbv(0, min=0, max=depth))\n mem_di = Signal(intbv(0)[width:0])\n mem_do = Signal(intbv(0)[width:0])\n mem = ram_sdp_ar(clk=clk, we=mem_we, addrw=mem_addrw, addrr=\n mem_addrr, di=mem_di, do=mem_do)\n\n @always_comb\n def x():\n mem_we.next = we_safe\n mem_addrw.next = wr_ptr\n mem_addrr.next = rd_ptr\n mem_di.next = din\n dout.next = mem_do\n return instances()\n",
"role": "user"
},
{
"content": "fifo",
"role": "assistant"
}
] |
def x(url, ftp=None):
assert isinstance(url, (str, _oldstr))
if ftp is not None:
assert isinstance(ftp, ftplib.FTP)
close_connection = False
ftp_server = 'ftp.ensembl.org'
ftp_user = 'anonymous'
if ftp is None:
ftp = ftplib.FTP(ftp_server)
ftp.login(ftp_user)
close_connection = True
data = []
ftp.retrbinary('RETR %s' % url, data.append)
data = ''.join(d.decode('utf-8') for d in data).split('\n')[:-1]
file_checksums = OrderedDict()
for d in data:
file_name = d[d.rindex(' ') + 1:]
sum_ = int(d[:d.index(' ')])
file_checksums[file_name] = sum_
logger.debug('Obtained checksums for %d files', len(file_checksums))
if close_connection:
ftp.close()
return file_checksums
| get_file_checksums | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(url, ftp=None):\n assert isinstance(url, (str, _oldstr))\n if ftp is not None:\n assert isinstance(ftp, ftplib.FTP)\n close_connection = False\n ftp_server = 'ftp.ensembl.org'\n ftp_user = 'anonymous'\n if ftp is None:\n ftp = ftplib.FTP(ftp_server)\n ftp.login(ftp_user)\n close_connection = True\n data = []\n ftp.retrbinary('RETR %s' % url, data.append)\n data = ''.join(d.decode('utf-8') for d in data).split('\\n')[:-1]\n file_checksums = OrderedDict()\n for d in data:\n file_name = d[d.rindex(' ') + 1:]\n sum_ = int(d[:d.index(' ')])\n file_checksums[file_name] = sum_\n logger.debug('Obtained checksums for %d files', len(file_checksums))\n if close_connection:\n ftp.close()\n return file_checksums\n",
"role": "user"
},
{
"content": "get_file_checksums",
"role": "assistant"
}
] |
def x(obj, ignore=(list, tuple, type(None))):
return obj if isinstance(obj, ignore) else [obj]
| listify | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(obj, ignore=(list, tuple, type(None))):\n return obj if isinstance(obj, ignore) else [obj]\n",
"role": "user"
},
{
"content": "listify",
"role": "assistant"
}
] |
def x(taxdump_file):
with tarfile.open(taxdump_file) as tf:
with tf.extractfile('division.dmp') as fh:
df = pd.read_csv(fh, header=None, sep='|', encoding='ascii')
df = df.iloc[:, [0, 2]]
df.iloc[:, 1] = df.iloc[:, 1].str.strip('\t')
divisions = {}
for _, row in df.iterrows():
divisions[row.iloc[1]] = row.iloc[0]
return divisions
| get_divisions | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(taxdump_file):\n with tarfile.open(taxdump_file) as tf:\n with tf.extractfile('division.dmp') as fh:\n df = pd.read_csv(fh, header=None, sep='|', encoding='ascii')\n df = df.iloc[:, [0, 2]]\n df.iloc[:, 1] = df.iloc[:, 1].str.strip('\\t')\n divisions = {}\n for _, row in df.iterrows():\n divisions[row.iloc[1]] = row.iloc[0]\n return divisions\n",
"role": "user"
},
{
"content": "get_divisions",
"role": "assistant"
}
] |
def x(taxdump_file, select_divisions=None, exclude_divisions=None):
if select_divisions and exclude_divisions:
raise ValueError(
'Cannot specify "select_divisions" and "exclude_divisions" at the same time.'
)
select_division_ids = None
exclude_division_ids = None
divisions = None
if select_divisions or exclude_divisions:
divisions = _get_divisions(taxdump_file)
if select_divisions:
select_division_ids = set([divisions[d] for d in select_divisions])
elif exclude_divisions:
exclude_division_ids = set([divisions[d] for d in exclude_divisions])
with tarfile.open(taxdump_file) as tf:
with tf.extractfile('nodes.dmp') as fh:
df = pd.read_csv(fh, header=None, sep='|', encoding='ascii')
df = df.iloc[:, [0, 2, 4]]
if select_division_ids:
df = df.loc[df.iloc[:, 2].isin(select_division_ids)]
elif exclude_division_ids:
df = df.loc[~df.iloc[:, 2].isin(exclude_division_ids)]
df.iloc[:, 1] = df.iloc[:, 1].str.strip('\t')
taxon_ids = df.iloc[:, 0].loc[df.iloc[:, 1] == 'species'].values
return taxon_ids
| get_species_taxon_ids | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(taxdump_file, select_divisions=None, exclude_divisions=None):\n if select_divisions and exclude_divisions:\n raise ValueError(\n 'Cannot specify \"select_divisions\" and \"exclude_divisions\" at the same time.'\n )\n select_division_ids = None\n exclude_division_ids = None\n divisions = None\n if select_divisions or exclude_divisions:\n divisions = _get_divisions(taxdump_file)\n if select_divisions:\n select_division_ids = set([divisions[d] for d in select_divisions])\n elif exclude_divisions:\n exclude_division_ids = set([divisions[d] for d in exclude_divisions])\n with tarfile.open(taxdump_file) as tf:\n with tf.extractfile('nodes.dmp') as fh:\n df = pd.read_csv(fh, header=None, sep='|', encoding='ascii')\n df = df.iloc[:, [0, 2, 4]]\n if select_division_ids:\n df = df.loc[df.iloc[:, 2].isin(select_division_ids)]\n elif exclude_division_ids:\n df = df.loc[~df.iloc[:, 2].isin(exclude_division_ids)]\n df.iloc[:, 1] = df.iloc[:, 1].str.strip('\\t')\n taxon_ids = df.iloc[:, 0].loc[df.iloc[:, 1] == 'species'].values\n return taxon_ids\n",
"role": "user"
},
{
"content": "get_species_taxon_ids",
"role": "assistant"
}
] |
def x(taxdump_file, select_divisions=None, exclude_divisions=None, nrows=None):
if select_divisions and exclude_divisions:
raise ValueError(
'Cannot specify "select_divisions" and "exclude_divisions" at the same time.'
)
select_taxon_ids = _get_species_taxon_ids(taxdump_file,
select_divisions=select_divisions, exclude_divisions=exclude_divisions)
select_taxon_ids = set(select_taxon_ids)
with tarfile.open(taxdump_file) as tf:
with tf.extractfile('names.dmp') as fh:
df = pd.read_csv(fh, header=None, sep='|', encoding='ascii',
nrows=nrows)
df = df.iloc[:, [0, 1, 3]]
df = df.loc[df.iloc[:, 0].isin(select_taxon_ids)]
df.iloc[:, 2] = df.iloc[:, 2].str.strip('\t')
df = df.loc[df.iloc[:, 2].isin(['scientific name', 'common name'])]
df.iloc[:, 1] = df.iloc[:, 1].str.strip('\t')
common_names = defaultdict(list)
cn = df.loc[df.iloc[:, 2] == 'common name']
for _, row in cn.iterrows():
common_names[row.iloc[0]].append(row.iloc[1])
sn = df.loc[df.iloc[:, 2] == 'scientific name']
species = []
for i, row in sn.iterrows():
species.append([row.iloc[0], row.iloc[1], '|'.join(common_names[row
.iloc[0]])])
species_df = pd.DataFrame(species).set_index(0)
species_df.columns = ['scientific_name', 'common_names']
species_df.index.name = 'taxon_id'
return species_df
| get_species | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(taxdump_file, select_divisions=None, exclude_divisions=None, nrows=None):\n if select_divisions and exclude_divisions:\n raise ValueError(\n 'Cannot specify \"select_divisions\" and \"exclude_divisions\" at the same time.'\n )\n select_taxon_ids = _get_species_taxon_ids(taxdump_file,\n select_divisions=select_divisions, exclude_divisions=exclude_divisions)\n select_taxon_ids = set(select_taxon_ids)\n with tarfile.open(taxdump_file) as tf:\n with tf.extractfile('names.dmp') as fh:\n df = pd.read_csv(fh, header=None, sep='|', encoding='ascii',\n nrows=nrows)\n df = df.iloc[:, [0, 1, 3]]\n df = df.loc[df.iloc[:, 0].isin(select_taxon_ids)]\n df.iloc[:, 2] = df.iloc[:, 2].str.strip('\\t')\n df = df.loc[df.iloc[:, 2].isin(['scientific name', 'common name'])]\n df.iloc[:, 1] = df.iloc[:, 1].str.strip('\\t')\n common_names = defaultdict(list)\n cn = df.loc[df.iloc[:, 2] == 'common name']\n for _, row in cn.iterrows():\n common_names[row.iloc[0]].append(row.iloc[1])\n sn = df.loc[df.iloc[:, 2] == 'scientific name']\n species = []\n for i, row in sn.iterrows():\n species.append([row.iloc[0], row.iloc[1], '|'.join(common_names[row\n .iloc[0]])])\n species_df = pd.DataFrame(species).set_index(0)\n species_df.columns = ['scientific_name', 'common_names']\n species_df.index.name = 'taxon_id'\n return species_df\n",
"role": "user"
},
{
"content": "get_species",
"role": "assistant"
}
] |
def x(self, **attributes):
auto_set = IxeObject.get_auto_set()
IxeObject.set_auto_set(False)
for name, value in attributes.items():
setattr(self, name, value)
if auto_set:
self.ix_set()
IxeObject.set_auto_set(auto_set)
| set_attributes | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, **attributes):\n auto_set = IxeObject.get_auto_set()\n IxeObject.set_auto_set(False)\n for name, value in attributes.items():\n setattr(self, name, value)\n if auto_set:\n self.ix_set()\n IxeObject.set_auto_set(auto_set)\n",
"role": "user"
},
{
"content": "set_attributes",
"role": "assistant"
}
] |
def x(recordFilename, hitId, fastaFilename, eCutoff=None, db='nt',
actualSequenceId=None):
print('TODO: This function is not finished yet.')
return
start = time()
if isinstance(recordFilename, str):
allhits = findHits(recordFilename, set([hitId]), limit=100)
else:
allhits = recordFilename
sequence = getSequence(hitId, db)
fasta, summary = summarizeHits(allhits, fastaFilename, eCutoff=eCutoff)
minX, maxX = summary['minX'], summary['maxX']
if actualSequenceId:
pass
print(summary['hitCount'])
print('seq len =', len(sequence))
fasta = summary['fasta']
consensusLen = maxX - minX
consensus = [None] * consensusLen
for item in summary['items']:
print('NEW HSP')
printHSP(item['origHsp'])
hsp = item['hsp']
print('HIT query-start=%d query-stop=%d subj-start=%d subj-stop=%d' %
(hsp['queryStart'], hsp['queryEnd'], hsp['subjectStart'], hsp[
'subjectEnd']))
if item['frame']['subject'] > 0:
query = fasta[item['sequenceId']].seq
else:
query = fasta[item['sequenceId']].reverse_complement().seq
print(' target:', sequence[hsp['queryStart']:hsp['queryEnd']].seq
)
print(' query:', query)
match = []
for index in range(hsp['subjectStart'], hsp['subjectEnd']):
queryIndex = index - hsp['queryStart']
match.append('.' if query[queryIndex] == sequence[index] else '*')
print(' match: %s%s' % (' ' * (hsp['subjectStart'] - hsp[
'queryStart']), ''.join(match)))
print(' score:', item['convertedE'])
print(' match len:', hsp['subjectEnd'] - hsp['subjectStart'])
print('subject frame:', item['frame']['subject'])
for queryIndex, sequenceIndex in enumerate(range(hsp['queryStart'],
hsp['queryEnd'])):
consensusIndex = sequenceIndex + minX
locus = consensus[consensusIndex]
if locus is None:
consensus[consensusIndex] = locus = defaultdict(int)
locus[query[queryIndex]] += 1
for index in range(minX, 0):
consensusIndex = index - minX
if consensus[consensusIndex]:
print('%d: %r' % (index, consensus[consensusIndex]))
for index in range(0, len(sequence)):
consensusIndex = index - minX
try:
if consensus[consensusIndex]:
print('%d: %r (%s)' % (index, consensus[index], sequence[
index]))
except KeyError:
break
for index in range(len(sequence), maxX):
consensusIndex = index - minX
if consensus[consensusIndex]:
print('%d: %r' % (index, consensus[consensusIndex]))
stop = time()
report('Consensus sequence generated in %.3f mins.' % ((stop - start) /
60.0))
return summary, consensus
| consensusSequence | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(recordFilename, hitId, fastaFilename, eCutoff=None, db='nt',\n actualSequenceId=None):\n print('TODO: This function is not finished yet.')\n return\n start = time()\n if isinstance(recordFilename, str):\n allhits = findHits(recordFilename, set([hitId]), limit=100)\n else:\n allhits = recordFilename\n sequence = getSequence(hitId, db)\n fasta, summary = summarizeHits(allhits, fastaFilename, eCutoff=eCutoff)\n minX, maxX = summary['minX'], summary['maxX']\n if actualSequenceId:\n pass\n print(summary['hitCount'])\n print('seq len =', len(sequence))\n fasta = summary['fasta']\n consensusLen = maxX - minX\n consensus = [None] * consensusLen\n for item in summary['items']:\n print('NEW HSP')\n printHSP(item['origHsp'])\n hsp = item['hsp']\n print('HIT query-start=%d query-stop=%d subj-start=%d subj-stop=%d' %\n (hsp['queryStart'], hsp['queryEnd'], hsp['subjectStart'], hsp[\n 'subjectEnd']))\n if item['frame']['subject'] > 0:\n query = fasta[item['sequenceId']].seq\n else:\n query = fasta[item['sequenceId']].reverse_complement().seq\n print(' target:', sequence[hsp['queryStart']:hsp['queryEnd']].seq\n )\n print(' query:', query)\n match = []\n for index in range(hsp['subjectStart'], hsp['subjectEnd']):\n queryIndex = index - hsp['queryStart']\n match.append('.' if query[queryIndex] == sequence[index] else '*')\n print(' match: %s%s' % (' ' * (hsp['subjectStart'] - hsp[\n 'queryStart']), ''.join(match)))\n print(' score:', item['convertedE'])\n print(' match len:', hsp['subjectEnd'] - hsp['subjectStart'])\n print('subject frame:', item['frame']['subject'])\n for queryIndex, sequenceIndex in enumerate(range(hsp['queryStart'],\n hsp['queryEnd'])):\n consensusIndex = sequenceIndex + minX\n locus = consensus[consensusIndex]\n if locus is None:\n consensus[consensusIndex] = locus = defaultdict(int)\n locus[query[queryIndex]] += 1\n for index in range(minX, 0):\n consensusIndex = index - minX\n if consensus[consensusIndex]:\n print('%d: %r' % (index, consensus[consensusIndex]))\n for index in range(0, len(sequence)):\n consensusIndex = index - minX\n try:\n if consensus[consensusIndex]:\n print('%d: %r (%s)' % (index, consensus[index], sequence[\n index]))\n except KeyError:\n break\n for index in range(len(sequence), maxX):\n consensusIndex = index - minX\n if consensus[consensusIndex]:\n print('%d: %r' % (index, consensus[consensusIndex]))\n stop = time()\n report('Consensus sequence generated in %.3f mins.' % ((stop - start) /\n 60.0))\n return summary, consensus\n",
"role": "user"
},
{
"content": "consensusSequence",
"role": "assistant"
}
] |
def x(self, posicao=0):
handle = self._url_open('detalheCEPAction.do', {'Metodo': 'detalhe',
'TipoCep': 2, 'Posicao': posicao + 1, 'CEP': ''})
html = handle.read()
return self._parse_detalhe(html)
| detalhe | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, posicao=0):\n handle = self._url_open('detalheCEPAction.do', {'Metodo': 'detalhe',\n 'TipoCep': 2, 'Posicao': posicao + 1, 'CEP': ''})\n html = handle.read()\n return self._parse_detalhe(html)\n",
"role": "user"
},
{
"content": "detalhe",
"role": "assistant"
}
] |
def x(self, localidade, uf):
url = 'consultaFaixaCepAction.do'
data = {'UF': uf, 'Localidade': localidade.encode('cp1252'), 'cfm': '1',
'Metodo': 'listaFaixaCEP', 'TipoConsulta': 'faixaCep', 'StartRow':
'1', 'EndRow': '10'}
html = self._url_open(url, data).read()
return self._parse_faixa(html)
| consulta_faixa | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, localidade, uf):\n url = 'consultaFaixaCepAction.do'\n data = {'UF': uf, 'Localidade': localidade.encode('cp1252'), 'cfm': '1',\n 'Metodo': 'listaFaixaCEP', 'TipoConsulta': 'faixaCep', 'StartRow':\n '1', 'EndRow': '10'}\n html = self._url_open(url, data).read()\n return self._parse_faixa(html)\n",
"role": "user"
},
{
"content": "consulta_faixa",
"role": "assistant"
}
] |
def x(self, endereco, primeiro=False, uf=None, localidade=None, tipo=None,
numero=None):
if uf is None:
url = 'consultaEnderecoAction.do'
data = {'relaxation': endereco.encode('ISO-8859-1'), 'TipoCep':
'ALL', 'semelhante': 'N', 'cfm': 1, 'Metodo': 'listaLogradouro',
'TipoConsulta': 'relaxation', 'StartRow': '1', 'EndRow': '10'}
else:
url = 'consultaLogradouroAction.do'
data = {'Logradouro': endereco.encode('ISO-8859-1'), 'UF': uf,
'TIPO': tipo, 'Localidade': localidade.encode('ISO-8859-1'),
'Numero': numero, 'cfm': 1, 'Metodo': 'listaLogradouro',
'TipoConsulta': 'logradouro', 'StartRow': '1', 'EndRow': '10'}
h = self._url_open(url, data)
html = h.read()
if primeiro:
return self.detalhe()
else:
return self._parse_tabela(html)
| consulta | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, endereco, primeiro=False, uf=None, localidade=None, tipo=None,\n numero=None):\n if uf is None:\n url = 'consultaEnderecoAction.do'\n data = {'relaxation': endereco.encode('ISO-8859-1'), 'TipoCep':\n 'ALL', 'semelhante': 'N', 'cfm': 1, 'Metodo': 'listaLogradouro',\n 'TipoConsulta': 'relaxation', 'StartRow': '1', 'EndRow': '10'}\n else:\n url = 'consultaLogradouroAction.do'\n data = {'Logradouro': endereco.encode('ISO-8859-1'), 'UF': uf,\n 'TIPO': tipo, 'Localidade': localidade.encode('ISO-8859-1'),\n 'Numero': numero, 'cfm': 1, 'Metodo': 'listaLogradouro',\n 'TipoConsulta': 'logradouro', 'StartRow': '1', 'EndRow': '10'}\n h = self._url_open(url, data)\n html = h.read()\n if primeiro:\n return self.detalhe()\n else:\n return self._parse_tabela(html)\n",
"role": "user"
},
{
"content": "consulta",
"role": "assistant"
}
] |
def x(self, url, *, method='GET', headers=None, data=None, result_callback=None
):
retries_left = self._conn_retries
while True:
s = self._make_session()
try:
cert = None
if self._client_cert and self._client_key:
cert = self._client_cert, self._client_key
elif self._client_cert:
cert = self._client_cert
if self._verify_cert:
verify = True
if self._ca_certs:
verify = self._ca_certs
else:
verify = False
auth = None
if self._username and self._password:
auth = self._username, self._password
response = s.request(method, url, data=data, timeout=self.
_connect_timeout, cert=cert, headers=headers, verify=verify,
auth=auth)
""":type: requests.models.Response
"""
if 400 <= response.status_code < 500:
raise ClientError(response.status_code, response.content)
elif response.status_code >= 500:
raise ServerError(response.status_code, response.content)
try:
if result_callback:
return result_callback(response.content)
except (ValueError, TypeError) as err:
raise MalformedResponse(err) from None
return response.content
except (requests.exceptions.RequestException, requests.exceptions.
BaseHTTPError) as exc:
if self._conn_retries is None or retries_left <= 0:
raise CommunicationError(exc) from None
else:
retries_left -= 1
retry_in = (self._conn_retries - retries_left) * 2
self._log.warning(
'Server communication error: %s. Retrying in %s seconds.',
exc, retry_in)
time.sleep(retry_in)
continue
finally:
s.close()
| request | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, url, *, method='GET', headers=None, data=None, result_callback=None\n ):\n retries_left = self._conn_retries\n while True:\n s = self._make_session()\n try:\n cert = None\n if self._client_cert and self._client_key:\n cert = self._client_cert, self._client_key\n elif self._client_cert:\n cert = self._client_cert\n if self._verify_cert:\n verify = True\n if self._ca_certs:\n verify = self._ca_certs\n else:\n verify = False\n auth = None\n if self._username and self._password:\n auth = self._username, self._password\n response = s.request(method, url, data=data, timeout=self.\n _connect_timeout, cert=cert, headers=headers, verify=verify,\n auth=auth)\n \"\"\":type: requests.models.Response\n \"\"\"\n if 400 <= response.status_code < 500:\n raise ClientError(response.status_code, response.content)\n elif response.status_code >= 500:\n raise ServerError(response.status_code, response.content)\n try:\n if result_callback:\n return result_callback(response.content)\n except (ValueError, TypeError) as err:\n raise MalformedResponse(err) from None\n return response.content\n except (requests.exceptions.RequestException, requests.exceptions.\n BaseHTTPError) as exc:\n if self._conn_retries is None or retries_left <= 0:\n raise CommunicationError(exc) from None\n else:\n retries_left -= 1\n retry_in = (self._conn_retries - retries_left) * 2\n self._log.warning(\n 'Server communication error: %s. Retrying in %s seconds.',\n exc, retry_in)\n time.sleep(retry_in)\n continue\n finally:\n s.close()\n",
"role": "user"
},
{
"content": "request",
"role": "assistant"
}
] |
def x():
sess = requests.Session()
sess.mount('http://', requests.adapters.HTTPAdapter(max_retries=False))
sess.mount('https://', requests.adapters.HTTPAdapter(max_retries=False))
return sess
| make_session | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n sess = requests.Session()\n sess.mount('http://', requests.adapters.HTTPAdapter(max_retries=False))\n sess.mount('https://', requests.adapters.HTTPAdapter(max_retries=False))\n return sess\n",
"role": "user"
},
{
"content": "make_session",
"role": "assistant"
}
] |
def x(tcga_id):
payload = {'attachment': 'true', 'filters': json.dumps({'op': 'and',
'content': [{'op': 'in', 'content': {'field':
'cases.project.program.name', 'value': ['TCGA']}}, {'op': 'in',
'content': {'field': 'project.project_id', 'value': [tcga_id]}}]}),
'fields': 'case_id', 'expand':
'demographic,diagnoses,family_histories,exposures', 'format':
'JSON', 'pretty': 'true', 'size': 10000, 'filename':
'clinical.project-%s' % tcga_id}
r = requests.get('https://gdc-api.nci.nih.gov/cases', params=payload)
j = json.loads(r.content.decode())
clinical = {}
valid = 0
for s in j:
if 'diagnoses' not in s:
continue
valid += 1
assert len(s['diagnoses']) == 1
diag = s['diagnoses'][0]
tcga_id = diag['submitter_id'][:12]
clinical[tcga_id] = diag
logger.info('Found clinical data for %d cases.', valid)
df = pd.DataFrame.from_dict(clinical).T
df.sort_index(inplace=True)
return df
| get_clinical_data | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(tcga_id):\n payload = {'attachment': 'true', 'filters': json.dumps({'op': 'and',\n 'content': [{'op': 'in', 'content': {'field':\n 'cases.project.program.name', 'value': ['TCGA']}}, {'op': 'in',\n 'content': {'field': 'project.project_id', 'value': [tcga_id]}}]}),\n 'fields': 'case_id', 'expand':\n 'demographic,diagnoses,family_histories,exposures', 'format':\n 'JSON', 'pretty': 'true', 'size': 10000, 'filename': \n 'clinical.project-%s' % tcga_id}\n r = requests.get('https://gdc-api.nci.nih.gov/cases', params=payload)\n j = json.loads(r.content.decode())\n clinical = {}\n valid = 0\n for s in j:\n if 'diagnoses' not in s:\n continue\n valid += 1\n assert len(s['diagnoses']) == 1\n diag = s['diagnoses'][0]\n tcga_id = diag['submitter_id'][:12]\n clinical[tcga_id] = diag\n logger.info('Found clinical data for %d cases.', valid)\n df = pd.DataFrame.from_dict(clinical).T\n df.sort_index(inplace=True)\n return df\n",
"role": "user"
},
{
"content": "get_clinical_data",
"role": "assistant"
}
] |
def x(tcga_id):
payload = {'attachment': 'true', 'filters': json.dumps({'op': 'and',
'content': [{'op': 'in', 'content': {'field':
'cases.project.program.name', 'value': ['TCGA']}}, {'op': 'in',
'content': {'field': 'project.project_id', 'value': [tcga_id]}}]}),
'fields': 'case_id', 'expand':
'samples,samples.portions,samples.portions.analytes,samples.portions.analytes.aliquots,samples.portions.analytes.aliquots.annotations,samples.portions.analytes.annotations,samples.portions.submitter_id,samples.portions.slides,samples.portions.annotations,samples.portions.center'
, 'format': 'JSON', 'pretty': 'true', 'size': 10000, 'filename':
'biospecimen.project-%s' % tcga_id}
r = requests.get('https://gdc-api.nci.nih.gov/cases', params=payload)
j = json.loads(r.content.decode())
biospec = {}
valid = 0
for case in j:
if 'samples' not in case:
continue
valid += 1
for s in case['samples']:
tcga_id = s['submitter_id'][:16]
del s['portions']
del s['submitter_id']
biospec[tcga_id] = s
logger.info('Found biospecimen data for %d cases.', valid)
df = pd.DataFrame.from_dict(biospec).T
df.sort_index(inplace=True)
return df
| get_biospecimen_data | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(tcga_id):\n payload = {'attachment': 'true', 'filters': json.dumps({'op': 'and',\n 'content': [{'op': 'in', 'content': {'field':\n 'cases.project.program.name', 'value': ['TCGA']}}, {'op': 'in',\n 'content': {'field': 'project.project_id', 'value': [tcga_id]}}]}),\n 'fields': 'case_id', 'expand':\n 'samples,samples.portions,samples.portions.analytes,samples.portions.analytes.aliquots,samples.portions.analytes.aliquots.annotations,samples.portions.analytes.annotations,samples.portions.submitter_id,samples.portions.slides,samples.portions.annotations,samples.portions.center'\n , 'format': 'JSON', 'pretty': 'true', 'size': 10000, 'filename': \n 'biospecimen.project-%s' % tcga_id}\n r = requests.get('https://gdc-api.nci.nih.gov/cases', params=payload)\n j = json.loads(r.content.decode())\n biospec = {}\n valid = 0\n for case in j:\n if 'samples' not in case:\n continue\n valid += 1\n for s in case['samples']:\n tcga_id = s['submitter_id'][:16]\n del s['portions']\n del s['submitter_id']\n biospec[tcga_id] = s\n logger.info('Found biospecimen data for %d cases.', valid)\n df = pd.DataFrame.from_dict(biospec).T\n df.sort_index(inplace=True)\n return df\n",
"role": "user"
},
{
"content": "get_biospecimen_data",
"role": "assistant"
}
] |
def x(tcga_id):
payload = {'filters': json.dumps({'op': 'and', 'content': [{'op': 'in',
'content': {'field': 'cases.project.program.name', 'value': ['TCGA'
]}}, {'op': 'in', 'content': {'field': 'cases.project.project_id',
'value': [tcga_id]}}, {'op': 'in', 'content': {'field':
'files.data_category', 'value': ['Copy Number Variation']}}, {'op':
'in', 'content': {'field': 'files.data_type', 'value': [
'Masked Copy Number Segment']}}]}), 'return_type': 'manifest',
'size': 10000}
r = requests.get('https://gdc-api.nci.nih.gov/files', params=payload)
df = pd.read_csv(io.StringIO(r.text), sep='\t', header=0)
logger.info('Obtained manifest with %d files.', df.shape[0])
return df
| get_masked_cnv_manifest | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(tcga_id):\n payload = {'filters': json.dumps({'op': 'and', 'content': [{'op': 'in',\n 'content': {'field': 'cases.project.program.name', 'value': ['TCGA'\n ]}}, {'op': 'in', 'content': {'field': 'cases.project.project_id',\n 'value': [tcga_id]}}, {'op': 'in', 'content': {'field':\n 'files.data_category', 'value': ['Copy Number Variation']}}, {'op':\n 'in', 'content': {'field': 'files.data_type', 'value': [\n 'Masked Copy Number Segment']}}]}), 'return_type': 'manifest',\n 'size': 10000}\n r = requests.get('https://gdc-api.nci.nih.gov/files', params=payload)\n df = pd.read_csv(io.StringIO(r.text), sep='\\t', header=0)\n logger.info('Obtained manifest with %d files.', df.shape[0])\n return df\n",
"role": "user"
},
{
"content": "get_masked_cnv_manifest",
"role": "assistant"
}
] |
def x(file_ids):
assert isinstance(file_ids, Iterable)
payload = {'filters': json.dumps({'op': 'in', 'content': {'field':
'files.file_id', 'value': list(file_ids)}}), 'fields':
'file_id,cases.samples.submitter_id', 'size': 10000}
r = requests.post('https://gdc-api.nci.nih.gov/files', data=payload)
j = json.loads(r.content.decode('utf-8'))
file_samples = OrderedDict()
for hit in j['data']['hits']:
file_id = hit['file_id']
assert len(hit['cases']) == 1
case = hit['cases'][0]
assert len(case['samples']) == 1
sample = case['samples'][0]
sample_barcode = sample['submitter_id']
file_samples[file_id] = sample_barcode
df = pd.DataFrame.from_dict(file_samples, orient='index')
df = df.reset_index()
df.columns = ['file_id', 'sample_barcode']
return df
| get_file_samples | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(file_ids):\n assert isinstance(file_ids, Iterable)\n payload = {'filters': json.dumps({'op': 'in', 'content': {'field':\n 'files.file_id', 'value': list(file_ids)}}), 'fields':\n 'file_id,cases.samples.submitter_id', 'size': 10000}\n r = requests.post('https://gdc-api.nci.nih.gov/files', data=payload)\n j = json.loads(r.content.decode('utf-8'))\n file_samples = OrderedDict()\n for hit in j['data']['hits']:\n file_id = hit['file_id']\n assert len(hit['cases']) == 1\n case = hit['cases'][0]\n assert len(case['samples']) == 1\n sample = case['samples'][0]\n sample_barcode = sample['submitter_id']\n file_samples[file_id] = sample_barcode\n df = pd.DataFrame.from_dict(file_samples, orient='index')\n df = df.reset_index()\n df.columns = ['file_id', 'sample_barcode']\n return df\n",
"role": "user"
},
{
"content": "get_file_samples",
"role": "assistant"
}
] |
def x(file_samples):
assert isinstance(file_samples, pd.DataFrame)
df = file_samples
df = df.sort_values('file_id')
logger.info('Original number of files: %d', len(df.index))
df.drop_duplicates('sample_barcode', keep='first', inplace=True)
logger.info(
'Number of files after removing duplicates from different aliquots: %d'
, len(df.index))
df['sample_barcode15'] = df['sample_barcode'].apply(lambda x: x[:15])
df.drop_duplicates('sample_barcode15', keep='first', inplace=True)
logger.info(
'Number of files after removing duplicates from different vials: %d',
len(df.index))
df.drop('sample_barcode15', axis=1, inplace=True)
df.sort_index(inplace=True)
return df
| get_unique_sample_files | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(file_samples):\n assert isinstance(file_samples, pd.DataFrame)\n df = file_samples\n df = df.sort_values('file_id')\n logger.info('Original number of files: %d', len(df.index))\n df.drop_duplicates('sample_barcode', keep='first', inplace=True)\n logger.info(\n 'Number of files after removing duplicates from different aliquots: %d'\n , len(df.index))\n df['sample_barcode15'] = df['sample_barcode'].apply(lambda x: x[:15])\n df.drop_duplicates('sample_barcode15', keep='first', inplace=True)\n logger.info(\n 'Number of files after removing duplicates from different vials: %d',\n len(df.index))\n df.drop('sample_barcode15', axis=1, inplace=True)\n df.sort_index(inplace=True)\n return df\n",
"role": "user"
},
{
"content": "get_unique_sample_files",
"role": "assistant"
}
] |
def x(biospecimen):
df = biospecimen
num_before = len(df.index)
df = df.loc[~df['is_ffpe']]
logger.info(
'Excluded %d files associated with FFPE samples (out of %d files in total).'
, num_before - len(df.index), num_before)
num_before = len(df.index)
df = df.loc[df['sample_type'] == 'Primary Tumor']
logger.info(
'Excluded %d files not corresponding to primary tumor samples (out of %d files in total).'
, num_before - len(df.index), num_before)
return df
| get_fresh_primary_tumors | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(biospecimen):\n df = biospecimen\n num_before = len(df.index)\n df = df.loc[~df['is_ffpe']]\n logger.info(\n 'Excluded %d files associated with FFPE samples (out of %d files in total).'\n , num_before - len(df.index), num_before)\n num_before = len(df.index)\n df = df.loc[df['sample_type'] == 'Primary Tumor']\n logger.info(\n 'Excluded %d files not corresponding to primary tumor samples (out of %d files in total).'\n , num_before - len(df.index), num_before)\n return df\n",
"role": "user"
},
{
"content": "get_fresh_primary_tumors",
"role": "assistant"
}
] |
def x(reads):
seen = set()
add = seen.add
for read in reads:
hash_ = md5(read.sequence.encode('UTF-8')).digest()
if hash_ not in seen:
add(hash_)
yield read
| dedupFasta | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(reads):\n seen = set()\n add = seen.add\n for read in reads:\n hash_ = md5(read.sequence.encode('UTF-8')).digest()\n if hash_ not in seen:\n add(hash_)\n yield read\n",
"role": "user"
},
{
"content": "dedupFasta",
"role": "assistant"
}
] |
def x(sequences):
sequences = sorted(sequences, key=lambda s: len(s.seq), reverse=True)
seen = set()
for s in sequences:
thisSeq = str(s.seq)
thisHash = md5(thisSeq.encode('UTF-8')).digest()
if thisHash not in seen:
newHash = md5()
for nucl in thisSeq:
newHash.update(nucl.encode('UTF-8'))
seen.add(newHash.digest())
for start in range(len(thisSeq) - 1):
seen.add(md5(thisSeq[start + 1:].encode('UTF-8')).digest())
yield s
| dePrefixAndSuffixFasta | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(sequences):\n sequences = sorted(sequences, key=lambda s: len(s.seq), reverse=True)\n seen = set()\n for s in sequences:\n thisSeq = str(s.seq)\n thisHash = md5(thisSeq.encode('UTF-8')).digest()\n if thisHash not in seen:\n newHash = md5()\n for nucl in thisSeq:\n newHash.update(nucl.encode('UTF-8'))\n seen.add(newHash.digest())\n for start in range(len(thisSeq) - 1):\n seen.add(md5(thisSeq[start + 1:].encode('UTF-8')).digest())\n yield s\n",
"role": "user"
},
{
"content": "dePrefixAndSuffixFasta",
"role": "assistant"
}
] |
def x(fastaFiles):
reads = {}
firstFile = fastaFiles.pop(0)
for seq in SeqIO.parse(firstFile, 'fasta'):
reads[seq.id] = seq
for fastaFile in fastaFiles:
for seq in SeqIO.parse(fastaFile, 'fasta'):
if seq.id in reads:
assert str(seq.seq) == str(reads[seq.id].seq)
reads.pop(seq.id, None)
return iter(reads.values())
| fastaSubtract | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fastaFiles):\n reads = {}\n firstFile = fastaFiles.pop(0)\n for seq in SeqIO.parse(firstFile, 'fasta'):\n reads[seq.id] = seq\n for fastaFile in fastaFiles:\n for seq in SeqIO.parse(fastaFile, 'fasta'):\n if seq.id in reads:\n assert str(seq.seq) == str(reads[seq.id].seq)\n reads.pop(seq.id, None)\n return iter(reads.values())\n",
"role": "user"
},
{
"content": "fastaSubtract",
"role": "assistant"
}
] |
def x(filename, sequences, readClass=DNARead, upperCase=False, idPrefix=
'command-line-read-'):
if filename:
reads = FastaReads(filename, readClass=readClass, upperCase=upperCase)
else:
reads = Reads()
if sequences:
for count, sequence in enumerate(sequences, start=1):
parts = sequence.rsplit(' ', 1)
if len(parts) == 2:
readId, sequence = parts
else:
readId = '%s%d' % (idPrefix, count)
if upperCase:
sequence = sequence.upper()
read = readClass(readId, sequence)
reads.add(read)
return reads
| combineReads | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(filename, sequences, readClass=DNARead, upperCase=False, idPrefix=\n 'command-line-read-'):\n if filename:\n reads = FastaReads(filename, readClass=readClass, upperCase=upperCase)\n else:\n reads = Reads()\n if sequences:\n for count, sequence in enumerate(sequences, start=1):\n parts = sequence.rsplit(' ', 1)\n if len(parts) == 2:\n readId, sequence = parts\n else:\n readId = '%s%d' % (idPrefix, count)\n if upperCase:\n sequence = sequence.upper()\n read = readClass(readId, sequence)\n reads.add(read)\n return reads\n",
"role": "user"
},
{
"content": "combineReads",
"role": "assistant"
}
] |
def x(self):
count = 0
for _file in self._files:
with asHandle(_file) as fp:
if self._upperCase:
for seq in SeqIO.parse(fp, 'fasta'):
read = self._readClass(seq.description, str(seq.seq.
upper()))
yield read
count += 1
else:
for seq in SeqIO.parse(fp, 'fasta'):
read = self._readClass(seq.description, str(seq.seq))
yield read
count += 1
| iter | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n count = 0\n for _file in self._files:\n with asHandle(_file) as fp:\n if self._upperCase:\n for seq in SeqIO.parse(fp, 'fasta'):\n read = self._readClass(seq.description, str(seq.seq.\n upper()))\n yield read\n count += 1\n else:\n for seq in SeqIO.parse(fp, 'fasta'):\n read = self._readClass(seq.description, str(seq.seq))\n yield read\n count += 1\n",
"role": "user"
},
{
"content": "iter",
"role": "assistant"
}
] |
def x(self):
if self._upperCase:
for id_ in self._fasta:
yield self._readClass(id_, str(self._fasta[id_]).upper())
else:
for id_ in self._fasta:
yield self._readClass(id_, str(self._fasta[id_]))
| iter | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self._upperCase:\n for id_ in self._fasta:\n yield self._readClass(id_, str(self._fasta[id_]).upper())\n else:\n for id_ in self._fasta:\n yield self._readClass(id_, str(self._fasta[id_]))\n",
"role": "user"
},
{
"content": "iter",
"role": "assistant"
}
] |
def x(self, fileNumber):
cur = self._connection.cursor()
cur.execute('SELECT name FROM files WHERE id = ?', (fileNumber,))
row = cur.fetchone()
if row is None:
return None
else:
return row[0]
| getFilename | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, fileNumber):\n cur = self._connection.cursor()\n cur.execute('SELECT name FROM files WHERE id = ?', (fileNumber,))\n row = cur.fetchone()\n if row is None:\n return None\n else:\n return row[0]\n",
"role": "user"
},
{
"content": "getFilename",
"role": "assistant"
}
] |
def x(self, filename):
cur = self._connection.cursor()
cur.execute('SELECT id FROM files WHERE name = ?', (filename,))
row = cur.fetchone()
if row is None:
return None
else:
return row[0]
| getFileNumber | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, filename):\n cur = self._connection.cursor()\n cur.execute('SELECT id FROM files WHERE name = ?', (filename,))\n row = cur.fetchone()\n if row is None:\n return None\n else:\n return row[0]\n",
"role": "user"
},
{
"content": "getFileNumber",
"role": "assistant"
}
] |
def x(self, filename):
cur = self._connection.cursor()
try:
cur.execute('INSERT INTO files(name) VALUES (?)', (filename,))
except sqlite3.IntegrityError as e:
if str(e).find('UNIQUE constraint failed') > -1:
raise ValueError('Duplicate file name: %r' % filename)
else:
raise
else:
fileNumber = cur.lastrowid
self._connection.commit()
return fileNumber
| addFilename | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, filename):\n cur = self._connection.cursor()\n try:\n cur.execute('INSERT INTO files(name) VALUES (?)', (filename,))\n except sqlite3.IntegrityError as e:\n if str(e).find('UNIQUE constraint failed') > -1:\n raise ValueError('Duplicate file name: %r' % filename)\n else:\n raise\n else:\n fileNumber = cur.lastrowid\n self._connection.commit()\n return fileNumber\n",
"role": "user"
},
{
"content": "addFilename",
"role": "assistant"
}
] |
def x(self, filename):
endswith = filename.lower().endswith
if endswith('.bgz') or endswith('.gz'):
useBgzf = True
elif endswith('.bz2'):
raise ValueError(
'Compressed FASTA is only supported in BGZF format. Use bgzip to compresss your FASTA.'
)
else:
useBgzf = False
fileNumber = self._addFilename(filename)
connection = self._connection
count = 0
try:
with connection:
if useBgzf:
try:
fp = bgzf.open(filename, 'rb')
except ValueError as e:
if str(e).find('BGZF') > -1:
raise ValueError(
'Compressed FASTA is only supported in BGZF format. Use the samtools bgzip utility (instead of gzip) to compresss your FASTA.'
)
else:
raise
else:
try:
for line in fp:
if line[0] == '>':
count += 1
id_ = line[1:].rstrip(' \t\n\r')
connection.execute(
'INSERT INTO sequences(id, fileNumber, offset) VALUES (?, ?, ?)'
, (id_, fileNumber, fp.tell()))
finally:
fp.close()
else:
with open(filename) as fp:
offset = 0
for line in fp:
offset += len(line)
if line[0] == '>':
count += 1
id_ = line[1:].rstrip(' \t\n\r')
connection.execute(
'INSERT INTO sequences(id, fileNumber, offset) VALUES (?, ?, ?)'
, (id_, fileNumber, offset))
except sqlite3.IntegrityError as e:
if str(e).find('UNIQUE constraint failed') > -1:
original = self._find(id_)
if original is None:
raise ValueError(
"FASTA sequence id '%s' found twice in file '%s'." % (
id_, filename))
else:
origFilename, _ = original
raise ValueError(
"FASTA sequence id '%s', found in file '%s', was previously added from file '%s'."
% (id_, filename, origFilename))
else:
raise
else:
return count
| addFile | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, filename):\n endswith = filename.lower().endswith\n if endswith('.bgz') or endswith('.gz'):\n useBgzf = True\n elif endswith('.bz2'):\n raise ValueError(\n 'Compressed FASTA is only supported in BGZF format. Use bgzip to compresss your FASTA.'\n )\n else:\n useBgzf = False\n fileNumber = self._addFilename(filename)\n connection = self._connection\n count = 0\n try:\n with connection:\n if useBgzf:\n try:\n fp = bgzf.open(filename, 'rb')\n except ValueError as e:\n if str(e).find('BGZF') > -1:\n raise ValueError(\n 'Compressed FASTA is only supported in BGZF format. Use the samtools bgzip utility (instead of gzip) to compresss your FASTA.'\n )\n else:\n raise\n else:\n try:\n for line in fp:\n if line[0] == '>':\n count += 1\n id_ = line[1:].rstrip(' \\t\\n\\r')\n connection.execute(\n 'INSERT INTO sequences(id, fileNumber, offset) VALUES (?, ?, ?)'\n , (id_, fileNumber, fp.tell()))\n finally:\n fp.close()\n else:\n with open(filename) as fp:\n offset = 0\n for line in fp:\n offset += len(line)\n if line[0] == '>':\n count += 1\n id_ = line[1:].rstrip(' \\t\\n\\r')\n connection.execute(\n 'INSERT INTO sequences(id, fileNumber, offset) VALUES (?, ?, ?)'\n , (id_, fileNumber, offset))\n except sqlite3.IntegrityError as e:\n if str(e).find('UNIQUE constraint failed') > -1:\n original = self._find(id_)\n if original is None:\n raise ValueError(\n \"FASTA sequence id '%s' found twice in file '%s'.\" % (\n id_, filename))\n else:\n origFilename, _ = original\n raise ValueError(\n \"FASTA sequence id '%s', found in file '%s', was previously added from file '%s'.\"\n % (id_, filename, origFilename))\n else:\n raise\n else:\n return count\n",
"role": "user"
},
{
"content": "addFile",
"role": "assistant"
}
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.