source_code
stringlengths
45
7.24k
function_name
stringlengths
1
111
messages
list
def x(connection, without_connection_test): database.set_connection(connection=connection) if not without_connection_test: test_connection(connection)
connection
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(connection, without_connection_test):\n database.set_connection(connection=connection)\n if not without_connection_test:\n test_connection(connection)\n", "role": "user" }, { "content": "connection", "role": "assistant" } ]
def x(initialParams, laterParams): err = [] for param in initialParams: if param in laterParams: if param not in VARIABLE_PARAMS and initialParams[param ] != laterParams[param]: err.append( '\tParam %r initial value %r differs from later value %r' % (param, initialParams[param], laterParams[param])) else: err.append( '\t%r found in initial parameters, not found in later parameters' % param) for param in laterParams: if param not in initialParams: err.append( '\t%r found in later parameters, not seen in initial parameters' % param) return 'Summary of differences:\n%s' % '\n'.join(err) if err else None
checkCompatibleParams
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(initialParams, laterParams):\n err = []\n for param in initialParams:\n if param in laterParams:\n if param not in VARIABLE_PARAMS and initialParams[param\n ] != laterParams[param]:\n err.append(\n '\\tParam %r initial value %r differs from later value %r' %\n (param, initialParams[param], laterParams[param]))\n else:\n err.append(\n '\\t%r found in initial parameters, not found in later parameters'\n % param)\n for param in laterParams:\n if param not in initialParams:\n err.append(\n '\\t%r found in later parameters, not seen in initial parameters'\n % param)\n return 'Summary of differences:\\n%s' % '\\n'.join(err) if err else None\n", "role": "user" }, { "content": "checkCompatibleParams", "role": "assistant" } ]
def x(self): src = self._source or '' coll = self._collection or '' desc = self._description or '' l = [self._id, src, coll, self._name, ','.join(sorted(self._genes)), desc] return l
to_list
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n src = self._source or ''\n coll = self._collection or ''\n desc = self._description or ''\n l = [self._id, src, coll, self._name, ','.join(sorted(self._genes)), desc]\n return l\n", "role": "user" }, { "content": "to_list", "role": "assistant" } ]
def x(cls, l): id_ = l[0] name = l[3] genes = l[4].split(',') src = l[1] or None coll = l[2] or None desc = l[5] or None return cls(id_, name, genes, src, coll, desc)
from_list
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, l):\n id_ = l[0]\n name = l[3]\n genes = l[4].split(',')\n src = l[1] or None\n coll = l[2] or None\n desc = l[5] or None\n return cls(id_, name, genes, src, coll, desc)\n", "role": "user" }, { "content": "from_list", "role": "assistant" } ]
def x(primer, seq): offsets = [] seq = seq.upper() primer = primer.upper() primerLen = len(primer) discarded = 0 offset = seq.find(primer) while offset > -1: offsets.append(discarded + offset) seq = seq[offset + primerLen:] discarded += offset + primerLen offset = seq.find(primer) return offsets
findPrimer
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(primer, seq):\n offsets = []\n seq = seq.upper()\n primer = primer.upper()\n primerLen = len(primer)\n discarded = 0\n offset = seq.find(primer)\n while offset > -1:\n offsets.append(discarded + offset)\n seq = seq[offset + primerLen:]\n discarded += offset + primerLen\n offset = seq.find(primer)\n return offsets\n", "role": "user" }, { "content": "findPrimer", "role": "assistant" } ]
def x(primer, seq): forward = findPrimer(primer, seq) reverse = findPrimer(reverse_complement(primer), seq) return forward, reverse
findPrimerBidi
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(primer, seq):\n forward = findPrimer(primer, seq)\n reverse = findPrimer(reverse_complement(primer), seq)\n return forward, reverse\n", "role": "user" }, { "content": "findPrimerBidi", "role": "assistant" } ]
def x(primer, seq): forward, reverse = findPrimerBidi(primer, seq) if forward: start = forward[-1] + len(primer) end = len(seq) for offset in reverse: if offset >= start: end = offset break else: start = 0 end = reverse[0] if reverse else len(seq) return start, end
findPrimerBidiLimits
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(primer, seq):\n forward, reverse = findPrimerBidi(primer, seq)\n if forward:\n start = forward[-1] + len(primer)\n end = len(seq)\n for offset in reverse:\n if offset >= start:\n end = offset\n break\n else:\n start = 0\n end = reverse[0] if reverse else len(seq)\n return start, end\n", "role": "user" }, { "content": "findPrimerBidiLimits", "role": "assistant" } ]
def x(names): def x(name): count = 0 for ch in name: if ch in string.digits: count += 1 else: break return 0 if count == 0 else int(name[0:count]) return sorted(names, key=lambda name: numericPrefix(basename(name)))
numericallySortFilenames
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(names):\n\n def x(name):\n count = 0\n for ch in name:\n if ch in string.digits:\n count += 1\n else:\n break\n return 0 if count == 0 else int(name[0:count])\n return sorted(names, key=lambda name: numericPrefix(basename(name)))\n", "role": "user" }, { "content": "numericallySortFilenames", "role": "assistant" } ]
def x(fileNameOrHandle, mode='r'): if isinstance(fileNameOrHandle, six.string_types): if fileNameOrHandle.endswith('.gz'): if six.PY3: yield gzip.open(fileNameOrHandle, mode='rt', encoding='UTF-8') else: yield gzip.GzipFile(fileNameOrHandle) elif fileNameOrHandle.endswith('.bz2'): if six.PY3: yield bz2.open(fileNameOrHandle, mode='rt', encoding='UTF-8') else: yield bz2.BZ2File(fileNameOrHandle) else: with open(fileNameOrHandle) as fp: yield fp else: yield fileNameOrHandle
asHandle
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(fileNameOrHandle, mode='r'):\n if isinstance(fileNameOrHandle, six.string_types):\n if fileNameOrHandle.endswith('.gz'):\n if six.PY3:\n yield gzip.open(fileNameOrHandle, mode='rt', encoding='UTF-8')\n else:\n yield gzip.GzipFile(fileNameOrHandle)\n elif fileNameOrHandle.endswith('.bz2'):\n if six.PY3:\n yield bz2.open(fileNameOrHandle, mode='rt', encoding='UTF-8')\n else:\n yield bz2.BZ2File(fileNameOrHandle)\n else:\n with open(fileNameOrHandle) as fp:\n yield fp\n else:\n yield fileNameOrHandle\n", "role": "user" }, { "content": "asHandle", "role": "assistant" } ]
def x(s, convertToZeroBased=False): result = set() for _range in s.split(','): match = _rangeRegex.match(_range) if match: start, end = match.groups() start = int(start) if end is None: end = start else: end = int(end) if start > end: start, end = end, start if convertToZeroBased: result.update(range(start - 1, end)) else: result.update(range(start, end + 1)) else: raise ValueError( 'Illegal range %r. Ranges must single numbers or number-number.' % _range) return result
parseRangeString
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(s, convertToZeroBased=False):\n result = set()\n for _range in s.split(','):\n match = _rangeRegex.match(_range)\n if match:\n start, end = match.groups()\n start = int(start)\n if end is None:\n end = start\n else:\n end = int(end)\n if start > end:\n start, end = end, start\n if convertToZeroBased:\n result.update(range(start - 1, end))\n else:\n result.update(range(start, end + 1))\n else:\n raise ValueError(\n 'Illegal range %r. Ranges must single numbers or number-number.'\n % _range)\n return result\n", "role": "user" }, { "content": "parseRangeString", "role": "assistant" } ]
def x(nucleotides, prefix=''): result = [] for offset in sorted(nucleotides): result.append('%s%d: %s' % (prefix, offset, baseCountsToStr( nucleotides[offset]))) return '\n'.join(result)
nucleotidesToStr
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(nucleotides, prefix=''):\n result = []\n for offset in sorted(nucleotides):\n result.append('%s%d: %s' % (prefix, offset, baseCountsToStr(\n nucleotides[offset])))\n return '\\n'.join(result)\n", "role": "user" }, { "content": "nucleotidesToStr", "role": "assistant" } ]
def x(mesg, count, len1, len2=None): def x(a, b): return 100.0 * a / b if b else 0.0 if count == 0: return '%s: %d' % (mesg, count) else: len2 = len2 or len1 if len1 == len2: return '%s: %d/%d (%.2f%%)' % (mesg, count, len1, percentage( count, len1)) else: return ( '%s: %d/%d (%.2f%%) of sequence 1, %d/%d (%.2f%%) of sequence 2' % (mesg, count, len1, percentage(count, len1), count, len2, percentage(count, len2)))
countPrint
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(mesg, count, len1, len2=None):\n\n def x(a, b):\n return 100.0 * a / b if b else 0.0\n if count == 0:\n return '%s: %d' % (mesg, count)\n else:\n len2 = len2 or len1\n if len1 == len2:\n return '%s: %d/%d (%.2f%%)' % (mesg, count, len1, percentage(\n count, len1))\n else:\n return (\n '%s: %d/%d (%.2f%%) of sequence 1, %d/%d (%.2f%%) of sequence 2'\n % (mesg, count, len1, percentage(count, len1), count, len2,\n percentage(count, len2)))\n", "role": "user" }, { "content": "countPrint", "role": "assistant" } ]
def x(client, script_dir, overwrite=True): local_dir = os.path.join(genometools._root, 'data', 'gcloud', 'scripts') match = _BUCKET_PAT.match(script_dir) script_bucket = match.group(1) script_prefix = match.group(2) depth = len(local_dir.split(os.sep)) for root, dirs, files in os.walk(local_dir): rel_path = '/'.join(root.split(os.sep)[depth:]) for f in files: local_path = os.path.join(root, f) if rel_path: remote_path = '/'.join([script_prefix, rel_path, f]) else: remote_path = '/'.join([script_prefix, f]) _LOGGER.info('Uploading "%s"...', remote_path) storage.upload_file(client, script_bucket, local_path, remote_path, overwrite=overwrite)
upload_scripts
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(client, script_dir, overwrite=True):\n local_dir = os.path.join(genometools._root, 'data', 'gcloud', 'scripts')\n match = _BUCKET_PAT.match(script_dir)\n script_bucket = match.group(1)\n script_prefix = match.group(2)\n depth = len(local_dir.split(os.sep))\n for root, dirs, files in os.walk(local_dir):\n rel_path = '/'.join(root.split(os.sep)[depth:])\n for f in files:\n local_path = os.path.join(root, f)\n if rel_path:\n remote_path = '/'.join([script_prefix, rel_path, f])\n else:\n remote_path = '/'.join([script_prefix, f])\n _LOGGER.info('Uploading \"%s\"...', remote_path)\n storage.upload_file(client, script_bucket, local_path,\n remote_path, overwrite=overwrite)\n", "role": "user" }, { "content": "upload_scripts", "role": "assistant" } ]
def x(self, filename, scoreClass): if filename.endswith('.json') or filename.endswith('.json.bz2'): return JSONRecordsReader(filename, scoreClass) else: raise ValueError('Unknown DIAMOND record file suffix for file %r.' % filename)
getReader
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, filename, scoreClass):\n if filename.endswith('.json') or filename.endswith('.json.bz2'):\n return JSONRecordsReader(filename, scoreClass)\n else:\n raise ValueError('Unknown DIAMOND record file suffix for file %r.' %\n filename)\n", "role": "user" }, { "content": "getReader", "role": "assistant" } ]
def x(self): reads = iter(self.reads) first = True for filename in self.filenames: if first: first = False reader = self._reader else: reader = self._getReader(filename, self.scoreClass) for readAlignments in reader.readAlignments(reads): yield readAlignments for read in reads: yield ReadAlignments(read, [])
iter
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n reads = iter(self.reads)\n first = True\n for filename in self.filenames:\n if first:\n first = False\n reader = self._reader\n else:\n reader = self._getReader(filename, self.scoreClass)\n for readAlignments in reader.readAlignments(reads):\n yield readAlignments\n for read in reads:\n yield ReadAlignments(read, [])\n", "role": "user" }, { "content": "iter", "role": "assistant" } ]
def x(self, title): if self._subjectTitleToSubject is None: if self._databaseFilename is None: self._subjectTitleToSubject = SqliteIndex(self. _sqliteDatabaseFilename, fastaDirectory=self. _databaseDirectory, readClass=AAReadWithX) else: titles = {} for read in FastaReads(self._databaseFilename, readClass= AAReadWithX): titles[read.id] = read self._subjectTitleToSubject = titles return self._subjectTitleToSubject[title]
getSubjectSequence
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, title):\n if self._subjectTitleToSubject is None:\n if self._databaseFilename is None:\n self._subjectTitleToSubject = SqliteIndex(self.\n _sqliteDatabaseFilename, fastaDirectory=self.\n _databaseDirectory, readClass=AAReadWithX)\n else:\n titles = {}\n for read in FastaReads(self._databaseFilename, readClass=\n AAReadWithX):\n titles[read.id] = read\n self._subjectTitleToSubject = titles\n return self._subjectTitleToSubject[title]\n", "role": "user" }, { "content": "getSubjectSequence", "role": "assistant" } ]
def x(project): urls = [] with ftputil.FTPHost(sra_host, sra_user, sra_password) as ftp_host: download_paths = [] exp_dir = '/sra/sra-instant/reads/ByStudy/sra/SRP/%s/%s/' % (project [:6], project) ftp_host.chdir(exp_dir) run_folders = ftp_host.listdir(ftp_host.curdir) for folder in run_folders: files = ftp_host.listdir(folder) assert len(files) == 1 for f in files: path = exp_dir + folder + '/' + f urls.append(path) return urls
get_project_urls
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(project):\n urls = []\n with ftputil.FTPHost(sra_host, sra_user, sra_password) as ftp_host:\n download_paths = []\n exp_dir = '/sra/sra-instant/reads/ByStudy/sra/SRP/%s/%s/' % (project\n [:6], project)\n ftp_host.chdir(exp_dir)\n run_folders = ftp_host.listdir(ftp_host.curdir)\n for folder in run_folders:\n files = ftp_host.listdir(folder)\n assert len(files) == 1\n for f in files:\n path = exp_dir + folder + '/' + f\n urls.append(path)\n return urls\n", "role": "user" }, { "content": "get_project_urls", "role": "assistant" } ]
def x(rst, clk, rx_rdy, rx_vld, rx_dat, tx_rdy, tx_vld, tx_dat): DATA_WIDTH = len(rx_dat) NUM_STAGES = 3 stage_en = Signal(intbv(0)[NUM_STAGES:]) pipe_ctrl = pipeline_control(rst=rst, clk=clk, rx_vld=rx_vld, rx_rdy= rx_rdy, tx_vld=tx_vld, tx_rdy=tx_rdy, stage_enable=stage_en) s0_dat = Signal(intbv(0)[DATA_WIDTH:]) @always_seq(clk.posedge, reset=rst) def x(): if stage_en[0]: s0_dat.next = rx_dat s1_dat = Signal(intbv(0)[DATA_WIDTH:]) @always_seq(clk.posedge, reset=rst) def x(): if stage_en[1]: s1_dat.next = ~s0_dat s2_dat = Signal(intbv(0)[DATA_WIDTH:]) @always_seq(clk.posedge, reset=rst) def x(): if stage_en[2]: s2_dat.next = s1_dat + 1 @always_comb def x(): tx_dat.next = s2_dat.signed() return instances()
twos_complement
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(rst, clk, rx_rdy, rx_vld, rx_dat, tx_rdy, tx_vld, tx_dat):\n DATA_WIDTH = len(rx_dat)\n NUM_STAGES = 3\n stage_en = Signal(intbv(0)[NUM_STAGES:])\n pipe_ctrl = pipeline_control(rst=rst, clk=clk, rx_vld=rx_vld, rx_rdy=\n rx_rdy, tx_vld=tx_vld, tx_rdy=tx_rdy, stage_enable=stage_en)\n s0_dat = Signal(intbv(0)[DATA_WIDTH:])\n\n @always_seq(clk.posedge, reset=rst)\n def x():\n if stage_en[0]:\n s0_dat.next = rx_dat\n s1_dat = Signal(intbv(0)[DATA_WIDTH:])\n\n @always_seq(clk.posedge, reset=rst)\n def x():\n if stage_en[1]:\n s1_dat.next = ~s0_dat\n s2_dat = Signal(intbv(0)[DATA_WIDTH:])\n\n @always_seq(clk.posedge, reset=rst)\n def x():\n if stage_en[2]:\n s2_dat.next = s1_dat + 1\n\n @always_comb\n def x():\n tx_dat.next = s2_dat.signed()\n return instances()\n", "role": "user" }, { "content": "twos_complement", "role": "assistant" } ]
def x(): desc = 'Generate a mapping of Entrez IDs to gene symbols.' parser = cli.get_argument_parser(desc=desc) parser.add_argument('-f', '--gene2acc-file', type=str, required=True, help=textwrap.dedent( """ Path of gene2accession.gz file (from ftp://ftp.ncbi.nlm.nih.gov/gene/DATA), or a filtered version thereof.""" )) parser.add_argument('-o', '--output-file', type=str, required=True, help=textwrap.dedent( """ Path of output file. If set to ``-``, print to ``stdout``, and redirect logging messages to ``stderr``.""" )) parser.add_argument('-l', '--log-file', type=str, default=None, help= 'Path of log file. If not specified, print to stdout.') parser.add_argument('-q', '--quiet', action='store_true', help= 'Suppress all output except warnings and errors.') parser.add_argument('-v', '--verbose', action='store_true', help= 'Enable verbose output. Ignored if ``--quiet`` is specified.') return parser
get_argument_parser
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n desc = 'Generate a mapping of Entrez IDs to gene symbols.'\n parser = cli.get_argument_parser(desc=desc)\n parser.add_argument('-f', '--gene2acc-file', type=str, required=True,\n help=textwrap.dedent(\n \"\"\" Path of gene2accession.gz file (from\n ftp://ftp.ncbi.nlm.nih.gov/gene/DATA), or a filtered version\n thereof.\"\"\"\n ))\n parser.add_argument('-o', '--output-file', type=str, required=True,\n help=textwrap.dedent(\n \"\"\" Path of output file. If set to ``-``, print to ``stdout``,\n and redirect logging messages to ``stderr``.\"\"\"\n ))\n parser.add_argument('-l', '--log-file', type=str, default=None, help=\n 'Path of log file. If not specified, print to stdout.')\n parser.add_argument('-q', '--quiet', action='store_true', help=\n 'Suppress all output except warnings and errors.')\n parser.add_argument('-v', '--verbose', action='store_true', help=\n 'Enable verbose output. Ignored if ``--quiet`` is specified.')\n return parser\n", "role": "user" }, { "content": "get_argument_parser", "role": "assistant" } ]
def x(file_path, logger): gene2acc = {} with misc.smart_open_read(file_path, mode='rb', try_gzip=True) as fh: reader = csv.reader(fh, dialect='excel-tab') next(reader) for i, l in enumerate(reader): id_ = int(l[1]) symbol = l[15] try: gene2acc[id_].append(symbol) except KeyError: gene2acc[id_] = [symbol] n = len(gene2acc) for k, v in gene2acc.items(): symbols = sorted(set(v)) assert len(symbols) == 1 gene2acc[k] = symbols[0] all_symbols = sorted(set(gene2acc.values())) m = len(all_symbols) logger.info('Found %d Entrez Gene IDs associated with %d gene symbols.', n, m) return gene2acc
read_gene2acc
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(file_path, logger):\n gene2acc = {}\n with misc.smart_open_read(file_path, mode='rb', try_gzip=True) as fh:\n reader = csv.reader(fh, dialect='excel-tab')\n next(reader)\n for i, l in enumerate(reader):\n id_ = int(l[1])\n symbol = l[15]\n try:\n gene2acc[id_].append(symbol)\n except KeyError:\n gene2acc[id_] = [symbol]\n n = len(gene2acc)\n for k, v in gene2acc.items():\n symbols = sorted(set(v))\n assert len(symbols) == 1\n gene2acc[k] = symbols[0]\n all_symbols = sorted(set(gene2acc.values()))\n m = len(all_symbols)\n logger.info('Found %d Entrez Gene IDs associated with %d gene symbols.',\n n, m)\n return gene2acc\n", "role": "user" }, { "content": "read_gene2acc", "role": "assistant" } ]
def x(file_path, entrez2gene, logger): with misc.smart_open_write(file_path, mode='wb') as ofh: writer = csv.writer(ofh, dialect='excel-tab', lineterminator=os.linesep ) for k in sorted(entrez2gene.keys(), key=lambda x: int(x)): writer.writerow([k, entrez2gene[k]]) logger.info('Output written to file "%s".', file_path)
write_entrez2gene
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(file_path, entrez2gene, logger):\n with misc.smart_open_write(file_path, mode='wb') as ofh:\n writer = csv.writer(ofh, dialect='excel-tab', lineterminator=os.linesep\n )\n for k in sorted(entrez2gene.keys(), key=lambda x: int(x)):\n writer.writerow([k, entrez2gene[k]])\n logger.info('Output written to file \"%s\".', file_path)\n", "role": "user" }, { "content": "write_entrez2gene", "role": "assistant" } ]
def x(args=None): vinfo = sys.version_info if not vinfo >= (2, 7): raise SystemError( 'Python interpreter version >= 2.7 required, found %d.%d instead.' % (vinfo.major, vinfo.minor)) if args is None: parser = get_argument_parser() args = parser.parse_args() gene2acc_file = args.gene2acc_file output_file = args.output_file log_file = args.log_file quiet = args.quiet verbose = args.verbose log_stream = sys.stdout if output_file == '-': log_stream = sys.stderr logger = misc.get_logger(log_stream=log_stream, log_file=log_file, quiet=quiet, verbose=verbose) entrez2gene = read_gene2acc(gene2acc_file, logger) write_entrez2gene(output_file, entrez2gene, logger) return 0
main
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(args=None):\n vinfo = sys.version_info\n if not vinfo >= (2, 7):\n raise SystemError(\n 'Python interpreter version >= 2.7 required, found %d.%d instead.'\n % (vinfo.major, vinfo.minor))\n if args is None:\n parser = get_argument_parser()\n args = parser.parse_args()\n gene2acc_file = args.gene2acc_file\n output_file = args.output_file\n log_file = args.log_file\n quiet = args.quiet\n verbose = args.verbose\n log_stream = sys.stdout\n if output_file == '-':\n log_stream = sys.stderr\n logger = misc.get_logger(log_stream=log_stream, log_file=log_file,\n quiet=quiet, verbose=verbose)\n entrez2gene = read_gene2acc(gene2acc_file, logger)\n write_entrez2gene(output_file, entrez2gene, logger)\n return 0\n", "role": "user" }, { "content": "main", "role": "assistant" } ]
def x(s): abbrev1 = None origS = s if ' ' in s: first, rest = s.split(' ', 1) s = first.title() + ' ' + rest.lower() else: s = s.title() if s in NAMES: abbrev1 = s elif s in ABBREV3_TO_ABBREV1: abbrev1 = ABBREV3_TO_ABBREV1[s] elif s in NAMES_TO_ABBREV1: abbrev1 = NAMES_TO_ABBREV1[s] else: def x(target): for abbrev1, codons in CODONS.items(): for codon in codons: if codon == target: return abbrev1 abbrev1 = findCodon(origS.upper()) if abbrev1: return AminoAcid(NAMES[abbrev1], ABBREV3[abbrev1], abbrev1, CODONS[ abbrev1], PROPERTIES[abbrev1], PROPERTY_DETAILS[abbrev1], PROPERTY_CLUSTERS[abbrev1])
find
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(s):\n abbrev1 = None\n origS = s\n if ' ' in s:\n first, rest = s.split(' ', 1)\n s = first.title() + ' ' + rest.lower()\n else:\n s = s.title()\n if s in NAMES:\n abbrev1 = s\n elif s in ABBREV3_TO_ABBREV1:\n abbrev1 = ABBREV3_TO_ABBREV1[s]\n elif s in NAMES_TO_ABBREV1:\n abbrev1 = NAMES_TO_ABBREV1[s]\n else:\n\n def x(target):\n for abbrev1, codons in CODONS.items():\n for codon in codons:\n if codon == target:\n return abbrev1\n abbrev1 = findCodon(origS.upper())\n if abbrev1:\n return AminoAcid(NAMES[abbrev1], ABBREV3[abbrev1], abbrev1, CODONS[\n abbrev1], PROPERTIES[abbrev1], PROPERTY_DETAILS[abbrev1],\n PROPERTY_CLUSTERS[abbrev1])\n", "role": "user" }, { "content": "find", "role": "assistant" } ]
def x(sequence, propertyNames, propertyValues, missingAAValue): propertyNames = sorted(map(str.lower, set(propertyNames))) knownProperties = set() for names in propertyValues.values(): knownProperties.update(names) unknown = set(propertyNames) - knownProperties if unknown: raise ValueError('Unknown propert%s: %s.' % ('y' if len(unknown) == 1 else 'ies', ', '.join(unknown))) aas = sequence.sequence.upper() result = {} for propertyName in propertyNames: result[propertyName] = values = [] append = values.append for aa in aas: try: properties = propertyValues[aa] except KeyError: append(missingAAValue) else: append(properties[propertyName]) return result
propertiesOrClustersForSequence
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(sequence, propertyNames, propertyValues, missingAAValue):\n propertyNames = sorted(map(str.lower, set(propertyNames)))\n knownProperties = set()\n for names in propertyValues.values():\n knownProperties.update(names)\n unknown = set(propertyNames) - knownProperties\n if unknown:\n raise ValueError('Unknown propert%s: %s.' % ('y' if len(unknown) ==\n 1 else 'ies', ', '.join(unknown)))\n aas = sequence.sequence.upper()\n result = {}\n for propertyName in propertyNames:\n result[propertyName] = values = []\n append = values.append\n for aa in aas:\n try:\n properties = propertyValues[aa]\n except KeyError:\n append(missingAAValue)\n else:\n append(properties[propertyName])\n return result\n", "role": "user" }, { "content": "propertiesOrClustersForSequence", "role": "assistant" } ]
def x(aaMatch, read1, read2, indent='', offsets=None): match = aaMatch['match'] matchCount = match['matchCount'] gapMismatchCount = match['gapMismatchCount'] gapGapMismatchCount = match['gapGapMismatchCount'] nonGapMismatchCount = match['nonGapMismatchCount'] if offsets: len1 = len2 = len(offsets) else: len1, len2 = map(len, (read1, read2)) result = [] append = result.append append(countPrint('%sMatches' % indent, matchCount, len1, len2)) mismatchCount = (gapMismatchCount + gapGapMismatchCount + nonGapMismatchCount) append(countPrint('%sMismatches' % indent, mismatchCount, len1, len2)) append(countPrint('%s Not involving gaps (i.e., conflicts)' % indent, nonGapMismatchCount, len1, len2)) append(countPrint('%s Involving a gap in one sequence' % indent, gapMismatchCount, len1, len2)) append(countPrint('%s Involving a gap in both sequences' % indent, gapGapMismatchCount, len1, len2)) for read, key in zip((read1, read2), ('read1', 'read2')): append('%s Id: %s' % (indent, read.id)) length = len(read) append('%s Length: %d' % (indent, length)) gapCount = len(aaMatch[key]['gapOffsets']) append(countPrint('%s Gaps' % indent, gapCount, length)) if gapCount: append('%s Gap locations (1-based): %s' % (indent, ', '.join (map(lambda offset: str(offset + 1), sorted(aaMatch[key][ 'gapOffsets']))))) extraCount = aaMatch[key]['extraCount'] if extraCount: append(countPrint('%s Extra nucleotides at end' % indent, extraCount, length)) return '\n'.join(result)
matchToString
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(aaMatch, read1, read2, indent='', offsets=None):\n match = aaMatch['match']\n matchCount = match['matchCount']\n gapMismatchCount = match['gapMismatchCount']\n gapGapMismatchCount = match['gapGapMismatchCount']\n nonGapMismatchCount = match['nonGapMismatchCount']\n if offsets:\n len1 = len2 = len(offsets)\n else:\n len1, len2 = map(len, (read1, read2))\n result = []\n append = result.append\n append(countPrint('%sMatches' % indent, matchCount, len1, len2))\n mismatchCount = (gapMismatchCount + gapGapMismatchCount +\n nonGapMismatchCount)\n append(countPrint('%sMismatches' % indent, mismatchCount, len1, len2))\n append(countPrint('%s Not involving gaps (i.e., conflicts)' % indent,\n nonGapMismatchCount, len1, len2))\n append(countPrint('%s Involving a gap in one sequence' % indent,\n gapMismatchCount, len1, len2))\n append(countPrint('%s Involving a gap in both sequences' % indent,\n gapGapMismatchCount, len1, len2))\n for read, key in zip((read1, read2), ('read1', 'read2')):\n append('%s Id: %s' % (indent, read.id))\n length = len(read)\n append('%s Length: %d' % (indent, length))\n gapCount = len(aaMatch[key]['gapOffsets'])\n append(countPrint('%s Gaps' % indent, gapCount, length))\n if gapCount:\n append('%s Gap locations (1-based): %s' % (indent, ', '.join\n (map(lambda offset: str(offset + 1), sorted(aaMatch[key][\n 'gapOffsets'])))))\n extraCount = aaMatch[key]['extraCount']\n if extraCount:\n append(countPrint('%s Extra nucleotides at end' % indent,\n extraCount, length))\n return '\\n'.join(result)\n", "role": "user" }, { "content": "matchToString", "role": "assistant" } ]
def x(read1, read2, gapChars='-', offsets=None): matchCount = 0 gapMismatchCount = nonGapMismatchCount = gapGapMismatchCount = 0 read1ExtraCount = read2ExtraCount = 0 read1GapOffsets = [] read2GapOffsets = [] for offset, (a, b) in enumerate(zip_longest(read1.sequence.upper(), read2.sequence.upper())): if offsets is not None and offset not in offsets: continue if a is None: assert b is not None read2ExtraCount += 1 if b in gapChars: read2GapOffsets.append(offset) elif b is None: read1ExtraCount += 1 if a in gapChars: read1GapOffsets.append(offset) elif a in gapChars: read1GapOffsets.append(offset) if b in gapChars: gapGapMismatchCount += 1 read2GapOffsets.append(offset) else: gapMismatchCount += 1 elif b in gapChars: gapMismatchCount += 1 read2GapOffsets.append(offset) elif a == b: matchCount += 1 else: nonGapMismatchCount += 1 return {'match': {'matchCount': matchCount, 'gapMismatchCount': gapMismatchCount, 'gapGapMismatchCount': gapGapMismatchCount, 'nonGapMismatchCount': nonGapMismatchCount}, 'read1': {'extraCount': read1ExtraCount, 'gapOffsets': read1GapOffsets}, 'read2': { 'extraCount': read2ExtraCount, 'gapOffsets': read2GapOffsets}}
compareAaReads
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(read1, read2, gapChars='-', offsets=None):\n matchCount = 0\n gapMismatchCount = nonGapMismatchCount = gapGapMismatchCount = 0\n read1ExtraCount = read2ExtraCount = 0\n read1GapOffsets = []\n read2GapOffsets = []\n for offset, (a, b) in enumerate(zip_longest(read1.sequence.upper(),\n read2.sequence.upper())):\n if offsets is not None and offset not in offsets:\n continue\n if a is None:\n assert b is not None\n read2ExtraCount += 1\n if b in gapChars:\n read2GapOffsets.append(offset)\n elif b is None:\n read1ExtraCount += 1\n if a in gapChars:\n read1GapOffsets.append(offset)\n elif a in gapChars:\n read1GapOffsets.append(offset)\n if b in gapChars:\n gapGapMismatchCount += 1\n read2GapOffsets.append(offset)\n else:\n gapMismatchCount += 1\n elif b in gapChars:\n gapMismatchCount += 1\n read2GapOffsets.append(offset)\n elif a == b:\n matchCount += 1\n else:\n nonGapMismatchCount += 1\n return {'match': {'matchCount': matchCount, 'gapMismatchCount':\n gapMismatchCount, 'gapGapMismatchCount': gapGapMismatchCount,\n 'nonGapMismatchCount': nonGapMismatchCount}, 'read1': {'extraCount':\n read1ExtraCount, 'gapOffsets': read1GapOffsets}, 'read2': {\n 'extraCount': read2ExtraCount, 'gapOffsets': read2GapOffsets}}\n", "role": "user" }, { "content": "compareAaReads", "role": "assistant" } ]
def x(colors, args): result = defaultdict(set) for colorInfo in colors: readIds = colorInfo.split() color = readIds.pop(0) for readId in readIds: if os.path.isfile(readId): filename = readId if args.fasta: reads = FastaReads(filename) else: reads = FastqReads(filename) for read in reads: result[color].add(read.id) else: result[color].add(readId) return result
parseColors
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(colors, args):\n result = defaultdict(set)\n for colorInfo in colors:\n readIds = colorInfo.split()\n color = readIds.pop(0)\n for readId in readIds:\n if os.path.isfile(readId):\n filename = readId\n if args.fasta:\n reads = FastaReads(filename)\n else:\n reads = FastqReads(filename)\n for read in reads:\n result[color].add(read.id)\n else:\n result[color].add(readId)\n return result\n", "role": "user" }, { "content": "parseColors", "role": "assistant" } ]
def x(download_file, release=None): if release is None: release = get_latest_release() url = ( 'http://viewvc.geneontology.org/viewvc/GO-SVN/ontology-releases/%s/go-basic.obo' % release) misc.http_download(url, download_file)
download_release
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(download_file, release=None):\n if release is None:\n release = get_latest_release()\n url = (\n 'http://viewvc.geneontology.org/viewvc/GO-SVN/ontology-releases/%s/go-basic.obo'\n % release)\n misc.http_download(url, download_file)\n", "role": "user" }, { "content": "download_release", "role": "assistant" } ]
def x(): with closing(requests.get( 'http://geneontology.org/ontology/go-basic.obo', stream=True)) as r: for i, l in enumerate(r.iter_lines(decode_unicode=True)): if i == 1: assert l.split(':')[0] == 'data-version' date = l.split('/')[-1] break return date
get_current_ontology_date
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n with closing(requests.get(\n 'http://geneontology.org/ontology/go-basic.obo', stream=True)) as r:\n for i, l in enumerate(r.iter_lines(decode_unicode=True)):\n if i == 1:\n assert l.split(':')[0] == 'data-version'\n date = l.split('/')[-1]\n break\n return date\n", "role": "user" }, { "content": "get_current_ontology_date", "role": "assistant" } ]
def x(function, name): try: return TaskResult(name, True, None, function()) except Exception as exc: return TaskResult(name, False, exc, None)
execute
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(function, name):\n try:\n return TaskResult(name, True, None, function())\n except Exception as exc:\n return TaskResult(name, False, exc, None)\n", "role": "user" }, { "content": "execute", "role": "assistant" } ]
def x(value, arg): try: length = int(arg) except ValueError: return value if len(value) > length: return value[:length] + '...' return value
truncatechars
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(value, arg):\n try:\n length = int(arg)\n except ValueError:\n return value\n if len(value) > length:\n return value[:length] + '...'\n return value\n", "role": "user" }, { "content": "truncatechars", "role": "assistant" } ]
def x(desc, default_field_name='gene'): parser = cli.get_argument_parser(desc=desc) parser.add_argument('-a', '--annotation-file', default='-', type=str, help=textwrap.dedent( """ Path of Ensembl gene annotation file (in GTF format). The file may be gzip'ed. If set to ``-``, read from ``stdin``.""" )) parser.add_argument('-o', '--output-file', required=True, type=str, help=textwrap.dedent( """ Path of output file. If set to ``-``, print to ``stdout``, and redirect logging messages to ``stderr``.""" )) parser.add_argument('-c', '--chromosome-pattern', type=str, required= False, default=None, help=textwrap.dedent( """ Regular expression that chromosome names have to match. [None] """ )) cli.add_reporting_args(parser) return parser
get_gtf_argument_parser
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(desc, default_field_name='gene'):\n parser = cli.get_argument_parser(desc=desc)\n parser.add_argument('-a', '--annotation-file', default='-', type=str,\n help=textwrap.dedent(\n \"\"\" Path of Ensembl gene annotation file (in GTF format). The file\n may be gzip'ed. If set to ``-``, read from ``stdin``.\"\"\"\n ))\n parser.add_argument('-o', '--output-file', required=True, type=str,\n help=textwrap.dedent(\n \"\"\" Path of output file. If set to ``-``, print to ``stdout``,\n and redirect logging messages to ``stderr``.\"\"\"\n ))\n parser.add_argument('-c', '--chromosome-pattern', type=str, required=\n False, default=None, help=textwrap.dedent(\n \"\"\" Regular expression that chromosome names have to match. [None]\n \"\"\"\n ))\n cli.add_reporting_args(parser)\n return parser\n", "role": "user" }, { "content": "get_gtf_argument_parser", "role": "assistant" } ]
def x(url_map): from flask import current_app from . import current_jsonschemas url_map.add(Rule('{0}/<path:path>'.format(current_app.config[ 'JSONSCHEMAS_ENDPOINT']), endpoint=current_jsonschemas.get_schema, host=current_app.config['JSONSCHEMAS_HOST']))
jsonresolver_loader
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(url_map):\n from flask import current_app\n from . import current_jsonschemas\n url_map.add(Rule('{0}/<path:path>'.format(current_app.config[\n 'JSONSCHEMAS_ENDPOINT']), endpoint=current_jsonschemas.get_schema,\n host=current_app.config['JSONSCHEMAS_HOST']))\n", "role": "user" }, { "content": "jsonresolver_loader", "role": "assistant" } ]
def x(layouts): layout = layouts[0].clone() for l in layouts[1:]: layout.files.update(l.files) layout.domains.update(l.domains) for k, v in l.entities.items(): if k not in layout.entities: layout.entities[k] = v else: layout.entities[k].files.update(v.files) return layout
merge_layouts
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(layouts):\n layout = layouts[0].clone()\n for l in layouts[1:]:\n layout.files.update(l.files)\n layout.domains.update(l.domains)\n for k, v in l.entities.items():\n if k not in layout.entities:\n layout.entities[k] = v\n else:\n layout.entities[k].files.update(v.files)\n return layout\n", "role": "user" }, { "content": "merge_layouts", "role": "assistant" } ]
def x(self, entities=None, extensions=None, domains=None, regex_search=False): if extensions is not None: if isinstance(extensions, six.string_types): extensions = [extensions] extensions = '(' + '|'.join(extensions) + ')$' if re.search(extensions, self.filename) is None: return False if domains is not None: domains = listify(domains) if not set(self.domains) & set(domains): return False if entities is not None: for name, val in entities.items(): if (name not in self.tags) ^ (val is None): return False if val is None: continue def x(x): patt = '%s' % x if isinstance(x, (int, float)): patt = '0*' + patt if not regex_search: patt = '^%s$' % patt return patt ent_patts = [make_patt(x) for x in listify(val)] patt = '|'.join(ent_patts) if re.search(patt, str(self.tags[name].value)) is None: return False return True
matches
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, entities=None, extensions=None, domains=None, regex_search=False):\n if extensions is not None:\n if isinstance(extensions, six.string_types):\n extensions = [extensions]\n extensions = '(' + '|'.join(extensions) + ')$'\n if re.search(extensions, self.filename) is None:\n return False\n if domains is not None:\n domains = listify(domains)\n if not set(self.domains) & set(domains):\n return False\n if entities is not None:\n for name, val in entities.items():\n if (name not in self.tags) ^ (val is None):\n return False\n if val is None:\n continue\n\n def x(x):\n patt = '%s' % x\n if isinstance(x, (int, float)):\n patt = '0*' + patt\n if not regex_search:\n patt = '^%s$' % patt\n return patt\n ent_patts = [make_patt(x) for x in listify(val)]\n patt = '|'.join(ent_patts)\n if re.search(patt, str(self.tags[name].value)) is None:\n return False\n return True\n", "role": "user" }, { "content": "matches", "role": "assistant" } ]
def x(self): keys = list(self.entities.keys()) replaced = [] for i, k in enumerate(keys): if iskeyword(k): replaced.append(k) keys[i] = '%s_' % k if replaced: safe = [('%s_' % k) for k in replaced] warnings.warn( 'Entity names cannot be reserved keywords when representing a File as a namedtuple. Replacing entities %s with safe versions %s.' % (keys, safe)) entities = dict(zip(keys, self.entities.values())) _File = namedtuple('File', 'filename ' + ' '.join(entities.keys())) return _File(filename=self.path, **entities)
as_named_tuple
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n keys = list(self.entities.keys())\n replaced = []\n for i, k in enumerate(keys):\n if iskeyword(k):\n replaced.append(k)\n keys[i] = '%s_' % k\n if replaced:\n safe = [('%s_' % k) for k in replaced]\n warnings.warn(\n 'Entity names cannot be reserved keywords when representing a File as a namedtuple. Replacing entities %s with safe versions %s.'\n % (keys, safe))\n entities = dict(zip(keys, self.entities.values()))\n _File = namedtuple('File', 'filename ' + ' '.join(entities.keys()))\n return _File(filename=self.path, **entities)\n", "role": "user" }, { "content": "as_named_tuple", "role": "assistant" } ]
def x(self, path_patterns, symbolic_link=False, root=None, conflicts='fail'): new_filename = build_path(self.entities, path_patterns) if not new_filename: return None if new_filename[-1] == os.sep: new_filename += self.filename if isabs(self.path) or root is None: path = self.path else: path = join(root, self.path) if not exists(path): raise ValueError( "Target filename to copy/symlink (%s) doesn't exist." % path) if symbolic_link: contents = None link_to = path else: with open(path, 'r') as f: contents = f.read() link_to = None write_contents_to_file(new_filename, contents=contents, link_to=link_to, content_mode='text', root=root, conflicts=conflicts)
copy
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, path_patterns, symbolic_link=False, root=None, conflicts='fail'):\n new_filename = build_path(self.entities, path_patterns)\n if not new_filename:\n return None\n if new_filename[-1] == os.sep:\n new_filename += self.filename\n if isabs(self.path) or root is None:\n path = self.path\n else:\n path = join(root, self.path)\n if not exists(path):\n raise ValueError(\n \"Target filename to copy/symlink (%s) doesn't exist.\" % path)\n if symbolic_link:\n contents = None\n link_to = path\n else:\n with open(path, 'r') as f:\n contents = f.read()\n link_to = None\n write_contents_to_file(new_filename, contents=contents, link_to=link_to,\n content_mode='text', root=root, conflicts=conflicts)\n", "role": "user" }, { "content": "copy", "role": "assistant" } ]
def x(self, f, update_file=False): if self.map_func is not None: val = self.map_func(f) else: m = self.regex.search(f.path) val = m.group(1) if m is not None else None return self._astype(val)
match_file
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, f, update_file=False):\n if self.map_func is not None:\n val = self.map_func(f)\n else:\n m = self.regex.search(f.path)\n val = m.group(1) if m is not None else None\n return self._astype(val)\n", "role": "user" }, { "content": "match_file", "role": "assistant" } ]
def x(self, domain): if isinstance(domain, six.string_types): if domain in self.domains: return self.domains[domain] elif exists(domain): with open(domain, 'r') as fobj: domain = json.load(fobj) else: raise ValueError( "No domain could be found/loaded from input '{}'; value must be either the name of an existing Domain, or a valid path to a configuration file." .format(domain)) name = domain['name'] if name in self.domains: msg = ( "Domain with name '{}' already exists; returning existing Domain configuration." .format(name)) warnings.warn(msg) return self.domains[name] entities = domain.get('entities', []) domain = Domain(domain) for e in entities: self.add_entity(domain=domain, **e) self.domains[name] = domain return self.domains[name]
get_or_load_domain
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, domain):\n if isinstance(domain, six.string_types):\n if domain in self.domains:\n return self.domains[domain]\n elif exists(domain):\n with open(domain, 'r') as fobj:\n domain = json.load(fobj)\n else:\n raise ValueError(\n \"No domain could be found/loaded from input '{}'; value must be either the name of an existing Domain, or a valid path to a configuration file.\"\n .format(domain))\n name = domain['name']\n if name in self.domains:\n msg = (\n \"Domain with name '{}' already exists; returning existing Domain configuration.\"\n .format(name))\n warnings.warn(msg)\n return self.domains[name]\n entities = domain.get('entities', [])\n domain = Domain(domain)\n for e in entities:\n self.add_entity(domain=domain, **e)\n self.domains[name] = domain\n return self.domains[name]\n", "role": "user" }, { "content": "get_or_load_domain", "role": "assistant" } ]
def x(self, f, domains=None): filename = f if isinstance(f, six.string_types) else f.path if domains is None: domains = list(self.domains.values()) domains = list(domains) domains.insert(0, self) for dom in domains: if dom.include: for regex in dom.include: if re.search(regex, filename): return True return False else: for regex in dom.exclude: if re.search(regex, filename, flags=re.UNICODE): return False return True
check_inclusions
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, f, domains=None):\n filename = f if isinstance(f, six.string_types) else f.path\n if domains is None:\n domains = list(self.domains.values())\n domains = list(domains)\n domains.insert(0, self)\n for dom in domains:\n if dom.include:\n for regex in dom.include:\n if re.search(regex, filename):\n return True\n return False\n else:\n for regex in dom.exclude:\n if re.search(regex, filename, flags=re.UNICODE):\n return False\n return True\n", "role": "user" }, { "content": "check_inclusions", "role": "assistant" } ]
def x(self, entity): if entity in self.entities: return self.entities[entity] _ent = [e for e in self.entities.values() if e.name == entity] if len(_ent) > 1: raise ValueError( "Entity name '%s' matches %d entities. To avoid ambiguity, please prefix the entity name with its domain (e.g., 'bids.%s'." % (entity, len(_ent), entity)) if _ent: return _ent[0] raise ValueError("No entity '%s' found." % entity)
find_entity
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, entity):\n if entity in self.entities:\n return self.entities[entity]\n _ent = [e for e in self.entities.values() if e.name == entity]\n if len(_ent) > 1:\n raise ValueError(\n \"Entity name '%s' matches %d entities. To avoid ambiguity, please prefix the entity name with its domain (e.g., 'bids.%s'.\"\n % (entity, len(_ent), entity))\n if _ent:\n return _ent[0]\n raise ValueError(\"No entity '%s' found.\" % entity)\n", "role": "user" }, { "content": "find_entity", "role": "assistant" } ]
def x(self, filename): data = {} for f in self.files.values(): entities = {v.entity.id: v.value for k, v in f.tags.items()} data[f.path] = {'domains': f.domains, 'entities': entities} with open(filename, 'w') as outfile: json.dump(data, outfile)
save_index
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, filename):\n data = {}\n for f in self.files.values():\n entities = {v.entity.id: v.value for k, v in f.tags.items()}\n data[f.path] = {'domains': f.domains, 'entities': entities}\n with open(filename, 'w') as outfile:\n json.dump(data, outfile)\n", "role": "user" }, { "content": "save_index", "role": "assistant" } ]
def x(self, filename, reindex=False): self._reset_index() with open(filename, 'r') as fobj: data = json.load(fobj) for path, file in data.items(): ents, domains = file['entities'], file['domains'] root, f = dirname(path), basename(path) if reindex: self._index_file(root, f, domains) else: f = self._make_file_object(root, f) tags = {k: Tag(self.entities[k], v) for k, v in ents.items()} f.tags = tags self.files[f.path] = f for ent, val in f.entities.items(): self.entities[ent].add_file(f.path, val)
load_index
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, filename, reindex=False):\n self._reset_index()\n with open(filename, 'r') as fobj:\n data = json.load(fobj)\n for path, file in data.items():\n ents, domains = file['entities'], file['domains']\n root, f = dirname(path), basename(path)\n if reindex:\n self._index_file(root, f, domains)\n else:\n f = self._make_file_object(root, f)\n tags = {k: Tag(self.entities[k], v) for k, v in ents.items()}\n f.tags = tags\n self.files[f.path] = f\n for ent, val in f.entities.items():\n self.entities[ent].add_file(f.path, val)\n", "role": "user" }, { "content": "load_index", "role": "assistant" } ]
def x(self, domain, **kwargs): map_func = kwargs.get('map_func', None) if map_func is not None and not callable(kwargs['map_func']): if self.entity_mapper is None: raise ValueError( "Mapping function '%s' specified for Entity '%s', but no entity mapper was passed when initializing the current Layout. Please make sure the 'entity_mapper' argument is set." % (map_func, kwargs['name'])) map_func = getattr(self.entity_mapper, kwargs['map_func']) kwargs['map_func'] = map_func ent = Entity(domain=domain, **kwargs) domain.add_entity(ent) if ent.mandatory: self.mandatory.add(ent.id) if ent.directory is not None: ent.directory = ent.directory.replace('{{root}}', self.root) self.entities[ent.id] = ent for alias in ent.aliases: self.entities[alias] = ent if self.dynamic_getters: func = partial(getattr(self, 'get'), target=ent.name, return_type='id') func_name = inflect.engine().plural(ent.name) setattr(self, 'get_%s' % func_name, func)
add_entity
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, domain, **kwargs):\n map_func = kwargs.get('map_func', None)\n if map_func is not None and not callable(kwargs['map_func']):\n if self.entity_mapper is None:\n raise ValueError(\n \"Mapping function '%s' specified for Entity '%s', but no entity mapper was passed when initializing the current Layout. Please make sure the 'entity_mapper' argument is set.\"\n % (map_func, kwargs['name']))\n map_func = getattr(self.entity_mapper, kwargs['map_func'])\n kwargs['map_func'] = map_func\n ent = Entity(domain=domain, **kwargs)\n domain.add_entity(ent)\n if ent.mandatory:\n self.mandatory.add(ent.id)\n if ent.directory is not None:\n ent.directory = ent.directory.replace('{{root}}', self.root)\n self.entities[ent.id] = ent\n for alias in ent.aliases:\n self.entities[alias] = ent\n if self.dynamic_getters:\n func = partial(getattr(self, 'get'), target=ent.name, return_type='id')\n func_name = inflect.engine().plural(ent.name)\n setattr(self, 'get_%s' % func_name, func)\n", "role": "user" }, { "content": "add_entity", "role": "assistant" } ]
def x(self, return_type='tuple', target=None, extensions=None, domains=None, regex_search=None, **kwargs): if regex_search is None: regex_search = self.regex_search result = [] filters = {} filters.update(kwargs) for filename, file in self.files.items(): if not file._matches(filters, extensions, domains, regex_search): continue result.append(file) if not self.absolute_paths: for i, f in enumerate(result): f = copy(f) f.path = relpath(f.path, self.root) result[i] = f if return_type == 'file': return natural_sort([f.path for f in result]) if return_type == 'tuple': result = [r.as_named_tuple() for r in result] return natural_sort(result, field='filename') if return_type.startswith('obj'): return result else: valid_entities = self.get_domain_entities(domains) if target is None: raise ValueError( 'If return_type is "id" or "dir", a valid target entity must also be specified.' ) result = [x for x in result if target in x.entities] if return_type == 'id': result = list(set([x.entities[target] for x in result])) return natural_sort(result) elif return_type == 'dir': template = valid_entities[target].directory if template is None: raise ValueError( 'Return type set to directory, but no directory template is defined for the target entity ("%s").' % target) to_rep = re.findall('\\{(.*?)\\}', template) for ent in to_rep: patt = valid_entities[ent].pattern template = template.replace('{%s}' % ent, patt) template += '[^\\%s]*$' % os.path.sep matches = [f.dirname for f in result if re.search(template, f. dirname)] return natural_sort(list(set(matches))) else: raise ValueError( "Invalid return_type specified (must be one of 'tuple', 'file', 'id', or 'dir'." )
get
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, return_type='tuple', target=None, extensions=None, domains=None,\n regex_search=None, **kwargs):\n if regex_search is None:\n regex_search = self.regex_search\n result = []\n filters = {}\n filters.update(kwargs)\n for filename, file in self.files.items():\n if not file._matches(filters, extensions, domains, regex_search):\n continue\n result.append(file)\n if not self.absolute_paths:\n for i, f in enumerate(result):\n f = copy(f)\n f.path = relpath(f.path, self.root)\n result[i] = f\n if return_type == 'file':\n return natural_sort([f.path for f in result])\n if return_type == 'tuple':\n result = [r.as_named_tuple() for r in result]\n return natural_sort(result, field='filename')\n if return_type.startswith('obj'):\n return result\n else:\n valid_entities = self.get_domain_entities(domains)\n if target is None:\n raise ValueError(\n 'If return_type is \"id\" or \"dir\", a valid target entity must also be specified.'\n )\n result = [x for x in result if target in x.entities]\n if return_type == 'id':\n result = list(set([x.entities[target] for x in result]))\n return natural_sort(result)\n elif return_type == 'dir':\n template = valid_entities[target].directory\n if template is None:\n raise ValueError(\n 'Return type set to directory, but no directory template is defined for the target entity (\"%s\").'\n % target)\n to_rep = re.findall('\\\\{(.*?)\\\\}', template)\n for ent in to_rep:\n patt = valid_entities[ent].pattern\n template = template.replace('{%s}' % ent, patt)\n template += '[^\\\\%s]*$' % os.path.sep\n matches = [f.dirname for f in result if re.search(template, f.\n dirname)]\n return natural_sort(list(set(matches)))\n else:\n raise ValueError(\n \"Invalid return_type specified (must be one of 'tuple', 'file', 'id', or 'dir'.\"\n )\n", "role": "user" }, { "content": "get", "role": "assistant" } ]
def x(self, entity, files=False): return self._find_entity(entity).count(files)
count
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, entity, files=False):\n return self._find_entity(entity).count(files)\n", "role": "user" }, { "content": "count", "role": "assistant" } ]
def x(self, **kwargs): try: import pandas as pd except ImportError: raise ImportError( "What are you doing trying to export a Layout as a pandas DataFrame when you don't have pandas installed? Eh? Eh?" ) if kwargs: files = self.get(return_type='obj', **kwargs) else: files = self.files.values() data = pd.DataFrame.from_records([f.entities for f in files]) data.insert(0, 'path', [f.path for f in files]) return data
as_data_frame
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, **kwargs):\n try:\n import pandas as pd\n except ImportError:\n raise ImportError(\n \"What are you doing trying to export a Layout as a pandas DataFrame when you don't have pandas installed? Eh? Eh?\"\n )\n if kwargs:\n files = self.get(return_type='obj', **kwargs)\n else:\n files = self.files.values()\n data = pd.DataFrame.from_records([f.entities for f in files])\n data.insert(0, 'path', [f.path for f in files])\n return data\n", "role": "user" }, { "content": "as_data_frame", "role": "assistant" } ]
def x(self, path, return_type='file', strict=True, all_=False, ignore_strict_entities=None, full_search=False, **kwargs): entities = {} for ent in self.entities.values(): m = ent.regex.search(path) if m: entities[ent.name] = ent._astype(m.group(1)) if strict and ignore_strict_entities is not None: for k in ignore_strict_entities: entities.pop(k, None) results = self.get(return_type='file', **kwargs) folders = defaultdict(list) for filename in results: f = self.get_file(filename) folders[f.dirname].append(f) def x(f): f_ents = f.entities keys = set(entities.keys()) & set(f_ents.keys()) shared = len(keys) return [shared, sum([(entities[k] == f_ents[k]) for k in keys])] matches = [] search_paths = [] while True: if path in folders and folders[path]: search_paths.append(path) parent = dirname(path) if parent == path: break path = parent if full_search: unchecked = set(folders.keys()) - set(search_paths) search_paths.extend(path for path in unchecked if folders[path]) for path in search_paths: num_ents = [([f] + count_matches(f)) for f in folders[path]] if strict: num_ents = [f for f in num_ents if f[1] == f[2]] num_ents.sort(key=lambda x: x[2], reverse=True) if num_ents: matches.append(num_ents[0][0]) if not all_: break matches = [(m.path if return_type == 'file' else m.as_named_tuple()) for m in matches] return matches if all_ else matches[0] if matches else None
get_nearest
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, path, return_type='file', strict=True, all_=False,\n ignore_strict_entities=None, full_search=False, **kwargs):\n entities = {}\n for ent in self.entities.values():\n m = ent.regex.search(path)\n if m:\n entities[ent.name] = ent._astype(m.group(1))\n if strict and ignore_strict_entities is not None:\n for k in ignore_strict_entities:\n entities.pop(k, None)\n results = self.get(return_type='file', **kwargs)\n folders = defaultdict(list)\n for filename in results:\n f = self.get_file(filename)\n folders[f.dirname].append(f)\n\n def x(f):\n f_ents = f.entities\n keys = set(entities.keys()) & set(f_ents.keys())\n shared = len(keys)\n return [shared, sum([(entities[k] == f_ents[k]) for k in keys])]\n matches = []\n search_paths = []\n while True:\n if path in folders and folders[path]:\n search_paths.append(path)\n parent = dirname(path)\n if parent == path:\n break\n path = parent\n if full_search:\n unchecked = set(folders.keys()) - set(search_paths)\n search_paths.extend(path for path in unchecked if folders[path])\n for path in search_paths:\n num_ents = [([f] + count_matches(f)) for f in folders[path]]\n if strict:\n num_ents = [f for f in num_ents if f[1] == f[2]]\n num_ents.sort(key=lambda x: x[2], reverse=True)\n if num_ents:\n matches.append(num_ents[0][0])\n if not all_:\n break\n matches = [(m.path if return_type == 'file' else m.as_named_tuple()) for\n m in matches]\n return matches if all_ else matches[0] if matches else None\n", "role": "user" }, { "content": "get_nearest", "role": "assistant" } ]
def x(self, source, path_patterns=None, strict=False, domains=None): if isinstance(source, six.string_types): if source not in self.files: source = join(self.root, source) source = self.get_file(source) if isinstance(source, File): source = source.entities if path_patterns is None: if domains is None: domains = list(self.domains.keys()) path_patterns = [] for dom in listify(domains): path_patterns.extend(self.domains[dom].path_patterns) return build_path(source, path_patterns, strict)
build_path
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, source, path_patterns=None, strict=False, domains=None):\n if isinstance(source, six.string_types):\n if source not in self.files:\n source = join(self.root, source)\n source = self.get_file(source)\n if isinstance(source, File):\n source = source.entities\n if path_patterns is None:\n if domains is None:\n domains = list(self.domains.keys())\n path_patterns = []\n for dom in listify(domains):\n path_patterns.extend(self.domains[dom].path_patterns)\n return build_path(source, path_patterns, strict)\n", "role": "user" }, { "content": "build_path", "role": "assistant" } ]
def x(self, entities, path_patterns=None, contents=None, link_to=None, content_mode='text', conflicts='fail', strict=False, domains=None, index=False, index_domains=None): path = self.build_path(entities, path_patterns, strict, domains) if path is None: raise ValueError( 'Cannot construct any valid filename for the passed entities given available path patterns.' ) write_contents_to_file(path, contents=contents, link_to=link_to, content_mode=content_mode, conflicts=conflicts, root=self.root) if index: if index_domains is None: index_domains = list(self.domains.keys()) self._index_file(self.root, path, index_domains)
write_contents_to_file
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, entities, path_patterns=None, contents=None, link_to=None,\n content_mode='text', conflicts='fail', strict=False, domains=None,\n index=False, index_domains=None):\n path = self.build_path(entities, path_patterns, strict, domains)\n if path is None:\n raise ValueError(\n 'Cannot construct any valid filename for the passed entities given available path patterns.'\n )\n write_contents_to_file(path, contents=contents, link_to=link_to,\n content_mode=content_mode, conflicts=conflicts, root=self.root)\n if index:\n if index_domains is None:\n index_domains = list(self.domains.keys())\n self._index_file(self.root, path, index_domains)\n", "role": "user" }, { "content": "write_contents_to_file", "role": "assistant" } ]
def x(name, log_stream=sys.stdout, log_file=None, log_level=logging.INFO, keep_old_handlers=False, propagate=False): logger = logging.getLogger(name) logger.setLevel(log_level) logger.propagate = propagate if not keep_old_handlers: logger.handlers = [] log_fmt = '[%(asctime)s] %(levelname)s: %(message)s' log_datefmt = '%Y-%m-%d %H:%M:%S' formatter = logging.Formatter(log_fmt, log_datefmt) if log_stream is not None: stream_handler = logging.StreamHandler(log_stream) stream_handler.setFormatter(formatter) logger.addHandler(stream_handler) if log_file is not None: file_handler = logging.FileHandler(log_file) file_handler.setFormatter(formatter) logger.addHandler(file_handler) if log_stream is None and log_file is None: logger.addHandler(logging.NullHandler()) return logger
configure_logger
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(name, log_stream=sys.stdout, log_file=None, log_level=logging.INFO,\n keep_old_handlers=False, propagate=False):\n logger = logging.getLogger(name)\n logger.setLevel(log_level)\n logger.propagate = propagate\n if not keep_old_handlers:\n logger.handlers = []\n log_fmt = '[%(asctime)s] %(levelname)s: %(message)s'\n log_datefmt = '%Y-%m-%d %H:%M:%S'\n formatter = logging.Formatter(log_fmt, log_datefmt)\n if log_stream is not None:\n stream_handler = logging.StreamHandler(log_stream)\n stream_handler.setFormatter(formatter)\n logger.addHandler(stream_handler)\n if log_file is not None:\n file_handler = logging.FileHandler(log_file)\n file_handler.setFormatter(formatter)\n logger.addHandler(file_handler)\n if log_stream is None and log_file is None:\n logger.addHandler(logging.NullHandler())\n return logger\n", "role": "user" }, { "content": "configure_logger", "role": "assistant" } ]
def x(name='', log_stream=None, log_file=None, quiet=False, verbose=False): log_level = logging.INFO if quiet: log_level = logging.WARNING elif verbose: log_level = logging.DEBUG if log_stream is None: log_stream = sys.stdout new_logger = configure_logger(name, log_stream=log_stream, log_file= log_file, log_level=log_level) return new_logger
get_logger
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(name='', log_stream=None, log_file=None, quiet=False, verbose=False):\n log_level = logging.INFO\n if quiet:\n log_level = logging.WARNING\n elif verbose:\n log_level = logging.DEBUG\n if log_stream is None:\n log_stream = sys.stdout\n new_logger = configure_logger(name, log_stream=log_stream, log_file=\n log_file, log_level=log_level)\n return new_logger\n", "role": "user" }, { "content": "get_logger", "role": "assistant" } ]
def x(milliseconds, func, *args, **kwargs): stopper = threading.Event() def x(seconds, func, *args, **kwargs): def x(): if stopper.isSet(): return interval(seconds, func, *args, **kwargs) try: func(*args, **kwargs) except: logging.error('Error during interval') logging.error(traceback.format_exc()) thread = threading.Timer(seconds, wrapper) thread.daemon = True thread.start() interval(milliseconds / 1000, func, *args, **kwargs) return stopper
start
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(milliseconds, func, *args, **kwargs):\n stopper = threading.Event()\n\n def x(seconds, func, *args, **kwargs):\n\n def x():\n if stopper.isSet():\n return\n interval(seconds, func, *args, **kwargs)\n try:\n func(*args, **kwargs)\n except:\n logging.error('Error during interval')\n logging.error(traceback.format_exc())\n thread = threading.Timer(seconds, wrapper)\n thread.daemon = True\n thread.start()\n interval(milliseconds / 1000, func, *args, **kwargs)\n return stopper\n", "role": "user" }, { "content": "start", "role": "assistant" } ]
def x(codons1, codons2): result = defaultdict(list) for c1 in codons1: for c2 in codons2: distance = findDistance(c1, c2) result[distance].append([c1, c2]) return result
codonInformation
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(codons1, codons2):\n result = defaultdict(list)\n for c1 in codons1:\n for c2 in codons2:\n distance = findDistance(c1, c2)\n result[distance].append([c1, c2])\n return result\n", "role": "user" }, { "content": "codonInformation", "role": "assistant" } ]
def x(api_client): try: pprint((yield from api_client.echo())) except errors.RequestError as exc: log.exception('Exception occurred: %s', exc) yield gen.Task(lambda *args, **kwargs: ioloop.IOLoop.current().stop())
example_async_client
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(api_client):\n try:\n pprint((yield from api_client.echo()))\n except errors.RequestError as exc:\n log.exception('Exception occurred: %s', exc)\n yield gen.Task(lambda *args, **kwargs: ioloop.IOLoop.current().stop())\n", "role": "user" }, { "content": "example_async_client", "role": "assistant" } ]
def x(api_client): try: pprint(api_client.echo()) except errors.RequestError as exc: log.exception('Exception occurred: %s', exc)
example_sync_client
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(api_client):\n try:\n pprint(api_client.echo())\n except errors.RequestError as exc:\n log.exception('Exception occurred: %s', exc)\n", "role": "user" }, { "content": "example_sync_client", "role": "assistant" } ]
def x(): logging.basicConfig(level=logging.INFO) example_sync_client(SyncAPIClient()) example_async_client(AsyncAPIClient()) io_loop = ioloop.IOLoop.current() io_loop.start()
main
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n logging.basicConfig(level=logging.INFO)\n example_sync_client(SyncAPIClient())\n example_async_client(AsyncAPIClient())\n io_loop = ioloop.IOLoop.current()\n io_loop.start()\n", "role": "user" }, { "content": "main", "role": "assistant" } ]
def x(args=None): vinfo = sys.version_info if not vinfo >= (2, 7): raise SystemError( 'Python interpreter version >= 2.7 required, found %d.%d instead.' % (vinfo.major, vinfo.minor)) if args is None: parser = get_argument_parser() args = parser.parse_args() input_file = args.annotation_file output_file = args.output_file chrom_pat = args.chromosome_pattern log_file = args.log_file quiet = args.quiet verbose = args.verbose log_stream = sys.stdout if output_file == '-': log_stream = sys.stderr logger = misc.get_logger(log_stream=log_stream, log_file=log_file, quiet=quiet, verbose=verbose) if chrom_pat is not None: logger.info( 'Regular expression used for filtering chromosome names: "%s"', chrom_pat) if input_file == '-': input_file = sys.stdin if output_file == '-': output_file = sys.stdout genes = ensembl.get_protein_coding_genes(input_file, chromosome_pattern =chrom_pat) genes.to_csv(output_file, sep='\t', index=False) return 0
main
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(args=None):\n vinfo = sys.version_info\n if not vinfo >= (2, 7):\n raise SystemError(\n 'Python interpreter version >= 2.7 required, found %d.%d instead.'\n % (vinfo.major, vinfo.minor))\n if args is None:\n parser = get_argument_parser()\n args = parser.parse_args()\n input_file = args.annotation_file\n output_file = args.output_file\n chrom_pat = args.chromosome_pattern\n log_file = args.log_file\n quiet = args.quiet\n verbose = args.verbose\n log_stream = sys.stdout\n if output_file == '-':\n log_stream = sys.stderr\n logger = misc.get_logger(log_stream=log_stream, log_file=log_file,\n quiet=quiet, verbose=verbose)\n if chrom_pat is not None:\n logger.info(\n 'Regular expression used for filtering chromosome names: \"%s\"',\n chrom_pat)\n if input_file == '-':\n input_file = sys.stdin\n if output_file == '-':\n output_file = sys.stdout\n genes = ensembl.get_protein_coding_genes(input_file, chromosome_pattern\n =chrom_pat)\n genes.to_csv(output_file, sep='\\t', index=False)\n return 0\n", "role": "user" }, { "content": "main", "role": "assistant" } ]
def x(self, *stats): self.statistics = OrderedDict() for port in self.ports: port_stats = IxeStatTotal(port).get_attributes(FLAG_RDONLY, *stats) port_stats.update({(c + '_rate'): v for c, v in IxeStatRate(port). get_attributes(FLAG_RDONLY, *stats).items()}) self.statistics[str(port)] = port_stats return self.statistics
read_stats
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, *stats):\n self.statistics = OrderedDict()\n for port in self.ports:\n port_stats = IxeStatTotal(port).get_attributes(FLAG_RDONLY, *stats)\n port_stats.update({(c + '_rate'): v for c, v in IxeStatRate(port).\n get_attributes(FLAG_RDONLY, *stats).items()})\n self.statistics[str(port)] = port_stats\n return self.statistics\n", "role": "user" }, { "content": "read_stats", "role": "assistant" } ]
def x(self, *stats): from ixexplorer.ixe_stream import IxePacketGroupStream sleep_time = 0.1 if not stats: stats = [m.attrname for m in IxePgStats.__tcl_members__ if m.flags & FLAG_RDONLY] sleep_time = 1 for port in self.tx_ports_streams: port.api.call_rc('streamTransmitStats get {} 1 4096'.format(port.uri)) for rx_port in self.rx_ports: rx_port.api.call_rc('packetGroupStats get {} 0 65536'.format( rx_port.uri)) time.sleep(sleep_time) self.statistics = OrderedDict() for tx_port, streams in self.tx_ports_streams.items(): for stream in streams: stream_stats = OrderedDict() tx_port.api.call_rc('streamTransmitStats get {} 1 4096'.format( tx_port.uri)) stream_tx_stats = IxeStreamTxStats(tx_port, stream.index) stream_stats_tx = {c: v for c, v in stream_tx_stats. get_attributes(FLAG_RDONLY).items()} stream_stats['tx'] = stream_stats_tx stream_stat_pgid = IxePacketGroupStream(stream).groupId stream_stats_pg = pg_stats_dict() for port in self.session.ports.values(): stream_stats_pg[str(port)] = OrderedDict(zip(stats, [-1] * len(stats))) for rx_port in self.rx_ports: if not stream.rx_ports or rx_port in stream.rx_ports: rx_port.api.call_rc('packetGroupStats get {} 0 65536'. format(rx_port.uri)) pg_stats = IxePgStats(rx_port, stream_stat_pgid) stream_stats_pg[str(rx_port)] = pg_stats.read_stats(*stats) stream_stats['rx'] = stream_stats_pg self.statistics[str(stream)] = stream_stats return self.statistics
read_stats
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, *stats):\n from ixexplorer.ixe_stream import IxePacketGroupStream\n sleep_time = 0.1\n if not stats:\n stats = [m.attrname for m in IxePgStats.__tcl_members__ if m.flags &\n FLAG_RDONLY]\n sleep_time = 1\n for port in self.tx_ports_streams:\n port.api.call_rc('streamTransmitStats get {} 1 4096'.format(port.uri))\n for rx_port in self.rx_ports:\n rx_port.api.call_rc('packetGroupStats get {} 0 65536'.format(\n rx_port.uri))\n time.sleep(sleep_time)\n self.statistics = OrderedDict()\n for tx_port, streams in self.tx_ports_streams.items():\n for stream in streams:\n stream_stats = OrderedDict()\n tx_port.api.call_rc('streamTransmitStats get {} 1 4096'.format(\n tx_port.uri))\n stream_tx_stats = IxeStreamTxStats(tx_port, stream.index)\n stream_stats_tx = {c: v for c, v in stream_tx_stats.\n get_attributes(FLAG_RDONLY).items()}\n stream_stats['tx'] = stream_stats_tx\n stream_stat_pgid = IxePacketGroupStream(stream).groupId\n stream_stats_pg = pg_stats_dict()\n for port in self.session.ports.values():\n stream_stats_pg[str(port)] = OrderedDict(zip(stats, [-1] *\n len(stats)))\n for rx_port in self.rx_ports:\n if not stream.rx_ports or rx_port in stream.rx_ports:\n rx_port.api.call_rc('packetGroupStats get {} 0 65536'.\n format(rx_port.uri))\n pg_stats = IxePgStats(rx_port, stream_stat_pgid)\n stream_stats_pg[str(rx_port)] = pg_stats.read_stats(*stats)\n stream_stats['rx'] = stream_stats_pg\n self.statistics[str(stream)] = stream_stats\n return self.statistics\n", "role": "user" }, { "content": "read_stats", "role": "assistant" } ]
def x(rst, clk, req_vec, gnt_vec=None, gnt_idx=None, gnt_vld=None, gnt_rdy= None, ARBITER_TYPE='priority'): if ARBITER_TYPE == 'priority': _arb = arbiter_priority(req_vec, gnt_vec, gnt_idx, gnt_vld) elif ARBITER_TYPE == 'roundrobin': _arb = arbiter_roundrobin(rst, clk, req_vec, gnt_vec, gnt_idx, gnt_vld, gnt_rdy) else: assert 'Arbiter: Unknown arbiter type: {}'.format(ARBITER_TYPE) return _arb
arbiter
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(rst, clk, req_vec, gnt_vec=None, gnt_idx=None, gnt_vld=None, gnt_rdy=\n None, ARBITER_TYPE='priority'):\n if ARBITER_TYPE == 'priority':\n _arb = arbiter_priority(req_vec, gnt_vec, gnt_idx, gnt_vld)\n elif ARBITER_TYPE == 'roundrobin':\n _arb = arbiter_roundrobin(rst, clk, req_vec, gnt_vec, gnt_idx,\n gnt_vld, gnt_rdy)\n else:\n assert 'Arbiter: Unknown arbiter type: {}'.format(ARBITER_TYPE)\n return _arb\n", "role": "user" }, { "content": "arbiter", "role": "assistant" } ]
def x(req_vec, gnt_vec=None, gnt_idx=None, gnt_vld=None): REQ_NUM = len(req_vec) gnt_vec_s = Signal(intbv(0)[REQ_NUM:]) gnt_idx_s = Signal(intbv(0, min=0, max=REQ_NUM)) gnt_vld_s = Signal(bool(0)) @always_comb def x(): gnt_vec_s.next = 0 gnt_idx_s.next = 0 gnt_vld_s.next = 0 for i in range(REQ_NUM): if req_vec[i] == 1: gnt_vec_s.next[i] = 1 gnt_idx_s.next = i gnt_vld_s.next = 1 break if gnt_vec != None: _vec = assign(gnt_vec, gnt_vec_s) if gnt_idx != None: _idx = assign(gnt_idx, gnt_idx_s) if gnt_vld != None: _vld = assign(gnt_vld, gnt_vld_s) return instances()
arbiter_priority
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(req_vec, gnt_vec=None, gnt_idx=None, gnt_vld=None):\n REQ_NUM = len(req_vec)\n gnt_vec_s = Signal(intbv(0)[REQ_NUM:])\n gnt_idx_s = Signal(intbv(0, min=0, max=REQ_NUM))\n gnt_vld_s = Signal(bool(0))\n\n @always_comb\n def x():\n gnt_vec_s.next = 0\n gnt_idx_s.next = 0\n gnt_vld_s.next = 0\n for i in range(REQ_NUM):\n if req_vec[i] == 1:\n gnt_vec_s.next[i] = 1\n gnt_idx_s.next = i\n gnt_vld_s.next = 1\n break\n if gnt_vec != None:\n _vec = assign(gnt_vec, gnt_vec_s)\n if gnt_idx != None:\n _idx = assign(gnt_idx, gnt_idx_s)\n if gnt_vld != None:\n _vld = assign(gnt_vld, gnt_vld_s)\n return instances()\n", "role": "user" }, { "content": "arbiter_priority", "role": "assistant" } ]
def x(rst, clk, req_vec, gnt_vec=None, gnt_idx=None, gnt_vld=None, gnt_rdy=None ): REQ_NUM = len(req_vec) ptr = Signal(intbv(0, min=0, max=REQ_NUM)) gnt_vec_s = Signal(intbv(0)[REQ_NUM:]) gnt_idx_s = Signal(intbv(0, min=0, max=REQ_NUM)) gnt_vld_s = Signal(bool(0)) @always(clk.posedge) def x(): if rst: ptr.next = REQ_NUM - 1 elif gnt_rdy and gnt_vld_s: ptr.next = gnt_idx_s @always_comb def x(): gnt_vec_s.next = 0 gnt_idx_s.next = 0 gnt_vld_s.next = 0 for i in range(REQ_NUM): if i > ptr: if req_vec[i] == 1: gnt_vec_s.next[i] = 1 gnt_idx_s.next = i gnt_vld_s.next = 1 return for i in range(REQ_NUM): if req_vec[i] == 1: gnt_vec_s.next[i] = 1 gnt_idx_s.next = i gnt_vld_s.next = 1 return if gnt_vec != None: _vec = assign(gnt_vec, gnt_vec_s) if gnt_idx != None: _idx = assign(gnt_idx, gnt_idx_s) if gnt_vld != None: _vld = assign(gnt_vld, gnt_vld_s) return instances()
arbiter_roundrobin
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(rst, clk, req_vec, gnt_vec=None, gnt_idx=None, gnt_vld=None, gnt_rdy=None\n ):\n REQ_NUM = len(req_vec)\n ptr = Signal(intbv(0, min=0, max=REQ_NUM))\n gnt_vec_s = Signal(intbv(0)[REQ_NUM:])\n gnt_idx_s = Signal(intbv(0, min=0, max=REQ_NUM))\n gnt_vld_s = Signal(bool(0))\n\n @always(clk.posedge)\n def x():\n if rst:\n ptr.next = REQ_NUM - 1\n elif gnt_rdy and gnt_vld_s:\n ptr.next = gnt_idx_s\n\n @always_comb\n def x():\n gnt_vec_s.next = 0\n gnt_idx_s.next = 0\n gnt_vld_s.next = 0\n for i in range(REQ_NUM):\n if i > ptr:\n if req_vec[i] == 1:\n gnt_vec_s.next[i] = 1\n gnt_idx_s.next = i\n gnt_vld_s.next = 1\n return\n for i in range(REQ_NUM):\n if req_vec[i] == 1:\n gnt_vec_s.next[i] = 1\n gnt_idx_s.next = i\n gnt_vld_s.next = 1\n return\n if gnt_vec != None:\n _vec = assign(gnt_vec, gnt_vec_s)\n if gnt_idx != None:\n _idx = assign(gnt_idx, gnt_idx_s)\n if gnt_vld != None:\n _vld = assign(gnt_vld, gnt_vld_s)\n return instances()\n", "role": "user" }, { "content": "arbiter_roundrobin", "role": "assistant" } ]
def x(seq, normalize=True): seq = normalize_sequence(seq) if normalize else seq return str(vmc_digest(seq, digest_size=24))
seq_seqhash
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(seq, normalize=True):\n seq = normalize_sequence(seq) if normalize else seq\n return str(vmc_digest(seq, digest_size=24))\n", "role": "user" }, { "content": "seq_seqhash", "role": "assistant" } ]
def x(seq, normalize=True): seq = normalize_sequence(seq) if normalize else seq bseq = seq.encode('ascii') return base64.b64encode(hashlib.sha1(bseq).digest()).decode('ascii' ).rstrip('=')
seq_seguid
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(seq, normalize=True):\n seq = normalize_sequence(seq) if normalize else seq\n bseq = seq.encode('ascii')\n return base64.b64encode(hashlib.sha1(bseq).digest()).decode('ascii'\n ).rstrip('=')\n", "role": "user" }, { "content": "seq_seguid", "role": "assistant" } ]
def x(seq, normalize=True): seq = normalize_sequence(seq) if normalize else seq bseq = seq.encode('ascii') return hashlib.md5(bseq).hexdigest()
seq_md5
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(seq, normalize=True):\n seq = normalize_sequence(seq) if normalize else seq\n bseq = seq.encode('ascii')\n return hashlib.md5(bseq).hexdigest()\n", "role": "user" }, { "content": "seq_md5", "role": "assistant" } ]
def x(seq, normalize=True): seq = normalize_sequence(seq) if normalize else seq bseq = seq.encode('ascii') return hashlib.sha1(bseq).hexdigest()
seq_sha1
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(seq, normalize=True):\n seq = normalize_sequence(seq) if normalize else seq\n bseq = seq.encode('ascii')\n return hashlib.sha1(bseq).hexdigest()\n", "role": "user" }, { "content": "seq_sha1", "role": "assistant" } ]
def x(seq, normalize=True): seq = normalize_sequence(seq) if normalize else seq bseq = seq.encode('ascii') return hashlib.sha512(bseq).hexdigest()
seq_sha512
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(seq, normalize=True):\n seq = normalize_sequence(seq) if normalize else seq\n bseq = seq.encode('ascii')\n return hashlib.sha512(bseq).hexdigest()\n", "role": "user" }, { "content": "seq_sha512", "role": "assistant" } ]
def x(seq, normalize=True): seq = normalize_sequence(seq) if normalize else seq return {'namespace': 'VMC', 'accession': 'GS_' + str(vmc_digest(seq))}
seq_vmc_identifier
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(seq, normalize=True):\n seq = normalize_sequence(seq) if normalize else seq\n return {'namespace': 'VMC', 'accession': 'GS_' + str(vmc_digest(seq))}\n", "role": "user" }, { "content": "seq_vmc_identifier", "role": "assistant" } ]
def x(credentials, instance_config, instance_name, script_dir, index_dir, fastq_file, output_dir, num_threads=None, seed_start_lmax=None, mismatch_nmax=None, multimap_nmax=None, splice_min_overhang=None, out_mult_nmax=None, sort_bam=True, keep_unmapped=False, self_destruct= True, compressed=True, **kwargs): if sort_bam: out_sam_type = 'BAM SortedByCoordinate' else: out_sam_type = 'BAM Unsorted' fastq_files = fastq_file if isinstance(fastq_files, (str, _oldstr)): fastq_files = [fastq_file] template = _TEMPLATE_ENV.get_template(os.path.join('map_single-end.sh')) startup_script = template.render(script_dir=script_dir, index_dir= index_dir, fastq_files=fastq_files, output_dir=output_dir, num_threads=num_threads, seed_start_lmax=seed_start_lmax, self_destruct=self_destruct, mismatch_nmax=mismatch_nmax, multimap_nmax=multimap_nmax, splice_min_overhang= splice_min_overhang, out_mult_nmax=out_mult_nmax, keep_unmapped= keep_unmapped, compressed=compressed, out_sam_type=out_sam_type) if len(startup_script) > 32768: raise ValueError('Startup script larger than 32,768 bytes!') op_name = instance_config.create_instance(credentials, instance_name, startup_script=startup_script, **kwargs) return op_name
map_single_end
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(credentials, instance_config, instance_name, script_dir, index_dir,\n fastq_file, output_dir, num_threads=None, seed_start_lmax=None,\n mismatch_nmax=None, multimap_nmax=None, splice_min_overhang=None,\n out_mult_nmax=None, sort_bam=True, keep_unmapped=False, self_destruct=\n True, compressed=True, **kwargs):\n if sort_bam:\n out_sam_type = 'BAM SortedByCoordinate'\n else:\n out_sam_type = 'BAM Unsorted'\n fastq_files = fastq_file\n if isinstance(fastq_files, (str, _oldstr)):\n fastq_files = [fastq_file]\n template = _TEMPLATE_ENV.get_template(os.path.join('map_single-end.sh'))\n startup_script = template.render(script_dir=script_dir, index_dir=\n index_dir, fastq_files=fastq_files, output_dir=output_dir,\n num_threads=num_threads, seed_start_lmax=seed_start_lmax,\n self_destruct=self_destruct, mismatch_nmax=mismatch_nmax,\n multimap_nmax=multimap_nmax, splice_min_overhang=\n splice_min_overhang, out_mult_nmax=out_mult_nmax, keep_unmapped=\n keep_unmapped, compressed=compressed, out_sam_type=out_sam_type)\n if len(startup_script) > 32768:\n raise ValueError('Startup script larger than 32,768 bytes!')\n op_name = instance_config.create_instance(credentials, instance_name,\n startup_script=startup_script, **kwargs)\n return op_name\n", "role": "user" }, { "content": "map_single_end", "role": "assistant" } ]
def x(credentials, instance_config, instance_name, script_dir, genome_file, output_dir, annotation_file=None, splice_overhang=100, num_threads=8, chromosome_bin_bits=18, genome_memory_limit=31000000000, self_destruct= True, **kwargs): template = _TEMPLATE_ENV.get_template(os.path.join('generate_index.sh')) startup_script = template.render(script_dir=script_dir, genome_file= genome_file, annotation_file=annotation_file, splice_overhang= splice_overhang, output_dir=output_dir, num_threads=num_threads, chromosome_bin_bits=chromosome_bin_bits, genome_memory_limit= genome_memory_limit, self_destruct=self_destruct) op_name = instance_config.create_instance(credentials, instance_name, startup_script=startup_script, **kwargs) return op_name
generate_index
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(credentials, instance_config, instance_name, script_dir, genome_file,\n output_dir, annotation_file=None, splice_overhang=100, num_threads=8,\n chromosome_bin_bits=18, genome_memory_limit=31000000000, self_destruct=\n True, **kwargs):\n template = _TEMPLATE_ENV.get_template(os.path.join('generate_index.sh'))\n startup_script = template.render(script_dir=script_dir, genome_file=\n genome_file, annotation_file=annotation_file, splice_overhang=\n splice_overhang, output_dir=output_dir, num_threads=num_threads,\n chromosome_bin_bits=chromosome_bin_bits, genome_memory_limit=\n genome_memory_limit, self_destruct=self_destruct)\n op_name = instance_config.create_instance(credentials, instance_name,\n startup_script=startup_script, **kwargs)\n return op_name\n", "role": "user" }, { "content": "generate_index", "role": "assistant" } ]
def x(rst, clk, full, we, din, empty, re, dout, afull=None, aempty=None, afull_th=None, aempty_th=None, ovf=None, udf=None, count=None, count_max=None, depth=None, width=None): if width == None: width = 0 if din is not None: width = len(din) if depth == None: depth = 2 full_flg = Signal(bool(1)) empty_flg = Signal(bool(1)) we_safe = Signal(bool(0)) re_safe = Signal(bool(0)) rd_ptr = Signal(intbv(0, min=0, max=depth)) rd_ptr_new = Signal(intbv(0, min=0, max=depth)) wr_ptr = Signal(intbv(0, min=0, max=depth)) wr_ptr_new = Signal(intbv(0, min=0, max=depth)) @always_comb def x(): full.next = full_flg empty.next = empty_flg we_safe.next = we and not full_flg re_safe.next = re and not empty_flg @always_comb def x(): rd_ptr_new.next = (rd_ptr + 1) % depth wr_ptr_new.next = (wr_ptr + 1) % depth @always(clk.posedge) def x(): if rst: wr_ptr.next = 0 rd_ptr.next = 0 full_flg.next = 0 empty_flg.next = 1 else: if we_safe: wr_ptr.next = wr_ptr_new if re_safe: rd_ptr.next = rd_ptr_new if we_safe: empty_flg.next = 0 elif re_safe and rd_ptr_new == wr_ptr: empty_flg.next = 1 if re_safe: full_flg.next = 0 elif we_safe and wr_ptr_new == rd_ptr: full_flg.next = 1 """ Count """ if count != None or count_max != None or afull != None or aempty != None: count_r = Signal(intbv(0, min=0, max=depth + 1)) count_new = Signal(intbv(0, min=-1, max=depth + 2)) if count != None: assert count.max > depth @always_comb def x(): count.next = count_r @always_comb def x(): if we_safe and not re_safe: count_new.next = count_r + 1 elif not we_safe and re_safe: count_new.next = count_r - 1 else: count_new.next = count_r @always(clk.posedge) def x(): if rst: count_r.next = 0 else: count_r.next = count_new """ Count max """ if count_max != None: assert count_max.max > depth count_max_r = Signal(intbv(0, min=0, max=count_max.max)) @always(clk.posedge) def x(): if rst: count_max_r.next = 0 elif count_max_r < count_new: count_max_r.next = count_new @always_comb def x(): count_max.next = count_max_r """ AlmostFull flag """ if afull != None: if afull_th == None: afull_th = depth // 2 @always(clk.posedge) def x(): if rst: afull.next = 0 else: afull.next = count_new >= depth - afull_th """ AlmostEmpty flag """ if aempty != None: if aempty_th == None: aempty_th = depth // 2 @always(clk.posedge) def x(): if rst: aempty.next = 1 else: aempty.next = count_new <= aempty_th """ Overflow flag """ if ovf != None: @always(clk.posedge) def x(): if rst: ovf.next = 0 elif we and full_flg: ovf.next = 1 """ Underflow flag """ if udf != None: @always(clk.posedge) def x(): if rst: udf.next = 0 elif re and empty_flg: udf.next = 1 if width > 0: mem_we = Signal(bool(0)) mem_addrw = Signal(intbv(0, min=0, max=depth)) mem_addrr = Signal(intbv(0, min=0, max=depth)) mem_di = Signal(intbv(0)[width:0]) mem_do = Signal(intbv(0)[width:0]) mem = ram_sdp_ar(clk=clk, we=mem_we, addrw=mem_addrw, addrr= mem_addrr, di=mem_di, do=mem_do) @always_comb def x(): mem_we.next = we_safe mem_addrw.next = wr_ptr mem_addrr.next = rd_ptr mem_di.next = din dout.next = mem_do return instances()
fifo
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(rst, clk, full, we, din, empty, re, dout, afull=None, aempty=None,\n afull_th=None, aempty_th=None, ovf=None, udf=None, count=None,\n count_max=None, depth=None, width=None):\n if width == None:\n width = 0\n if din is not None:\n width = len(din)\n if depth == None:\n depth = 2\n full_flg = Signal(bool(1))\n empty_flg = Signal(bool(1))\n we_safe = Signal(bool(0))\n re_safe = Signal(bool(0))\n rd_ptr = Signal(intbv(0, min=0, max=depth))\n rd_ptr_new = Signal(intbv(0, min=0, max=depth))\n wr_ptr = Signal(intbv(0, min=0, max=depth))\n wr_ptr_new = Signal(intbv(0, min=0, max=depth))\n\n @always_comb\n def x():\n full.next = full_flg\n empty.next = empty_flg\n we_safe.next = we and not full_flg\n re_safe.next = re and not empty_flg\n\n @always_comb\n def x():\n rd_ptr_new.next = (rd_ptr + 1) % depth\n wr_ptr_new.next = (wr_ptr + 1) % depth\n\n @always(clk.posedge)\n def x():\n if rst:\n wr_ptr.next = 0\n rd_ptr.next = 0\n full_flg.next = 0\n empty_flg.next = 1\n else:\n if we_safe:\n wr_ptr.next = wr_ptr_new\n if re_safe:\n rd_ptr.next = rd_ptr_new\n if we_safe:\n empty_flg.next = 0\n elif re_safe and rd_ptr_new == wr_ptr:\n empty_flg.next = 1\n if re_safe:\n full_flg.next = 0\n elif we_safe and wr_ptr_new == rd_ptr:\n full_flg.next = 1\n \"\"\" Count \"\"\"\n if count != None or count_max != None or afull != None or aempty != None:\n count_r = Signal(intbv(0, min=0, max=depth + 1))\n count_new = Signal(intbv(0, min=-1, max=depth + 2))\n if count != None:\n assert count.max > depth\n\n @always_comb\n def x():\n count.next = count_r\n\n @always_comb\n def x():\n if we_safe and not re_safe:\n count_new.next = count_r + 1\n elif not we_safe and re_safe:\n count_new.next = count_r - 1\n else:\n count_new.next = count_r\n\n @always(clk.posedge)\n def x():\n if rst:\n count_r.next = 0\n else:\n count_r.next = count_new\n \"\"\" Count max \"\"\"\n if count_max != None:\n assert count_max.max > depth\n count_max_r = Signal(intbv(0, min=0, max=count_max.max))\n\n @always(clk.posedge)\n def x():\n if rst:\n count_max_r.next = 0\n elif count_max_r < count_new:\n count_max_r.next = count_new\n\n @always_comb\n def x():\n count_max.next = count_max_r\n \"\"\" AlmostFull flag \"\"\"\n if afull != None:\n if afull_th == None:\n afull_th = depth // 2\n\n @always(clk.posedge)\n def x():\n if rst:\n afull.next = 0\n else:\n afull.next = count_new >= depth - afull_th\n \"\"\" AlmostEmpty flag \"\"\"\n if aempty != None:\n if aempty_th == None:\n aempty_th = depth // 2\n\n @always(clk.posedge)\n def x():\n if rst:\n aempty.next = 1\n else:\n aempty.next = count_new <= aempty_th\n \"\"\" Overflow flag \"\"\"\n if ovf != None:\n\n @always(clk.posedge)\n def x():\n if rst:\n ovf.next = 0\n elif we and full_flg:\n ovf.next = 1\n \"\"\" Underflow flag \"\"\"\n if udf != None:\n\n @always(clk.posedge)\n def x():\n if rst:\n udf.next = 0\n elif re and empty_flg:\n udf.next = 1\n if width > 0:\n mem_we = Signal(bool(0))\n mem_addrw = Signal(intbv(0, min=0, max=depth))\n mem_addrr = Signal(intbv(0, min=0, max=depth))\n mem_di = Signal(intbv(0)[width:0])\n mem_do = Signal(intbv(0)[width:0])\n mem = ram_sdp_ar(clk=clk, we=mem_we, addrw=mem_addrw, addrr=\n mem_addrr, di=mem_di, do=mem_do)\n\n @always_comb\n def x():\n mem_we.next = we_safe\n mem_addrw.next = wr_ptr\n mem_addrr.next = rd_ptr\n mem_di.next = din\n dout.next = mem_do\n return instances()\n", "role": "user" }, { "content": "fifo", "role": "assistant" } ]
def x(url, ftp=None): assert isinstance(url, (str, _oldstr)) if ftp is not None: assert isinstance(ftp, ftplib.FTP) close_connection = False ftp_server = 'ftp.ensembl.org' ftp_user = 'anonymous' if ftp is None: ftp = ftplib.FTP(ftp_server) ftp.login(ftp_user) close_connection = True data = [] ftp.retrbinary('RETR %s' % url, data.append) data = ''.join(d.decode('utf-8') for d in data).split('\n')[:-1] file_checksums = OrderedDict() for d in data: file_name = d[d.rindex(' ') + 1:] sum_ = int(d[:d.index(' ')]) file_checksums[file_name] = sum_ logger.debug('Obtained checksums for %d files', len(file_checksums)) if close_connection: ftp.close() return file_checksums
get_file_checksums
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(url, ftp=None):\n assert isinstance(url, (str, _oldstr))\n if ftp is not None:\n assert isinstance(ftp, ftplib.FTP)\n close_connection = False\n ftp_server = 'ftp.ensembl.org'\n ftp_user = 'anonymous'\n if ftp is None:\n ftp = ftplib.FTP(ftp_server)\n ftp.login(ftp_user)\n close_connection = True\n data = []\n ftp.retrbinary('RETR %s' % url, data.append)\n data = ''.join(d.decode('utf-8') for d in data).split('\\n')[:-1]\n file_checksums = OrderedDict()\n for d in data:\n file_name = d[d.rindex(' ') + 1:]\n sum_ = int(d[:d.index(' ')])\n file_checksums[file_name] = sum_\n logger.debug('Obtained checksums for %d files', len(file_checksums))\n if close_connection:\n ftp.close()\n return file_checksums\n", "role": "user" }, { "content": "get_file_checksums", "role": "assistant" } ]
def x(obj, ignore=(list, tuple, type(None))): return obj if isinstance(obj, ignore) else [obj]
listify
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(obj, ignore=(list, tuple, type(None))):\n return obj if isinstance(obj, ignore) else [obj]\n", "role": "user" }, { "content": "listify", "role": "assistant" } ]
def x(taxdump_file): with tarfile.open(taxdump_file) as tf: with tf.extractfile('division.dmp') as fh: df = pd.read_csv(fh, header=None, sep='|', encoding='ascii') df = df.iloc[:, [0, 2]] df.iloc[:, 1] = df.iloc[:, 1].str.strip('\t') divisions = {} for _, row in df.iterrows(): divisions[row.iloc[1]] = row.iloc[0] return divisions
get_divisions
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(taxdump_file):\n with tarfile.open(taxdump_file) as tf:\n with tf.extractfile('division.dmp') as fh:\n df = pd.read_csv(fh, header=None, sep='|', encoding='ascii')\n df = df.iloc[:, [0, 2]]\n df.iloc[:, 1] = df.iloc[:, 1].str.strip('\\t')\n divisions = {}\n for _, row in df.iterrows():\n divisions[row.iloc[1]] = row.iloc[0]\n return divisions\n", "role": "user" }, { "content": "get_divisions", "role": "assistant" } ]
def x(taxdump_file, select_divisions=None, exclude_divisions=None): if select_divisions and exclude_divisions: raise ValueError( 'Cannot specify "select_divisions" and "exclude_divisions" at the same time.' ) select_division_ids = None exclude_division_ids = None divisions = None if select_divisions or exclude_divisions: divisions = _get_divisions(taxdump_file) if select_divisions: select_division_ids = set([divisions[d] for d in select_divisions]) elif exclude_divisions: exclude_division_ids = set([divisions[d] for d in exclude_divisions]) with tarfile.open(taxdump_file) as tf: with tf.extractfile('nodes.dmp') as fh: df = pd.read_csv(fh, header=None, sep='|', encoding='ascii') df = df.iloc[:, [0, 2, 4]] if select_division_ids: df = df.loc[df.iloc[:, 2].isin(select_division_ids)] elif exclude_division_ids: df = df.loc[~df.iloc[:, 2].isin(exclude_division_ids)] df.iloc[:, 1] = df.iloc[:, 1].str.strip('\t') taxon_ids = df.iloc[:, 0].loc[df.iloc[:, 1] == 'species'].values return taxon_ids
get_species_taxon_ids
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(taxdump_file, select_divisions=None, exclude_divisions=None):\n if select_divisions and exclude_divisions:\n raise ValueError(\n 'Cannot specify \"select_divisions\" and \"exclude_divisions\" at the same time.'\n )\n select_division_ids = None\n exclude_division_ids = None\n divisions = None\n if select_divisions or exclude_divisions:\n divisions = _get_divisions(taxdump_file)\n if select_divisions:\n select_division_ids = set([divisions[d] for d in select_divisions])\n elif exclude_divisions:\n exclude_division_ids = set([divisions[d] for d in exclude_divisions])\n with tarfile.open(taxdump_file) as tf:\n with tf.extractfile('nodes.dmp') as fh:\n df = pd.read_csv(fh, header=None, sep='|', encoding='ascii')\n df = df.iloc[:, [0, 2, 4]]\n if select_division_ids:\n df = df.loc[df.iloc[:, 2].isin(select_division_ids)]\n elif exclude_division_ids:\n df = df.loc[~df.iloc[:, 2].isin(exclude_division_ids)]\n df.iloc[:, 1] = df.iloc[:, 1].str.strip('\\t')\n taxon_ids = df.iloc[:, 0].loc[df.iloc[:, 1] == 'species'].values\n return taxon_ids\n", "role": "user" }, { "content": "get_species_taxon_ids", "role": "assistant" } ]
def x(taxdump_file, select_divisions=None, exclude_divisions=None, nrows=None): if select_divisions and exclude_divisions: raise ValueError( 'Cannot specify "select_divisions" and "exclude_divisions" at the same time.' ) select_taxon_ids = _get_species_taxon_ids(taxdump_file, select_divisions=select_divisions, exclude_divisions=exclude_divisions) select_taxon_ids = set(select_taxon_ids) with tarfile.open(taxdump_file) as tf: with tf.extractfile('names.dmp') as fh: df = pd.read_csv(fh, header=None, sep='|', encoding='ascii', nrows=nrows) df = df.iloc[:, [0, 1, 3]] df = df.loc[df.iloc[:, 0].isin(select_taxon_ids)] df.iloc[:, 2] = df.iloc[:, 2].str.strip('\t') df = df.loc[df.iloc[:, 2].isin(['scientific name', 'common name'])] df.iloc[:, 1] = df.iloc[:, 1].str.strip('\t') common_names = defaultdict(list) cn = df.loc[df.iloc[:, 2] == 'common name'] for _, row in cn.iterrows(): common_names[row.iloc[0]].append(row.iloc[1]) sn = df.loc[df.iloc[:, 2] == 'scientific name'] species = [] for i, row in sn.iterrows(): species.append([row.iloc[0], row.iloc[1], '|'.join(common_names[row .iloc[0]])]) species_df = pd.DataFrame(species).set_index(0) species_df.columns = ['scientific_name', 'common_names'] species_df.index.name = 'taxon_id' return species_df
get_species
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(taxdump_file, select_divisions=None, exclude_divisions=None, nrows=None):\n if select_divisions and exclude_divisions:\n raise ValueError(\n 'Cannot specify \"select_divisions\" and \"exclude_divisions\" at the same time.'\n )\n select_taxon_ids = _get_species_taxon_ids(taxdump_file,\n select_divisions=select_divisions, exclude_divisions=exclude_divisions)\n select_taxon_ids = set(select_taxon_ids)\n with tarfile.open(taxdump_file) as tf:\n with tf.extractfile('names.dmp') as fh:\n df = pd.read_csv(fh, header=None, sep='|', encoding='ascii',\n nrows=nrows)\n df = df.iloc[:, [0, 1, 3]]\n df = df.loc[df.iloc[:, 0].isin(select_taxon_ids)]\n df.iloc[:, 2] = df.iloc[:, 2].str.strip('\\t')\n df = df.loc[df.iloc[:, 2].isin(['scientific name', 'common name'])]\n df.iloc[:, 1] = df.iloc[:, 1].str.strip('\\t')\n common_names = defaultdict(list)\n cn = df.loc[df.iloc[:, 2] == 'common name']\n for _, row in cn.iterrows():\n common_names[row.iloc[0]].append(row.iloc[1])\n sn = df.loc[df.iloc[:, 2] == 'scientific name']\n species = []\n for i, row in sn.iterrows():\n species.append([row.iloc[0], row.iloc[1], '|'.join(common_names[row\n .iloc[0]])])\n species_df = pd.DataFrame(species).set_index(0)\n species_df.columns = ['scientific_name', 'common_names']\n species_df.index.name = 'taxon_id'\n return species_df\n", "role": "user" }, { "content": "get_species", "role": "assistant" } ]
def x(self, **attributes): auto_set = IxeObject.get_auto_set() IxeObject.set_auto_set(False) for name, value in attributes.items(): setattr(self, name, value) if auto_set: self.ix_set() IxeObject.set_auto_set(auto_set)
set_attributes
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, **attributes):\n auto_set = IxeObject.get_auto_set()\n IxeObject.set_auto_set(False)\n for name, value in attributes.items():\n setattr(self, name, value)\n if auto_set:\n self.ix_set()\n IxeObject.set_auto_set(auto_set)\n", "role": "user" }, { "content": "set_attributes", "role": "assistant" } ]
def x(recordFilename, hitId, fastaFilename, eCutoff=None, db='nt', actualSequenceId=None): print('TODO: This function is not finished yet.') return start = time() if isinstance(recordFilename, str): allhits = findHits(recordFilename, set([hitId]), limit=100) else: allhits = recordFilename sequence = getSequence(hitId, db) fasta, summary = summarizeHits(allhits, fastaFilename, eCutoff=eCutoff) minX, maxX = summary['minX'], summary['maxX'] if actualSequenceId: pass print(summary['hitCount']) print('seq len =', len(sequence)) fasta = summary['fasta'] consensusLen = maxX - minX consensus = [None] * consensusLen for item in summary['items']: print('NEW HSP') printHSP(item['origHsp']) hsp = item['hsp'] print('HIT query-start=%d query-stop=%d subj-start=%d subj-stop=%d' % (hsp['queryStart'], hsp['queryEnd'], hsp['subjectStart'], hsp[ 'subjectEnd'])) if item['frame']['subject'] > 0: query = fasta[item['sequenceId']].seq else: query = fasta[item['sequenceId']].reverse_complement().seq print(' target:', sequence[hsp['queryStart']:hsp['queryEnd']].seq ) print(' query:', query) match = [] for index in range(hsp['subjectStart'], hsp['subjectEnd']): queryIndex = index - hsp['queryStart'] match.append('.' if query[queryIndex] == sequence[index] else '*') print(' match: %s%s' % (' ' * (hsp['subjectStart'] - hsp[ 'queryStart']), ''.join(match))) print(' score:', item['convertedE']) print(' match len:', hsp['subjectEnd'] - hsp['subjectStart']) print('subject frame:', item['frame']['subject']) for queryIndex, sequenceIndex in enumerate(range(hsp['queryStart'], hsp['queryEnd'])): consensusIndex = sequenceIndex + minX locus = consensus[consensusIndex] if locus is None: consensus[consensusIndex] = locus = defaultdict(int) locus[query[queryIndex]] += 1 for index in range(minX, 0): consensusIndex = index - minX if consensus[consensusIndex]: print('%d: %r' % (index, consensus[consensusIndex])) for index in range(0, len(sequence)): consensusIndex = index - minX try: if consensus[consensusIndex]: print('%d: %r (%s)' % (index, consensus[index], sequence[ index])) except KeyError: break for index in range(len(sequence), maxX): consensusIndex = index - minX if consensus[consensusIndex]: print('%d: %r' % (index, consensus[consensusIndex])) stop = time() report('Consensus sequence generated in %.3f mins.' % ((stop - start) / 60.0)) return summary, consensus
consensusSequence
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(recordFilename, hitId, fastaFilename, eCutoff=None, db='nt',\n actualSequenceId=None):\n print('TODO: This function is not finished yet.')\n return\n start = time()\n if isinstance(recordFilename, str):\n allhits = findHits(recordFilename, set([hitId]), limit=100)\n else:\n allhits = recordFilename\n sequence = getSequence(hitId, db)\n fasta, summary = summarizeHits(allhits, fastaFilename, eCutoff=eCutoff)\n minX, maxX = summary['minX'], summary['maxX']\n if actualSequenceId:\n pass\n print(summary['hitCount'])\n print('seq len =', len(sequence))\n fasta = summary['fasta']\n consensusLen = maxX - minX\n consensus = [None] * consensusLen\n for item in summary['items']:\n print('NEW HSP')\n printHSP(item['origHsp'])\n hsp = item['hsp']\n print('HIT query-start=%d query-stop=%d subj-start=%d subj-stop=%d' %\n (hsp['queryStart'], hsp['queryEnd'], hsp['subjectStart'], hsp[\n 'subjectEnd']))\n if item['frame']['subject'] > 0:\n query = fasta[item['sequenceId']].seq\n else:\n query = fasta[item['sequenceId']].reverse_complement().seq\n print(' target:', sequence[hsp['queryStart']:hsp['queryEnd']].seq\n )\n print(' query:', query)\n match = []\n for index in range(hsp['subjectStart'], hsp['subjectEnd']):\n queryIndex = index - hsp['queryStart']\n match.append('.' if query[queryIndex] == sequence[index] else '*')\n print(' match: %s%s' % (' ' * (hsp['subjectStart'] - hsp[\n 'queryStart']), ''.join(match)))\n print(' score:', item['convertedE'])\n print(' match len:', hsp['subjectEnd'] - hsp['subjectStart'])\n print('subject frame:', item['frame']['subject'])\n for queryIndex, sequenceIndex in enumerate(range(hsp['queryStart'],\n hsp['queryEnd'])):\n consensusIndex = sequenceIndex + minX\n locus = consensus[consensusIndex]\n if locus is None:\n consensus[consensusIndex] = locus = defaultdict(int)\n locus[query[queryIndex]] += 1\n for index in range(minX, 0):\n consensusIndex = index - minX\n if consensus[consensusIndex]:\n print('%d: %r' % (index, consensus[consensusIndex]))\n for index in range(0, len(sequence)):\n consensusIndex = index - minX\n try:\n if consensus[consensusIndex]:\n print('%d: %r (%s)' % (index, consensus[index], sequence[\n index]))\n except KeyError:\n break\n for index in range(len(sequence), maxX):\n consensusIndex = index - minX\n if consensus[consensusIndex]:\n print('%d: %r' % (index, consensus[consensusIndex]))\n stop = time()\n report('Consensus sequence generated in %.3f mins.' % ((stop - start) /\n 60.0))\n return summary, consensus\n", "role": "user" }, { "content": "consensusSequence", "role": "assistant" } ]
def x(self, posicao=0): handle = self._url_open('detalheCEPAction.do', {'Metodo': 'detalhe', 'TipoCep': 2, 'Posicao': posicao + 1, 'CEP': ''}) html = handle.read() return self._parse_detalhe(html)
detalhe
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, posicao=0):\n handle = self._url_open('detalheCEPAction.do', {'Metodo': 'detalhe',\n 'TipoCep': 2, 'Posicao': posicao + 1, 'CEP': ''})\n html = handle.read()\n return self._parse_detalhe(html)\n", "role": "user" }, { "content": "detalhe", "role": "assistant" } ]
def x(self, localidade, uf): url = 'consultaFaixaCepAction.do' data = {'UF': uf, 'Localidade': localidade.encode('cp1252'), 'cfm': '1', 'Metodo': 'listaFaixaCEP', 'TipoConsulta': 'faixaCep', 'StartRow': '1', 'EndRow': '10'} html = self._url_open(url, data).read() return self._parse_faixa(html)
consulta_faixa
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, localidade, uf):\n url = 'consultaFaixaCepAction.do'\n data = {'UF': uf, 'Localidade': localidade.encode('cp1252'), 'cfm': '1',\n 'Metodo': 'listaFaixaCEP', 'TipoConsulta': 'faixaCep', 'StartRow':\n '1', 'EndRow': '10'}\n html = self._url_open(url, data).read()\n return self._parse_faixa(html)\n", "role": "user" }, { "content": "consulta_faixa", "role": "assistant" } ]
def x(self, endereco, primeiro=False, uf=None, localidade=None, tipo=None, numero=None): if uf is None: url = 'consultaEnderecoAction.do' data = {'relaxation': endereco.encode('ISO-8859-1'), 'TipoCep': 'ALL', 'semelhante': 'N', 'cfm': 1, 'Metodo': 'listaLogradouro', 'TipoConsulta': 'relaxation', 'StartRow': '1', 'EndRow': '10'} else: url = 'consultaLogradouroAction.do' data = {'Logradouro': endereco.encode('ISO-8859-1'), 'UF': uf, 'TIPO': tipo, 'Localidade': localidade.encode('ISO-8859-1'), 'Numero': numero, 'cfm': 1, 'Metodo': 'listaLogradouro', 'TipoConsulta': 'logradouro', 'StartRow': '1', 'EndRow': '10'} h = self._url_open(url, data) html = h.read() if primeiro: return self.detalhe() else: return self._parse_tabela(html)
consulta
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, endereco, primeiro=False, uf=None, localidade=None, tipo=None,\n numero=None):\n if uf is None:\n url = 'consultaEnderecoAction.do'\n data = {'relaxation': endereco.encode('ISO-8859-1'), 'TipoCep':\n 'ALL', 'semelhante': 'N', 'cfm': 1, 'Metodo': 'listaLogradouro',\n 'TipoConsulta': 'relaxation', 'StartRow': '1', 'EndRow': '10'}\n else:\n url = 'consultaLogradouroAction.do'\n data = {'Logradouro': endereco.encode('ISO-8859-1'), 'UF': uf,\n 'TIPO': tipo, 'Localidade': localidade.encode('ISO-8859-1'),\n 'Numero': numero, 'cfm': 1, 'Metodo': 'listaLogradouro',\n 'TipoConsulta': 'logradouro', 'StartRow': '1', 'EndRow': '10'}\n h = self._url_open(url, data)\n html = h.read()\n if primeiro:\n return self.detalhe()\n else:\n return self._parse_tabela(html)\n", "role": "user" }, { "content": "consulta", "role": "assistant" } ]
def x(self, url, *, method='GET', headers=None, data=None, result_callback=None ): retries_left = self._conn_retries while True: s = self._make_session() try: cert = None if self._client_cert and self._client_key: cert = self._client_cert, self._client_key elif self._client_cert: cert = self._client_cert if self._verify_cert: verify = True if self._ca_certs: verify = self._ca_certs else: verify = False auth = None if self._username and self._password: auth = self._username, self._password response = s.request(method, url, data=data, timeout=self. _connect_timeout, cert=cert, headers=headers, verify=verify, auth=auth) """:type: requests.models.Response """ if 400 <= response.status_code < 500: raise ClientError(response.status_code, response.content) elif response.status_code >= 500: raise ServerError(response.status_code, response.content) try: if result_callback: return result_callback(response.content) except (ValueError, TypeError) as err: raise MalformedResponse(err) from None return response.content except (requests.exceptions.RequestException, requests.exceptions. BaseHTTPError) as exc: if self._conn_retries is None or retries_left <= 0: raise CommunicationError(exc) from None else: retries_left -= 1 retry_in = (self._conn_retries - retries_left) * 2 self._log.warning( 'Server communication error: %s. Retrying in %s seconds.', exc, retry_in) time.sleep(retry_in) continue finally: s.close()
request
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, url, *, method='GET', headers=None, data=None, result_callback=None\n ):\n retries_left = self._conn_retries\n while True:\n s = self._make_session()\n try:\n cert = None\n if self._client_cert and self._client_key:\n cert = self._client_cert, self._client_key\n elif self._client_cert:\n cert = self._client_cert\n if self._verify_cert:\n verify = True\n if self._ca_certs:\n verify = self._ca_certs\n else:\n verify = False\n auth = None\n if self._username and self._password:\n auth = self._username, self._password\n response = s.request(method, url, data=data, timeout=self.\n _connect_timeout, cert=cert, headers=headers, verify=verify,\n auth=auth)\n \"\"\":type: requests.models.Response\n \"\"\"\n if 400 <= response.status_code < 500:\n raise ClientError(response.status_code, response.content)\n elif response.status_code >= 500:\n raise ServerError(response.status_code, response.content)\n try:\n if result_callback:\n return result_callback(response.content)\n except (ValueError, TypeError) as err:\n raise MalformedResponse(err) from None\n return response.content\n except (requests.exceptions.RequestException, requests.exceptions.\n BaseHTTPError) as exc:\n if self._conn_retries is None or retries_left <= 0:\n raise CommunicationError(exc) from None\n else:\n retries_left -= 1\n retry_in = (self._conn_retries - retries_left) * 2\n self._log.warning(\n 'Server communication error: %s. Retrying in %s seconds.',\n exc, retry_in)\n time.sleep(retry_in)\n continue\n finally:\n s.close()\n", "role": "user" }, { "content": "request", "role": "assistant" } ]
def x(): sess = requests.Session() sess.mount('http://', requests.adapters.HTTPAdapter(max_retries=False)) sess.mount('https://', requests.adapters.HTTPAdapter(max_retries=False)) return sess
make_session
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n sess = requests.Session()\n sess.mount('http://', requests.adapters.HTTPAdapter(max_retries=False))\n sess.mount('https://', requests.adapters.HTTPAdapter(max_retries=False))\n return sess\n", "role": "user" }, { "content": "make_session", "role": "assistant" } ]
def x(tcga_id): payload = {'attachment': 'true', 'filters': json.dumps({'op': 'and', 'content': [{'op': 'in', 'content': {'field': 'cases.project.program.name', 'value': ['TCGA']}}, {'op': 'in', 'content': {'field': 'project.project_id', 'value': [tcga_id]}}]}), 'fields': 'case_id', 'expand': 'demographic,diagnoses,family_histories,exposures', 'format': 'JSON', 'pretty': 'true', 'size': 10000, 'filename': 'clinical.project-%s' % tcga_id} r = requests.get('https://gdc-api.nci.nih.gov/cases', params=payload) j = json.loads(r.content.decode()) clinical = {} valid = 0 for s in j: if 'diagnoses' not in s: continue valid += 1 assert len(s['diagnoses']) == 1 diag = s['diagnoses'][0] tcga_id = diag['submitter_id'][:12] clinical[tcga_id] = diag logger.info('Found clinical data for %d cases.', valid) df = pd.DataFrame.from_dict(clinical).T df.sort_index(inplace=True) return df
get_clinical_data
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(tcga_id):\n payload = {'attachment': 'true', 'filters': json.dumps({'op': 'and',\n 'content': [{'op': 'in', 'content': {'field':\n 'cases.project.program.name', 'value': ['TCGA']}}, {'op': 'in',\n 'content': {'field': 'project.project_id', 'value': [tcga_id]}}]}),\n 'fields': 'case_id', 'expand':\n 'demographic,diagnoses,family_histories,exposures', 'format':\n 'JSON', 'pretty': 'true', 'size': 10000, 'filename': \n 'clinical.project-%s' % tcga_id}\n r = requests.get('https://gdc-api.nci.nih.gov/cases', params=payload)\n j = json.loads(r.content.decode())\n clinical = {}\n valid = 0\n for s in j:\n if 'diagnoses' not in s:\n continue\n valid += 1\n assert len(s['diagnoses']) == 1\n diag = s['diagnoses'][0]\n tcga_id = diag['submitter_id'][:12]\n clinical[tcga_id] = diag\n logger.info('Found clinical data for %d cases.', valid)\n df = pd.DataFrame.from_dict(clinical).T\n df.sort_index(inplace=True)\n return df\n", "role": "user" }, { "content": "get_clinical_data", "role": "assistant" } ]
def x(tcga_id): payload = {'attachment': 'true', 'filters': json.dumps({'op': 'and', 'content': [{'op': 'in', 'content': {'field': 'cases.project.program.name', 'value': ['TCGA']}}, {'op': 'in', 'content': {'field': 'project.project_id', 'value': [tcga_id]}}]}), 'fields': 'case_id', 'expand': 'samples,samples.portions,samples.portions.analytes,samples.portions.analytes.aliquots,samples.portions.analytes.aliquots.annotations,samples.portions.analytes.annotations,samples.portions.submitter_id,samples.portions.slides,samples.portions.annotations,samples.portions.center' , 'format': 'JSON', 'pretty': 'true', 'size': 10000, 'filename': 'biospecimen.project-%s' % tcga_id} r = requests.get('https://gdc-api.nci.nih.gov/cases', params=payload) j = json.loads(r.content.decode()) biospec = {} valid = 0 for case in j: if 'samples' not in case: continue valid += 1 for s in case['samples']: tcga_id = s['submitter_id'][:16] del s['portions'] del s['submitter_id'] biospec[tcga_id] = s logger.info('Found biospecimen data for %d cases.', valid) df = pd.DataFrame.from_dict(biospec).T df.sort_index(inplace=True) return df
get_biospecimen_data
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(tcga_id):\n payload = {'attachment': 'true', 'filters': json.dumps({'op': 'and',\n 'content': [{'op': 'in', 'content': {'field':\n 'cases.project.program.name', 'value': ['TCGA']}}, {'op': 'in',\n 'content': {'field': 'project.project_id', 'value': [tcga_id]}}]}),\n 'fields': 'case_id', 'expand':\n 'samples,samples.portions,samples.portions.analytes,samples.portions.analytes.aliquots,samples.portions.analytes.aliquots.annotations,samples.portions.analytes.annotations,samples.portions.submitter_id,samples.portions.slides,samples.portions.annotations,samples.portions.center'\n , 'format': 'JSON', 'pretty': 'true', 'size': 10000, 'filename': \n 'biospecimen.project-%s' % tcga_id}\n r = requests.get('https://gdc-api.nci.nih.gov/cases', params=payload)\n j = json.loads(r.content.decode())\n biospec = {}\n valid = 0\n for case in j:\n if 'samples' not in case:\n continue\n valid += 1\n for s in case['samples']:\n tcga_id = s['submitter_id'][:16]\n del s['portions']\n del s['submitter_id']\n biospec[tcga_id] = s\n logger.info('Found biospecimen data for %d cases.', valid)\n df = pd.DataFrame.from_dict(biospec).T\n df.sort_index(inplace=True)\n return df\n", "role": "user" }, { "content": "get_biospecimen_data", "role": "assistant" } ]
def x(tcga_id): payload = {'filters': json.dumps({'op': 'and', 'content': [{'op': 'in', 'content': {'field': 'cases.project.program.name', 'value': ['TCGA' ]}}, {'op': 'in', 'content': {'field': 'cases.project.project_id', 'value': [tcga_id]}}, {'op': 'in', 'content': {'field': 'files.data_category', 'value': ['Copy Number Variation']}}, {'op': 'in', 'content': {'field': 'files.data_type', 'value': [ 'Masked Copy Number Segment']}}]}), 'return_type': 'manifest', 'size': 10000} r = requests.get('https://gdc-api.nci.nih.gov/files', params=payload) df = pd.read_csv(io.StringIO(r.text), sep='\t', header=0) logger.info('Obtained manifest with %d files.', df.shape[0]) return df
get_masked_cnv_manifest
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(tcga_id):\n payload = {'filters': json.dumps({'op': 'and', 'content': [{'op': 'in',\n 'content': {'field': 'cases.project.program.name', 'value': ['TCGA'\n ]}}, {'op': 'in', 'content': {'field': 'cases.project.project_id',\n 'value': [tcga_id]}}, {'op': 'in', 'content': {'field':\n 'files.data_category', 'value': ['Copy Number Variation']}}, {'op':\n 'in', 'content': {'field': 'files.data_type', 'value': [\n 'Masked Copy Number Segment']}}]}), 'return_type': 'manifest',\n 'size': 10000}\n r = requests.get('https://gdc-api.nci.nih.gov/files', params=payload)\n df = pd.read_csv(io.StringIO(r.text), sep='\\t', header=0)\n logger.info('Obtained manifest with %d files.', df.shape[0])\n return df\n", "role": "user" }, { "content": "get_masked_cnv_manifest", "role": "assistant" } ]
def x(file_ids): assert isinstance(file_ids, Iterable) payload = {'filters': json.dumps({'op': 'in', 'content': {'field': 'files.file_id', 'value': list(file_ids)}}), 'fields': 'file_id,cases.samples.submitter_id', 'size': 10000} r = requests.post('https://gdc-api.nci.nih.gov/files', data=payload) j = json.loads(r.content.decode('utf-8')) file_samples = OrderedDict() for hit in j['data']['hits']: file_id = hit['file_id'] assert len(hit['cases']) == 1 case = hit['cases'][0] assert len(case['samples']) == 1 sample = case['samples'][0] sample_barcode = sample['submitter_id'] file_samples[file_id] = sample_barcode df = pd.DataFrame.from_dict(file_samples, orient='index') df = df.reset_index() df.columns = ['file_id', 'sample_barcode'] return df
get_file_samples
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(file_ids):\n assert isinstance(file_ids, Iterable)\n payload = {'filters': json.dumps({'op': 'in', 'content': {'field':\n 'files.file_id', 'value': list(file_ids)}}), 'fields':\n 'file_id,cases.samples.submitter_id', 'size': 10000}\n r = requests.post('https://gdc-api.nci.nih.gov/files', data=payload)\n j = json.loads(r.content.decode('utf-8'))\n file_samples = OrderedDict()\n for hit in j['data']['hits']:\n file_id = hit['file_id']\n assert len(hit['cases']) == 1\n case = hit['cases'][0]\n assert len(case['samples']) == 1\n sample = case['samples'][0]\n sample_barcode = sample['submitter_id']\n file_samples[file_id] = sample_barcode\n df = pd.DataFrame.from_dict(file_samples, orient='index')\n df = df.reset_index()\n df.columns = ['file_id', 'sample_barcode']\n return df\n", "role": "user" }, { "content": "get_file_samples", "role": "assistant" } ]
def x(file_samples): assert isinstance(file_samples, pd.DataFrame) df = file_samples df = df.sort_values('file_id') logger.info('Original number of files: %d', len(df.index)) df.drop_duplicates('sample_barcode', keep='first', inplace=True) logger.info( 'Number of files after removing duplicates from different aliquots: %d' , len(df.index)) df['sample_barcode15'] = df['sample_barcode'].apply(lambda x: x[:15]) df.drop_duplicates('sample_barcode15', keep='first', inplace=True) logger.info( 'Number of files after removing duplicates from different vials: %d', len(df.index)) df.drop('sample_barcode15', axis=1, inplace=True) df.sort_index(inplace=True) return df
get_unique_sample_files
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(file_samples):\n assert isinstance(file_samples, pd.DataFrame)\n df = file_samples\n df = df.sort_values('file_id')\n logger.info('Original number of files: %d', len(df.index))\n df.drop_duplicates('sample_barcode', keep='first', inplace=True)\n logger.info(\n 'Number of files after removing duplicates from different aliquots: %d'\n , len(df.index))\n df['sample_barcode15'] = df['sample_barcode'].apply(lambda x: x[:15])\n df.drop_duplicates('sample_barcode15', keep='first', inplace=True)\n logger.info(\n 'Number of files after removing duplicates from different vials: %d',\n len(df.index))\n df.drop('sample_barcode15', axis=1, inplace=True)\n df.sort_index(inplace=True)\n return df\n", "role": "user" }, { "content": "get_unique_sample_files", "role": "assistant" } ]
def x(biospecimen): df = biospecimen num_before = len(df.index) df = df.loc[~df['is_ffpe']] logger.info( 'Excluded %d files associated with FFPE samples (out of %d files in total).' , num_before - len(df.index), num_before) num_before = len(df.index) df = df.loc[df['sample_type'] == 'Primary Tumor'] logger.info( 'Excluded %d files not corresponding to primary tumor samples (out of %d files in total).' , num_before - len(df.index), num_before) return df
get_fresh_primary_tumors
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(biospecimen):\n df = biospecimen\n num_before = len(df.index)\n df = df.loc[~df['is_ffpe']]\n logger.info(\n 'Excluded %d files associated with FFPE samples (out of %d files in total).'\n , num_before - len(df.index), num_before)\n num_before = len(df.index)\n df = df.loc[df['sample_type'] == 'Primary Tumor']\n logger.info(\n 'Excluded %d files not corresponding to primary tumor samples (out of %d files in total).'\n , num_before - len(df.index), num_before)\n return df\n", "role": "user" }, { "content": "get_fresh_primary_tumors", "role": "assistant" } ]
def x(reads): seen = set() add = seen.add for read in reads: hash_ = md5(read.sequence.encode('UTF-8')).digest() if hash_ not in seen: add(hash_) yield read
dedupFasta
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(reads):\n seen = set()\n add = seen.add\n for read in reads:\n hash_ = md5(read.sequence.encode('UTF-8')).digest()\n if hash_ not in seen:\n add(hash_)\n yield read\n", "role": "user" }, { "content": "dedupFasta", "role": "assistant" } ]
def x(sequences): sequences = sorted(sequences, key=lambda s: len(s.seq), reverse=True) seen = set() for s in sequences: thisSeq = str(s.seq) thisHash = md5(thisSeq.encode('UTF-8')).digest() if thisHash not in seen: newHash = md5() for nucl in thisSeq: newHash.update(nucl.encode('UTF-8')) seen.add(newHash.digest()) for start in range(len(thisSeq) - 1): seen.add(md5(thisSeq[start + 1:].encode('UTF-8')).digest()) yield s
dePrefixAndSuffixFasta
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(sequences):\n sequences = sorted(sequences, key=lambda s: len(s.seq), reverse=True)\n seen = set()\n for s in sequences:\n thisSeq = str(s.seq)\n thisHash = md5(thisSeq.encode('UTF-8')).digest()\n if thisHash not in seen:\n newHash = md5()\n for nucl in thisSeq:\n newHash.update(nucl.encode('UTF-8'))\n seen.add(newHash.digest())\n for start in range(len(thisSeq) - 1):\n seen.add(md5(thisSeq[start + 1:].encode('UTF-8')).digest())\n yield s\n", "role": "user" }, { "content": "dePrefixAndSuffixFasta", "role": "assistant" } ]
def x(fastaFiles): reads = {} firstFile = fastaFiles.pop(0) for seq in SeqIO.parse(firstFile, 'fasta'): reads[seq.id] = seq for fastaFile in fastaFiles: for seq in SeqIO.parse(fastaFile, 'fasta'): if seq.id in reads: assert str(seq.seq) == str(reads[seq.id].seq) reads.pop(seq.id, None) return iter(reads.values())
fastaSubtract
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(fastaFiles):\n reads = {}\n firstFile = fastaFiles.pop(0)\n for seq in SeqIO.parse(firstFile, 'fasta'):\n reads[seq.id] = seq\n for fastaFile in fastaFiles:\n for seq in SeqIO.parse(fastaFile, 'fasta'):\n if seq.id in reads:\n assert str(seq.seq) == str(reads[seq.id].seq)\n reads.pop(seq.id, None)\n return iter(reads.values())\n", "role": "user" }, { "content": "fastaSubtract", "role": "assistant" } ]
def x(filename, sequences, readClass=DNARead, upperCase=False, idPrefix= 'command-line-read-'): if filename: reads = FastaReads(filename, readClass=readClass, upperCase=upperCase) else: reads = Reads() if sequences: for count, sequence in enumerate(sequences, start=1): parts = sequence.rsplit(' ', 1) if len(parts) == 2: readId, sequence = parts else: readId = '%s%d' % (idPrefix, count) if upperCase: sequence = sequence.upper() read = readClass(readId, sequence) reads.add(read) return reads
combineReads
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(filename, sequences, readClass=DNARead, upperCase=False, idPrefix=\n 'command-line-read-'):\n if filename:\n reads = FastaReads(filename, readClass=readClass, upperCase=upperCase)\n else:\n reads = Reads()\n if sequences:\n for count, sequence in enumerate(sequences, start=1):\n parts = sequence.rsplit(' ', 1)\n if len(parts) == 2:\n readId, sequence = parts\n else:\n readId = '%s%d' % (idPrefix, count)\n if upperCase:\n sequence = sequence.upper()\n read = readClass(readId, sequence)\n reads.add(read)\n return reads\n", "role": "user" }, { "content": "combineReads", "role": "assistant" } ]
def x(self): count = 0 for _file in self._files: with asHandle(_file) as fp: if self._upperCase: for seq in SeqIO.parse(fp, 'fasta'): read = self._readClass(seq.description, str(seq.seq. upper())) yield read count += 1 else: for seq in SeqIO.parse(fp, 'fasta'): read = self._readClass(seq.description, str(seq.seq)) yield read count += 1
iter
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n count = 0\n for _file in self._files:\n with asHandle(_file) as fp:\n if self._upperCase:\n for seq in SeqIO.parse(fp, 'fasta'):\n read = self._readClass(seq.description, str(seq.seq.\n upper()))\n yield read\n count += 1\n else:\n for seq in SeqIO.parse(fp, 'fasta'):\n read = self._readClass(seq.description, str(seq.seq))\n yield read\n count += 1\n", "role": "user" }, { "content": "iter", "role": "assistant" } ]
def x(self): if self._upperCase: for id_ in self._fasta: yield self._readClass(id_, str(self._fasta[id_]).upper()) else: for id_ in self._fasta: yield self._readClass(id_, str(self._fasta[id_]))
iter
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self._upperCase:\n for id_ in self._fasta:\n yield self._readClass(id_, str(self._fasta[id_]).upper())\n else:\n for id_ in self._fasta:\n yield self._readClass(id_, str(self._fasta[id_]))\n", "role": "user" }, { "content": "iter", "role": "assistant" } ]
def x(self, fileNumber): cur = self._connection.cursor() cur.execute('SELECT name FROM files WHERE id = ?', (fileNumber,)) row = cur.fetchone() if row is None: return None else: return row[0]
getFilename
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, fileNumber):\n cur = self._connection.cursor()\n cur.execute('SELECT name FROM files WHERE id = ?', (fileNumber,))\n row = cur.fetchone()\n if row is None:\n return None\n else:\n return row[0]\n", "role": "user" }, { "content": "getFilename", "role": "assistant" } ]
def x(self, filename): cur = self._connection.cursor() cur.execute('SELECT id FROM files WHERE name = ?', (filename,)) row = cur.fetchone() if row is None: return None else: return row[0]
getFileNumber
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, filename):\n cur = self._connection.cursor()\n cur.execute('SELECT id FROM files WHERE name = ?', (filename,))\n row = cur.fetchone()\n if row is None:\n return None\n else:\n return row[0]\n", "role": "user" }, { "content": "getFileNumber", "role": "assistant" } ]
def x(self, filename): cur = self._connection.cursor() try: cur.execute('INSERT INTO files(name) VALUES (?)', (filename,)) except sqlite3.IntegrityError as e: if str(e).find('UNIQUE constraint failed') > -1: raise ValueError('Duplicate file name: %r' % filename) else: raise else: fileNumber = cur.lastrowid self._connection.commit() return fileNumber
addFilename
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, filename):\n cur = self._connection.cursor()\n try:\n cur.execute('INSERT INTO files(name) VALUES (?)', (filename,))\n except sqlite3.IntegrityError as e:\n if str(e).find('UNIQUE constraint failed') > -1:\n raise ValueError('Duplicate file name: %r' % filename)\n else:\n raise\n else:\n fileNumber = cur.lastrowid\n self._connection.commit()\n return fileNumber\n", "role": "user" }, { "content": "addFilename", "role": "assistant" } ]
def x(self, filename): endswith = filename.lower().endswith if endswith('.bgz') or endswith('.gz'): useBgzf = True elif endswith('.bz2'): raise ValueError( 'Compressed FASTA is only supported in BGZF format. Use bgzip to compresss your FASTA.' ) else: useBgzf = False fileNumber = self._addFilename(filename) connection = self._connection count = 0 try: with connection: if useBgzf: try: fp = bgzf.open(filename, 'rb') except ValueError as e: if str(e).find('BGZF') > -1: raise ValueError( 'Compressed FASTA is only supported in BGZF format. Use the samtools bgzip utility (instead of gzip) to compresss your FASTA.' ) else: raise else: try: for line in fp: if line[0] == '>': count += 1 id_ = line[1:].rstrip(' \t\n\r') connection.execute( 'INSERT INTO sequences(id, fileNumber, offset) VALUES (?, ?, ?)' , (id_, fileNumber, fp.tell())) finally: fp.close() else: with open(filename) as fp: offset = 0 for line in fp: offset += len(line) if line[0] == '>': count += 1 id_ = line[1:].rstrip(' \t\n\r') connection.execute( 'INSERT INTO sequences(id, fileNumber, offset) VALUES (?, ?, ?)' , (id_, fileNumber, offset)) except sqlite3.IntegrityError as e: if str(e).find('UNIQUE constraint failed') > -1: original = self._find(id_) if original is None: raise ValueError( "FASTA sequence id '%s' found twice in file '%s'." % ( id_, filename)) else: origFilename, _ = original raise ValueError( "FASTA sequence id '%s', found in file '%s', was previously added from file '%s'." % (id_, filename, origFilename)) else: raise else: return count
addFile
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, filename):\n endswith = filename.lower().endswith\n if endswith('.bgz') or endswith('.gz'):\n useBgzf = True\n elif endswith('.bz2'):\n raise ValueError(\n 'Compressed FASTA is only supported in BGZF format. Use bgzip to compresss your FASTA.'\n )\n else:\n useBgzf = False\n fileNumber = self._addFilename(filename)\n connection = self._connection\n count = 0\n try:\n with connection:\n if useBgzf:\n try:\n fp = bgzf.open(filename, 'rb')\n except ValueError as e:\n if str(e).find('BGZF') > -1:\n raise ValueError(\n 'Compressed FASTA is only supported in BGZF format. Use the samtools bgzip utility (instead of gzip) to compresss your FASTA.'\n )\n else:\n raise\n else:\n try:\n for line in fp:\n if line[0] == '>':\n count += 1\n id_ = line[1:].rstrip(' \\t\\n\\r')\n connection.execute(\n 'INSERT INTO sequences(id, fileNumber, offset) VALUES (?, ?, ?)'\n , (id_, fileNumber, fp.tell()))\n finally:\n fp.close()\n else:\n with open(filename) as fp:\n offset = 0\n for line in fp:\n offset += len(line)\n if line[0] == '>':\n count += 1\n id_ = line[1:].rstrip(' \\t\\n\\r')\n connection.execute(\n 'INSERT INTO sequences(id, fileNumber, offset) VALUES (?, ?, ?)'\n , (id_, fileNumber, offset))\n except sqlite3.IntegrityError as e:\n if str(e).find('UNIQUE constraint failed') > -1:\n original = self._find(id_)\n if original is None:\n raise ValueError(\n \"FASTA sequence id '%s' found twice in file '%s'.\" % (\n id_, filename))\n else:\n origFilename, _ = original\n raise ValueError(\n \"FASTA sequence id '%s', found in file '%s', was previously added from file '%s'.\"\n % (id_, filename, origFilename))\n else:\n raise\n else:\n return count\n", "role": "user" }, { "content": "addFile", "role": "assistant" } ]