source_code
stringlengths
45
7.24k
function_name
stringlengths
1
111
messages
list
def x(key, project): cred = get_storage_credentials(key) return storage.Client(project=project, credentials=cred)
get_client
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(key, project):\n cred = get_storage_credentials(key)\n return storage.Client(project=project, credentials=cred)\n", "role": "user" }, { "content": "get_client", "role": "assistant" } ]
def x(client, bucket, prefix=''): bucket = client.get_bucket(bucket) files = list(bucket.list_blobs(prefix=prefix)) return files
get_files
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(client, bucket, prefix=''):\n bucket = client.get_bucket(bucket)\n files = list(bucket.list_blobs(prefix=prefix))\n return files\n", "role": "user" }, { "content": "get_files", "role": "assistant" } ]
def x(client, bucket, remote_path, local_path, overwrite=False): bucket = client.get_bucket(bucket) if not overwrite and os.path.isfile(local_path): raise OSError('File already exists!') with open(local_path, 'wb') as ofh: bucket.get_blob(remote_path).download_to_file(ofh)
download_file
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(client, bucket, remote_path, local_path, overwrite=False):\n bucket = client.get_bucket(bucket)\n if not overwrite and os.path.isfile(local_path):\n raise OSError('File already exists!')\n with open(local_path, 'wb') as ofh:\n bucket.get_blob(remote_path).download_to_file(ofh)\n", "role": "user" }, { "content": "download_file", "role": "assistant" } ]
def x(client, bucket, remote_path): bucket = client.get_bucket(bucket) bucket.delete_blob(remote_path)
delete_file
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(client, bucket, remote_path):\n bucket = client.get_bucket(bucket)\n bucket.delete_blob(remote_path)\n", "role": "user" }, { "content": "delete_file", "role": "assistant" } ]
def x(client, bucket, local_path, remote_path, overwrite=False): bucket = client.get_bucket(bucket) blob = storage.Blob(remote_path, bucket) if not overwrite and blob.exists(): raise Conflict('File/object already exists on the bucket!') blob.upload_from_filename(local_path)
upload_file
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(client, bucket, local_path, remote_path, overwrite=False):\n bucket = client.get_bucket(bucket)\n blob = storage.Blob(remote_path, bucket)\n if not overwrite and blob.exists():\n raise Conflict('File/object already exists on the bucket!')\n blob.upload_from_filename(local_path)\n", "role": "user" }, { "content": "upload_file", "role": "assistant" } ]
def x(self, readAlignments): if self.limit is not None and self.count == self.limit: return False if self.maxAlignmentsPerRead is not None and len(readAlignments ) > self.maxAlignmentsPerRead: return False if self.readIdRegex and self.readIdRegex.search(readAlignments.read.id ) is None: return False if self.titleFilter: wantedAlignments = [] for alignment in readAlignments: if self.titleFilter.accept(alignment.subjectTitle ) != TitleFilter.REJECT: wantedAlignments.append(alignment) if wantedAlignments: readAlignments[:] = wantedAlignments else: return False minSequenceLen = self.minSequenceLen maxSequenceLen = self.maxSequenceLen if minSequenceLen is not None or maxSequenceLen is not None: wantedAlignments = [] for alignment in readAlignments: length = alignment.subjectLength if not (minSequenceLen is not None and length < minSequenceLen or maxSequenceLen is not None and length > self.maxSequenceLen): wantedAlignments.append(alignment) if wantedAlignments: readAlignments[:] = wantedAlignments else: return False if self.taxonomy is not None: wantedAlignments = [] for alignment in readAlignments: lineage = self.lineageFetcher.lineage(alignment.subjectTitle) if lineage: for taxonomyIdAndScientificName in lineage: if self.taxonomy in taxonomyIdAndScientificName: wantedAlignments.append(alignment) else: wantedAlignments.append(alignment) if wantedAlignments: readAlignments[:] = wantedAlignments else: return False if self.oneAlignmentPerRead and readAlignments: readAlignments[:] = [bestAlignment(readAlignments)] if self.maxHspsPerHit is not None: for alignment in readAlignments: hsps = alignment.hsps if len(hsps) > self.maxHspsPerHit: alignment.hsps = hsps[:self.maxHspsPerHit] if self.scoreCutoff is not None: wantedAlignments = [] for alignment in readAlignments: hsps = alignment.hsps wantedHsps = [] for hsp in hsps: if hsp.betterThan(self.scoreCutoff): wantedHsps.append(hsp) if wantedHsps: alignment.hsps = wantedHsps wantedAlignments.append(alignment) if wantedAlignments: readAlignments[:] = wantedAlignments else: return False minStart = self.minStart maxStop = self.maxStop if minStart is not None or maxStop is not None: wantedAlignments = [] for alignment in readAlignments: hsps = alignment.hsps wantedHsps = [] for hsp in hsps: if not (minStart is not None and hsp.readStartInSubject < minStart or maxStop is not None and hsp. readEndInSubject > maxStop): wantedHsps.append(hsp) if wantedHsps: alignment.hsps = wantedHsps wantedAlignments.append(alignment) if wantedAlignments: readAlignments[:] = wantedAlignments else: return False self.count += 1 return readAlignments
filter
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, readAlignments):\n if self.limit is not None and self.count == self.limit:\n return False\n if self.maxAlignmentsPerRead is not None and len(readAlignments\n ) > self.maxAlignmentsPerRead:\n return False\n if self.readIdRegex and self.readIdRegex.search(readAlignments.read.id\n ) is None:\n return False\n if self.titleFilter:\n wantedAlignments = []\n for alignment in readAlignments:\n if self.titleFilter.accept(alignment.subjectTitle\n ) != TitleFilter.REJECT:\n wantedAlignments.append(alignment)\n if wantedAlignments:\n readAlignments[:] = wantedAlignments\n else:\n return False\n minSequenceLen = self.minSequenceLen\n maxSequenceLen = self.maxSequenceLen\n if minSequenceLen is not None or maxSequenceLen is not None:\n wantedAlignments = []\n for alignment in readAlignments:\n length = alignment.subjectLength\n if not (minSequenceLen is not None and length < minSequenceLen or\n maxSequenceLen is not None and length > self.maxSequenceLen):\n wantedAlignments.append(alignment)\n if wantedAlignments:\n readAlignments[:] = wantedAlignments\n else:\n return False\n if self.taxonomy is not None:\n wantedAlignments = []\n for alignment in readAlignments:\n lineage = self.lineageFetcher.lineage(alignment.subjectTitle)\n if lineage:\n for taxonomyIdAndScientificName in lineage:\n if self.taxonomy in taxonomyIdAndScientificName:\n wantedAlignments.append(alignment)\n else:\n wantedAlignments.append(alignment)\n if wantedAlignments:\n readAlignments[:] = wantedAlignments\n else:\n return False\n if self.oneAlignmentPerRead and readAlignments:\n readAlignments[:] = [bestAlignment(readAlignments)]\n if self.maxHspsPerHit is not None:\n for alignment in readAlignments:\n hsps = alignment.hsps\n if len(hsps) > self.maxHspsPerHit:\n alignment.hsps = hsps[:self.maxHspsPerHit]\n if self.scoreCutoff is not None:\n wantedAlignments = []\n for alignment in readAlignments:\n hsps = alignment.hsps\n wantedHsps = []\n for hsp in hsps:\n if hsp.betterThan(self.scoreCutoff):\n wantedHsps.append(hsp)\n if wantedHsps:\n alignment.hsps = wantedHsps\n wantedAlignments.append(alignment)\n if wantedAlignments:\n readAlignments[:] = wantedAlignments\n else:\n return False\n minStart = self.minStart\n maxStop = self.maxStop\n if minStart is not None or maxStop is not None:\n wantedAlignments = []\n for alignment in readAlignments:\n hsps = alignment.hsps\n wantedHsps = []\n for hsp in hsps:\n if not (minStart is not None and hsp.readStartInSubject <\n minStart or maxStop is not None and hsp.\n readEndInSubject > maxStop):\n wantedHsps.append(hsp)\n if wantedHsps:\n alignment.hsps = wantedHsps\n wantedAlignments.append(alignment)\n if wantedAlignments:\n readAlignments[:] = wantedAlignments\n else:\n return False\n self.count += 1\n return readAlignments\n", "role": "user" }, { "content": "filter", "role": "assistant" } ]
def x(self): for readAlignments in self: for alignment in readAlignments: for hsp in alignment.hsps: yield hsp
hsps
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n for readAlignments in self:\n for alignment in readAlignments:\n for hsp in alignment.hsps:\n yield hsp\n", "role": "user" }, { "content": "hsps", "role": "assistant" } ]
def x(self, **kwargs): self._filters.append(ReadsAlignmentsFilter(**kwargs).filter) return self
filter
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, **kwargs):\n self._filters.append(ReadsAlignmentsFilter(**kwargs).filter)\n return self\n", "role": "user" }, { "content": "filter", "role": "assistant" } ]
def x(species, release=None, ftp=None): assert isinstance(species, (str, _oldstr)) or isinstance(species, Iterable) if release is not None: assert isinstance(release, int) if ftp is not None: assert isinstance(ftp, ftplib.FTP) close_connection = False ftp_server = 'ftp.ensembl.org' ftp_user = 'anonymous' if ftp is None: ftp = ftplib.FTP(ftp_server) ftp.login(ftp_user) close_connection = True if release is None: release = util.get_latest_release(ftp=ftp) species_data = OrderedDict() if isinstance(species, (str, _oldstr)): species_list = [species] else: species_list = species for spec in species_list: species_dir = '/pub/release-%d/gtf/%s' % (release, spec.lower()) data = [] ftp.dir(species_dir, data.append) gtf_file = [] for d in data: i = d.rindex(' ') fn = d[i + 1:] if fn.endswith('.%d.gtf.gz' % release): gtf_file.append(fn) assert len(gtf_file) == 1 gtf_file = gtf_file[0] _LOGGER.debug('GTF file: %s', gtf_file) checksum_url = '/'.join([species_dir, 'CHECKSUMS']) file_checksums = util.get_file_checksums(checksum_url, ftp=ftp) gtf_checksum = file_checksums[gtf_file] _LOGGER.debug('GTF file checksum: %d', gtf_checksum) gtf_url = 'ftp://%s%s/%s' % (ftp_server, species_dir, gtf_file) species_data[spec] = gtf_url, gtf_checksum if close_connection: ftp.close() return species_data
get_annotation_urls_and_checksums
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(species, release=None, ftp=None):\n assert isinstance(species, (str, _oldstr)) or isinstance(species, Iterable)\n if release is not None:\n assert isinstance(release, int)\n if ftp is not None:\n assert isinstance(ftp, ftplib.FTP)\n close_connection = False\n ftp_server = 'ftp.ensembl.org'\n ftp_user = 'anonymous'\n if ftp is None:\n ftp = ftplib.FTP(ftp_server)\n ftp.login(ftp_user)\n close_connection = True\n if release is None:\n release = util.get_latest_release(ftp=ftp)\n species_data = OrderedDict()\n if isinstance(species, (str, _oldstr)):\n species_list = [species]\n else:\n species_list = species\n for spec in species_list:\n species_dir = '/pub/release-%d/gtf/%s' % (release, spec.lower())\n data = []\n ftp.dir(species_dir, data.append)\n gtf_file = []\n for d in data:\n i = d.rindex(' ')\n fn = d[i + 1:]\n if fn.endswith('.%d.gtf.gz' % release):\n gtf_file.append(fn)\n assert len(gtf_file) == 1\n gtf_file = gtf_file[0]\n _LOGGER.debug('GTF file: %s', gtf_file)\n checksum_url = '/'.join([species_dir, 'CHECKSUMS'])\n file_checksums = util.get_file_checksums(checksum_url, ftp=ftp)\n gtf_checksum = file_checksums[gtf_file]\n _LOGGER.debug('GTF file checksum: %d', gtf_checksum)\n gtf_url = 'ftp://%s%s/%s' % (ftp_server, species_dir, gtf_file)\n species_data[spec] = gtf_url, gtf_checksum\n if close_connection:\n ftp.close()\n return species_data\n", "role": "user" }, { "content": "get_annotation_urls_and_checksums", "role": "assistant" } ]
def x(path_or_buffer, include_polymorphic_pseudogenes=True, remove_duplicates=True, **kwargs): valid_biotypes = set(['protein_coding']) if include_polymorphic_pseudogenes: valid_biotypes.add('polymorphic_pseudogene') df = get_genes(path_or_buffer, valid_biotypes, remove_duplicates= remove_duplicates, **kwargs) return df
get_protein_coding_genes
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(path_or_buffer, include_polymorphic_pseudogenes=True,\n remove_duplicates=True, **kwargs):\n valid_biotypes = set(['protein_coding'])\n if include_polymorphic_pseudogenes:\n valid_biotypes.add('polymorphic_pseudogene')\n df = get_genes(path_or_buffer, valid_biotypes, remove_duplicates=\n remove_duplicates, **kwargs)\n return df\n", "role": "user" }, { "content": "get_protein_coding_genes", "role": "assistant" } ]
def x(path_or_buffer, remove_duplicates=True, **kwargs): valid_biotypes = set(['lincRNA']) df = get_genes(path_or_buffer, valid_biotypes, remove_duplicates= remove_duplicates, **kwargs) return df
get_linc_rna_genes
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(path_or_buffer, remove_duplicates=True, **kwargs):\n valid_biotypes = set(['lincRNA'])\n df = get_genes(path_or_buffer, valid_biotypes, remove_duplicates=\n remove_duplicates, **kwargs)\n return df\n", "role": "user" }, { "content": "get_linc_rna_genes", "role": "assistant" } ]
def x(levels=0, dimensions=None, sequential=True, string=True, max_int=255, max_pkt_len=150, max_dim_size=3): MAX_INT = max_int if string: MAX_INT = min(255, max_int) MAX_DIM_SIZE = max_dim_size MAX_PKT_LEN = max_pkt_len def x(): next_i.i = (next_i.i + 1) % (MAX_INT + 1) return next_i.i next_i.i = 0 def x(length): if sequential: pld = [next_i() for _ in xrange(length)] else: pld = [random.randint(0, MAX_INT) for _ in xrange(length)] if string: pld = str(bytearray(pld)) return pld def x(level, pld): if level > 0: if pld == None: pld = random.randint(1, MAX_DIM_SIZE) if isinstance(pld, int): pld = pld * [None] if isinstance(pld, list): for i in range(len(pld)): pld[i] = next_level(level - 1, pld[i]) return pld else: raise TypeError('Expected None, int or list, got {}: {}'. format(type(pld), pld)) elif level == 0: if pld == None: pld = random.randint(1, MAX_PKT_LEN) if isinstance(pld, int): return payload(pld) else: raise TypeError('Expected None or int, got {}: {}'.format( type(pld), pld)) else: raise ValueError('Expected int>=0, got {}'.format(level)) pld = copy.deepcopy(dimensions) pld = next_level(levels, pld) return pld
payload_generator
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(levels=0, dimensions=None, sequential=True, string=True, max_int=255,\n max_pkt_len=150, max_dim_size=3):\n MAX_INT = max_int\n if string:\n MAX_INT = min(255, max_int)\n MAX_DIM_SIZE = max_dim_size\n MAX_PKT_LEN = max_pkt_len\n\n def x():\n next_i.i = (next_i.i + 1) % (MAX_INT + 1)\n return next_i.i\n next_i.i = 0\n\n def x(length):\n if sequential:\n pld = [next_i() for _ in xrange(length)]\n else:\n pld = [random.randint(0, MAX_INT) for _ in xrange(length)]\n if string:\n pld = str(bytearray(pld))\n return pld\n\n def x(level, pld):\n if level > 0:\n if pld == None:\n pld = random.randint(1, MAX_DIM_SIZE)\n if isinstance(pld, int):\n pld = pld * [None]\n if isinstance(pld, list):\n for i in range(len(pld)):\n pld[i] = next_level(level - 1, pld[i])\n return pld\n else:\n raise TypeError('Expected None, int or list, got {}: {}'.\n format(type(pld), pld))\n elif level == 0:\n if pld == None:\n pld = random.randint(1, MAX_PKT_LEN)\n if isinstance(pld, int):\n return payload(pld)\n else:\n raise TypeError('Expected None or int, got {}: {}'.format(\n type(pld), pld))\n else:\n raise ValueError('Expected int>=0, got {}'.format(level))\n pld = copy.deepcopy(dimensions)\n pld = next_level(levels, pld)\n return pld\n", "role": "user" }, { "content": "payload_generator", "role": "assistant" } ]
def x(title, db='nucleotide'): titleId = title.split(' ', 1)[0] try: gi = titleId.split('|')[1] except IndexError: gi = str(titleId) try: client = Entrez.efetch(db=db, rettype='gb', retmode='text', id=gi) except URLError: return None else: record = SeqIO.read(client, 'gb') client.close() return record
getSequence
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(title, db='nucleotide'):\n titleId = title.split(' ', 1)[0]\n try:\n gi = titleId.split('|')[1]\n except IndexError:\n gi = str(titleId)\n try:\n client = Entrez.efetch(db=db, rettype='gb', retmode='text', id=gi)\n except URLError:\n return None\n else:\n record = SeqIO.read(client, 'gb')\n client.close()\n return record\n", "role": "user" }, { "content": "getSequence", "role": "assistant" } ]
def x(self): lowest_layer = None for layer in self.cloud_layers: if layer.coverage not in [CloudLayer.BROKEN, CloudLayer.OVERCAST]: continue if lowest_layer is None: lowest_layer = layer continue if layer.height > lowest_layer.height: continue if (layer.height < lowest_layer.height or lowest_layer. get_coverage_percentage() < layer.get_coverage_percentage()): lowest_layer = layer return lowest_layer
get_ceiling_cloud_layer
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n lowest_layer = None\n for layer in self.cloud_layers:\n if layer.coverage not in [CloudLayer.BROKEN, CloudLayer.OVERCAST]:\n continue\n if lowest_layer is None:\n lowest_layer = layer\n continue\n if layer.height > lowest_layer.height:\n continue\n if (layer.height < lowest_layer.height or lowest_layer.\n get_coverage_percentage() < layer.get_coverage_percentage()):\n lowest_layer = layer\n return lowest_layer\n", "role": "user" }, { "content": "get_ceiling_cloud_layer", "role": "assistant" } ]
def x(self): self.raw_text = self.xml_data.find('raw_text').text self.station = WeatherStation(self.xml_data.find('station_id').text) self.station.latitude = float(self.xml_data.find('latitude').text) self.station.longitude = float(self.xml_data.find('longitude').text) self.station.elevation = float(self.xml_data.find('elevation_m').text ) * 3.28084
parse_xml_data
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n self.raw_text = self.xml_data.find('raw_text').text\n self.station = WeatherStation(self.xml_data.find('station_id').text)\n self.station.latitude = float(self.xml_data.find('latitude').text)\n self.station.longitude = float(self.xml_data.find('longitude').text)\n self.station.elevation = float(self.xml_data.find('elevation_m').text\n ) * 3.28084\n", "role": "user" }, { "content": "parse_xml_data", "role": "assistant" } ]
def x(self, mock_response=None): if mock_response is not None: body = mock_response else: api_url = self.get_api_url() body = urlopen(api_url).read() xml_root = ElementTree.fromstring(body) xml_warnings = xml_root.find('warnings') if len(xml_warnings.attrib) != 0: print('Data warnings found: %s' % xml_warnings.attrib) xml_errors = xml_root.find('errors') if len(xml_errors.attrib) != 0: raise Exception('Data errors found: %s' % xml_errors.attrib) self.xml_data = xml_root.find('data')
download_data
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, mock_response=None):\n if mock_response is not None:\n body = mock_response\n else:\n api_url = self.get_api_url()\n body = urlopen(api_url).read()\n xml_root = ElementTree.fromstring(body)\n xml_warnings = xml_root.find('warnings')\n if len(xml_warnings.attrib) != 0:\n print('Data warnings found: %s' % xml_warnings.attrib)\n xml_errors = xml_root.find('errors')\n if len(xml_errors.attrib) != 0:\n raise Exception('Data errors found: %s' % xml_errors.attrib)\n self.xml_data = xml_root.find('data')\n", "role": "user" }, { "content": "download_data", "role": "assistant" } ]
def x(file_handle, file_type): base_counts = defaultdict(int) read_number = 0 total_length = 0 length_list = [] records = SeqIO.parse(file_handle, file_type) for record in records: total_length += len(record) read_number += 1 length_list.append(len(record)) for base in record: base_counts[base] += 1 result = {'read_number': read_number, 'total_length': total_length, 'average_length': total_length / read_number if read_number > 0 else 0, 'max_length': max(length_list) if length_list else 0, 'min_length': min(length_list) if length_list else 0, 'median_length': median(length_list) if length_list else 0, 'base_counts': base_counts} return result
summarizeReads
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(file_handle, file_type):\n base_counts = defaultdict(int)\n read_number = 0\n total_length = 0\n length_list = []\n records = SeqIO.parse(file_handle, file_type)\n for record in records:\n total_length += len(record)\n read_number += 1\n length_list.append(len(record))\n for base in record:\n base_counts[base] += 1\n result = {'read_number': read_number, 'total_length': total_length,\n 'average_length': total_length / read_number if read_number > 0 else\n 0, 'max_length': max(length_list) if length_list else 0,\n 'min_length': min(length_list) if length_list else 0,\n 'median_length': median(length_list) if length_list else 0,\n 'base_counts': base_counts}\n return result\n", "role": "user" }, { "content": "summarizeReads", "role": "assistant" } ]
def x(read, categories, defaultCategory=None, suppressedCategory='...', minLength=1): result = [] append = result.append get = categories.get first = True currentCategory = None currentCount = 0 suppressing = False suppressedCount = 0 if minLength < 1: raise ValueError('minLength must be at least 1') for base in read.sequence: thisCategory = get(base, defaultCategory) if first: first = False currentCategory = thisCategory currentCount += 1 elif thisCategory == currentCategory: currentCount += 1 else: if currentCount < minLength: if suppressing: suppressedCount += currentCount else: suppressedCount = currentCount suppressing = True else: if suppressing: append((suppressedCategory, suppressedCount)) suppressedCount = 0 suppressing = False append((currentCategory, currentCount)) currentCategory = thisCategory currentCount = 1 if suppressing: append((suppressedCategory, suppressedCount + currentCount)) elif currentCount >= minLength: append((currentCategory, currentCount)) elif currentCount: append((suppressedCategory, currentCount)) return result
sequenceCategoryLengths
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(read, categories, defaultCategory=None, suppressedCategory='...',\n minLength=1):\n result = []\n append = result.append\n get = categories.get\n first = True\n currentCategory = None\n currentCount = 0\n suppressing = False\n suppressedCount = 0\n if minLength < 1:\n raise ValueError('minLength must be at least 1')\n for base in read.sequence:\n thisCategory = get(base, defaultCategory)\n if first:\n first = False\n currentCategory = thisCategory\n currentCount += 1\n elif thisCategory == currentCategory:\n currentCount += 1\n else:\n if currentCount < minLength:\n if suppressing:\n suppressedCount += currentCount\n else:\n suppressedCount = currentCount\n suppressing = True\n else:\n if suppressing:\n append((suppressedCategory, suppressedCount))\n suppressedCount = 0\n suppressing = False\n append((currentCategory, currentCount))\n currentCategory = thisCategory\n currentCount = 1\n if suppressing:\n append((suppressedCategory, suppressedCount + currentCount))\n elif currentCount >= minLength:\n append((currentCategory, currentCount))\n elif currentCount:\n append((suppressedCategory, currentCount))\n return result\n", "role": "user" }, { "content": "sequenceCategoryLengths", "role": "assistant" } ]
def x(title, target): targetLen = len(target) result = [] for word in title.split(): if len(word) >= targetLen: offset = word.lower().find(target.lower()) if offset > -1: result.append(word[:offset + targetLen]) break result.append(word) return ' '.join(result)
simplifyTitle
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(title, target):\n targetLen = len(target)\n result = []\n for word in title.split():\n if len(word) >= targetLen:\n offset = word.lower().find(target.lower())\n if offset > -1:\n result.append(word[:offset + targetLen])\n break\n result.append(word)\n return ' '.join(result)\n", "role": "user" }, { "content": "simplifyTitle", "role": "assistant" } ]
def x(self, name=None, hdl=None, analyze_cmd=None, elaborate_cmd=None, simulate_cmd=None): if not isinstance(name, str) or name.strip() == '': raise ValueError('Invalid _simulator name') if hdl.lower() not in ('vhdl', 'verilog'): raise ValueError('Invalid hdl {}'.format(hdl)) if not isinstance(analyze_cmd, str) or analyze_cmd.strip() == '': raise ValueError('Invalid analyzer command') if elaborate_cmd is not None: if not isinstance(elaborate_cmd, str) or elaborate_cmd.strip() == '': raise ValueError('Invalid elaborate_cmd command') if not isinstance(simulate_cmd, str) or simulate_cmd.strip() == '': raise ValueError('Invalid _simulator command') self.sim_reg[name] = hdl.lower(), analyze_cmd, elaborate_cmd, simulate_cmd
registerSimulator
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, name=None, hdl=None, analyze_cmd=None, elaborate_cmd=None,\n simulate_cmd=None):\n if not isinstance(name, str) or name.strip() == '':\n raise ValueError('Invalid _simulator name')\n if hdl.lower() not in ('vhdl', 'verilog'):\n raise ValueError('Invalid hdl {}'.format(hdl))\n if not isinstance(analyze_cmd, str) or analyze_cmd.strip() == '':\n raise ValueError('Invalid analyzer command')\n if elaborate_cmd is not None:\n if not isinstance(elaborate_cmd, str) or elaborate_cmd.strip() == '':\n raise ValueError('Invalid elaborate_cmd command')\n if not isinstance(simulate_cmd, str) or simulate_cmd.strip() == '':\n raise ValueError('Invalid _simulator command')\n self.sim_reg[name] = hdl.lower(), analyze_cmd, elaborate_cmd, simulate_cmd\n", "role": "user" }, { "content": "registerSimulator", "role": "assistant" } ]
def x(self, func, **kwargs): vals = {} vals['topname'] = func.func_name vals['unitname'] = func.func_name.lower() hdlsim = self._simulator if not hdlsim: raise ValueError('No _simulator specified') if not self.sim_reg.has_key(hdlsim): raise ValueError('Simulator {} is not registered'.format(hdlsim)) hdl, analyze_cmd, elaborate_cmd, simulate_cmd = self.sim_reg[hdlsim] if hdl == 'verilog': toVerilog(func, **kwargs) if self._trace: self._enableTracesVerilog('./tb_{topname}.v'.format(**vals)) elif hdl == 'vhdl': toVHDL(func, **kwargs) os.system(analyze_cmd.format(**vals)) if elaborate_cmd: os.system(elaborate_cmd.format(**vals)) return Cosimulation(simulate_cmd.format(**vals), **kwargs)
getCosimulation
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, func, **kwargs):\n vals = {}\n vals['topname'] = func.func_name\n vals['unitname'] = func.func_name.lower()\n hdlsim = self._simulator\n if not hdlsim:\n raise ValueError('No _simulator specified')\n if not self.sim_reg.has_key(hdlsim):\n raise ValueError('Simulator {} is not registered'.format(hdlsim))\n hdl, analyze_cmd, elaborate_cmd, simulate_cmd = self.sim_reg[hdlsim]\n if hdl == 'verilog':\n toVerilog(func, **kwargs)\n if self._trace:\n self._enableTracesVerilog('./tb_{topname}.v'.format(**vals))\n elif hdl == 'vhdl':\n toVHDL(func, **kwargs)\n os.system(analyze_cmd.format(**vals))\n if elaborate_cmd:\n os.system(elaborate_cmd.format(**vals))\n return Cosimulation(simulate_cmd.format(**vals), **kwargs)\n", "role": "user" }, { "content": "getCosimulation", "role": "assistant" } ]
def x(self, verilogFile): fname, _ = os.path.splitext(verilogFile) inserted = False for _, line in enumerate(fileinput.input(verilogFile, inplace=1)): sys.stdout.write(line) if line.startswith('end') and not inserted: sys.stdout.write('\n\n') sys.stdout.write('initial begin\n') sys.stdout.write(' $dumpfile("{}_cosim.vcd");\n'.format(fname)) sys.stdout.write(' $dumpvars(0, dut);\n') sys.stdout.write('end\n\n') inserted = True
enableTracesVerilog
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, verilogFile):\n fname, _ = os.path.splitext(verilogFile)\n inserted = False\n for _, line in enumerate(fileinput.input(verilogFile, inplace=1)):\n sys.stdout.write(line)\n if line.startswith('end') and not inserted:\n sys.stdout.write('\\n\\n')\n sys.stdout.write('initial begin\\n')\n sys.stdout.write(' $dumpfile(\"{}_cosim.vcd\");\\n'.format(fname))\n sys.stdout.write(' $dumpvars(0, dut);\\n')\n sys.stdout.write('end\\n\\n')\n inserted = True\n", "role": "user" }, { "content": "enableTracesVerilog", "role": "assistant" } ]
def x(self, func, **kwargs): if self._simulator == 'myhdl': if not self._trace: sim_dut = func(**kwargs) else: sim_dut = traceSignals(func, **kwargs) else: sim_dut = self._getCosimulation(func, **kwargs) return sim_dut
getDut
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, func, **kwargs):\n if self._simulator == 'myhdl':\n if not self._trace:\n sim_dut = func(**kwargs)\n else:\n sim_dut = traceSignals(func, **kwargs)\n else:\n sim_dut = self._getCosimulation(func, **kwargs)\n return sim_dut\n", "role": "user" }, { "content": "getDut", "role": "assistant" } ]
def x(self): for _file in self._files: with asHandle(_file) as fp: for sequenceId, sequence, quality in FastqGeneralIterator(fp): yield self.readClass(sequenceId, sequence, quality)
iter
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n for _file in self._files:\n with asHandle(_file) as fp:\n for sequenceId, sequence, quality in FastqGeneralIterator(fp):\n yield self.readClass(sequenceId, sequence, quality)\n", "role": "user" }, { "content": "iter", "role": "assistant" } ]
def x(params): assert params['fastaFile'][-1 ] == 'a', 'You must specify a file in fasta-format that ends in .fasta' fileCount = count = seqCount = 0 outfp = None with open(params['fastaFile']) as infp: for seq in SeqIO.parse(infp, 'fasta'): seqCount += 1 if count == params['seqsPerJob']: outfp.close() count = 0 if count == 0: outfp = open('%d.fasta' % fileCount, 'w') fileCount += 1 count += 1 outfp.write('>%s\n%s\n' % (seq.description, str(seq.seq))) outfp.close() return fileCount, seqCount
splitFASTA
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(params):\n assert params['fastaFile'][-1\n ] == 'a', 'You must specify a file in fasta-format that ends in .fasta'\n fileCount = count = seqCount = 0\n outfp = None\n with open(params['fastaFile']) as infp:\n for seq in SeqIO.parse(infp, 'fasta'):\n seqCount += 1\n if count == params['seqsPerJob']:\n outfp.close()\n count = 0\n if count == 0:\n outfp = open('%d.fasta' % fileCount, 'w')\n fileCount += 1\n count += 1\n outfp.write('>%s\\n%s\\n' % (seq.description, str(seq.seq)))\n outfp.close()\n return fileCount, seqCount\n", "role": "user" }, { "content": "splitFASTA", "role": "assistant" } ]
def x(self, task): if not self._valid_name(task.name): raise ValueError(task.name) self._tasks[task.name] = task incomplete_dependencies = set() for dependency in task.dependencies: if not self._valid_name(dependency) or dependency in self._failed: self._cascade_failure(task.name) break if dependency not in self._completed: incomplete_dependencies.add(dependency) else: try: self._graph.add(task.name, incomplete_dependencies) except ValueError: self._cascade_failure(task.name)
add_task
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, task):\n if not self._valid_name(task.name):\n raise ValueError(task.name)\n self._tasks[task.name] = task\n incomplete_dependencies = set()\n for dependency in task.dependencies:\n if not self._valid_name(dependency) or dependency in self._failed:\n self._cascade_failure(task.name)\n break\n if dependency not in self._completed:\n incomplete_dependencies.add(dependency)\n else:\n try:\n self._graph.add(task.name, incomplete_dependencies)\n except ValueError:\n self._cascade_failure(task.name)\n", "role": "user" }, { "content": "add_task", "role": "assistant" } ]
def x(self, name=None): if name is None: for possibility in self._graph.roots: if possibility not in self._running: name = possibility break else: return None elif name not in self._graph.roots or name in self._running: raise ValueError(name) self._running.add(name) return self._tasks[name]
start_task
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, name=None):\n if name is None:\n for possibility in self._graph.roots:\n if possibility not in self._running:\n name = possibility\n break\n else:\n return None\n elif name not in self._graph.roots or name in self._running:\n raise ValueError(name)\n self._running.add(name)\n return self._tasks[name]\n", "role": "user" }, { "content": "start_task", "role": "assistant" } ]
def x(self, name, success=True): self._running.remove(name) if success: self._completed.add(name) self._graph.remove(name, strategy=Strategy.orphan) else: self._cascade_failure(name)
end_task
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, name, success=True):\n self._running.remove(name)\n if success:\n self._completed.add(name)\n self._graph.remove(name, strategy=Strategy.orphan)\n else:\n self._cascade_failure(name)\n", "role": "user" }, { "content": "end_task", "role": "assistant" } ]
def x(self): self._failed.update(self._graph.nodes) self._graph = Graph() self._running = set()
fail_remaining
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n self._failed.update(self._graph.nodes)\n self._graph = Graph()\n self._running = set()\n", "role": "user" }, { "content": "fail_remaining", "role": "assistant" } ]
def x(self, name): if name in self._graph: self._failed.update(self._graph.remove(name, strategy=Strategy.remove)) else: self._failed.add(name)
cascade_failure
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, name):\n if name in self._graph:\n self._failed.update(self._graph.remove(name, strategy=Strategy.remove))\n else:\n self._failed.add(name)\n", "role": "user" }, { "content": "cascade_failure", "role": "assistant" } ]
def x(gi, ranges): record = getSequence(gi) if record is None: print("Looks like you're offline.") sys.exit(3) else: printed = set() if ranges: for start, end in ranges: for index, feature in enumerate(record.features): if start < int(feature.location.end) and end > int(feature .location.start) and index not in printed: print(feature) printed.add(index) else: for feature in record.features: print(feature)
main
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(gi, ranges):\n record = getSequence(gi)\n if record is None:\n print(\"Looks like you're offline.\")\n sys.exit(3)\n else:\n printed = set()\n if ranges:\n for start, end in ranges:\n for index, feature in enumerate(record.features):\n if start < int(feature.location.end) and end > int(feature\n .location.start) and index not in printed:\n print(feature)\n printed.add(index)\n else:\n for feature in record.features:\n print(feature)\n", "role": "user" }, { "content": "main", "role": "assistant" } ]
def x(credentials, project, zone, name, startup_script=None, startup_script_url=None, metadata=None, machine_type='f1-micro', tags= None, disk_size_gb=10, wait_until_done=False): if startup_script is not None and startup_script_url is not None: raise ValueError( 'Cannot specify a startup script string and URL at the same time!') access_token = credentials.get_access_token() if metadata is None: metadata = {} meta_items = [{'key': k, 'value': v} for k, v in metadata.items()] if tags is None: tags = [] if startup_script is not None: meta_items.insert(0, {'key': 'startup-script', 'value': startup_script} ) elif startup_script_url is not None: meta_items.insert(0, {'key': 'startup-script-url', 'value': startup_script_url}) payload = {'name': name, 'zone': 'projects/%s/zones/%s' % (project, zone), 'machineType': 'projects/%s/zones/%s/machineTypes/%s' % ( project, zone, machine_type), 'metadata': {'items': meta_items}, 'tags': {'items': tags}, 'disks': [{'type': 'PERSISTENT', 'boot': True, 'mode': 'READ_WRITE', 'autoDelete': True, 'deviceName': name, 'initializeParams': {'sourceImage': 'projects/ubuntu-os-cloud/global/images/ubuntu-1604-xenial-v20170815a', 'diskType': 'projects/%s/zones/%s/diskTypes/pd-standard' % (project, zone), 'diskSizeGb': str(disk_size_gb)}}], 'canIpForward': False, 'networkInterfaces': [{'network': 'projects/%s/global/networks/default' % project, 'subnetwork': 'projects/%s/regions/%s/subnetworks/default' % (project, zone[:-2]), 'accessConfigs': [{'name': 'External NAT', 'type': 'ONE_TO_ONE_NAT' }]}], 'description': '', 'scheduling': {'preemptible': False, 'onHostMaintenance': 'MIGRATE', 'automaticRestart': True}, 'serviceAccounts': [{'email': 'default', 'scopes': [ 'https://www.googleapis.com/auth/compute', 'https://www.googleapis.com/auth/devstorage.read_write', 'https://www.googleapis.com/auth/logging.write', 'https://www.googleapis.com/auth/monitoring.write', 'https://www.googleapis.com/auth/servicecontrol', 'https://www.googleapis.com/auth/service.management.readonly', 'https://www.googleapis.com/auth/trace.append']}]} headers = {'Authorization': 'Bearer %s' % access_token.access_token} _LOGGER.debug('Access token: %s' % access_token.access_token) _LOGGER.debug('Payload: %s', json.dumps(payload, sort_keys=True, indent=4)) r = requests.post( 'https://www.googleapis.com/compute/v1/projects/%s/zones/%s/instances' % (project, zone), headers=headers, json=payload) r.raise_for_status() op_name = r.json()['name'] _LOGGER.info('Submitted request to create intsance (HTTP code: %d).', r .status_code) if wait_until_done: _LOGGER.info('Waiting until operation is done...') wait_for_zone_op(access_token, project, zone, op_name) return op_name
create_instance
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(credentials, project, zone, name, startup_script=None,\n startup_script_url=None, metadata=None, machine_type='f1-micro', tags=\n None, disk_size_gb=10, wait_until_done=False):\n if startup_script is not None and startup_script_url is not None:\n raise ValueError(\n 'Cannot specify a startup script string and URL at the same time!')\n access_token = credentials.get_access_token()\n if metadata is None:\n metadata = {}\n meta_items = [{'key': k, 'value': v} for k, v in metadata.items()]\n if tags is None:\n tags = []\n if startup_script is not None:\n meta_items.insert(0, {'key': 'startup-script', 'value': startup_script}\n )\n elif startup_script_url is not None:\n meta_items.insert(0, {'key': 'startup-script-url', 'value':\n startup_script_url})\n payload = {'name': name, 'zone': 'projects/%s/zones/%s' % (project,\n zone), 'machineType': 'projects/%s/zones/%s/machineTypes/%s' % (\n project, zone, machine_type), 'metadata': {'items': meta_items},\n 'tags': {'items': tags}, 'disks': [{'type': 'PERSISTENT', 'boot': \n True, 'mode': 'READ_WRITE', 'autoDelete': True, 'deviceName': name,\n 'initializeParams': {'sourceImage':\n 'projects/ubuntu-os-cloud/global/images/ubuntu-1604-xenial-v20170815a',\n 'diskType': 'projects/%s/zones/%s/diskTypes/pd-standard' % (project,\n zone), 'diskSizeGb': str(disk_size_gb)}}], 'canIpForward': False,\n 'networkInterfaces': [{'network': \n 'projects/%s/global/networks/default' % project, 'subnetwork': \n 'projects/%s/regions/%s/subnetworks/default' % (project, zone[:-2]),\n 'accessConfigs': [{'name': 'External NAT', 'type': 'ONE_TO_ONE_NAT'\n }]}], 'description': '', 'scheduling': {'preemptible': False,\n 'onHostMaintenance': 'MIGRATE', 'automaticRestart': True},\n 'serviceAccounts': [{'email': 'default', 'scopes': [\n 'https://www.googleapis.com/auth/compute',\n 'https://www.googleapis.com/auth/devstorage.read_write',\n 'https://www.googleapis.com/auth/logging.write',\n 'https://www.googleapis.com/auth/monitoring.write',\n 'https://www.googleapis.com/auth/servicecontrol',\n 'https://www.googleapis.com/auth/service.management.readonly',\n 'https://www.googleapis.com/auth/trace.append']}]}\n headers = {'Authorization': 'Bearer %s' % access_token.access_token}\n _LOGGER.debug('Access token: %s' % access_token.access_token)\n _LOGGER.debug('Payload: %s', json.dumps(payload, sort_keys=True, indent=4))\n r = requests.post(\n 'https://www.googleapis.com/compute/v1/projects/%s/zones/%s/instances'\n % (project, zone), headers=headers, json=payload)\n r.raise_for_status()\n op_name = r.json()['name']\n _LOGGER.info('Submitted request to create intsance (HTTP code: %d).', r\n .status_code)\n if wait_until_done:\n _LOGGER.info('Waiting until operation is done...')\n wait_for_zone_op(access_token, project, zone, op_name)\n return op_name\n", "role": "user" }, { "content": "create_instance", "role": "assistant" } ]
def x(credentials, project, zone, name, wait_until_done=False): access_token = credentials.get_access_token() headers = {'Authorization': 'Bearer %s' % access_token.access_token} r = requests.delete( 'https://www.googleapis.com/compute/v1/projects/%s/zones/%s/instances/%s' % (project, zone, name), headers=headers) r.raise_for_status() op_name = r.json()['name'] _LOGGER.info('Submitted request to create intsance (HTTP code: %d).', r .status_code) if wait_until_done: _LOGGER.info('Waiting until operation is done...') wait_for_zone_op(access_token, project, zone, op_name) return op_name
delete_instance
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(credentials, project, zone, name, wait_until_done=False):\n access_token = credentials.get_access_token()\n headers = {'Authorization': 'Bearer %s' % access_token.access_token}\n r = requests.delete(\n 'https://www.googleapis.com/compute/v1/projects/%s/zones/%s/instances/%s'\n % (project, zone, name), headers=headers)\n r.raise_for_status()\n op_name = r.json()['name']\n _LOGGER.info('Submitted request to create intsance (HTTP code: %d).', r\n .status_code)\n if wait_until_done:\n _LOGGER.info('Waiting until operation is done...')\n wait_for_zone_op(access_token, project, zone, op_name)\n return op_name\n", "role": "user" }, { "content": "delete_instance", "role": "assistant" } ]
def x(credentials, project, zone, instance_name, interval_seconds=5): t0 = time.time() access_token = credentials.get_access_token() headers = {'Authorization': 'Bearer %s' % access_token.access_token} r = requests.get( 'https://www.googleapis.com/compute/v1/projects/%s/zones/%s/instances/%s' % (project, zone, instance_name), headers=headers) if r.status_code == 404: raise AssertionError('Instance "%s" does not exist!' % instance_name) r.raise_for_status() _LOGGER.debug('Instance "%s" exists.', instance_name) while True: time.sleep(interval_seconds) access_token = credentials.get_access_token() headers = {'Authorization': 'Bearer %s' % access_token.access_token} r = requests.get( 'https://www.googleapis.com/compute/v1/projects/%s/zones/%s/instances/%s' % (project, zone, instance_name), headers=headers) if r.status_code == 404: break r.raise_for_status() _LOGGER.debug('Instance "%s" still exists.', instance_name) t1 = time.time() t = t1 - t0 t_min = t / 60.0 _LOGGER.info('Instance was deleted after %.1f s (%.1f m).', t, t_min)
wait_for_instance_deletion
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(credentials, project, zone, instance_name, interval_seconds=5):\n t0 = time.time()\n access_token = credentials.get_access_token()\n headers = {'Authorization': 'Bearer %s' % access_token.access_token}\n r = requests.get(\n 'https://www.googleapis.com/compute/v1/projects/%s/zones/%s/instances/%s'\n % (project, zone, instance_name), headers=headers)\n if r.status_code == 404:\n raise AssertionError('Instance \"%s\" does not exist!' % instance_name)\n r.raise_for_status()\n _LOGGER.debug('Instance \"%s\" exists.', instance_name)\n while True:\n time.sleep(interval_seconds)\n access_token = credentials.get_access_token()\n headers = {'Authorization': 'Bearer %s' % access_token.access_token}\n r = requests.get(\n 'https://www.googleapis.com/compute/v1/projects/%s/zones/%s/instances/%s'\n % (project, zone, instance_name), headers=headers)\n if r.status_code == 404:\n break\n r.raise_for_status()\n _LOGGER.debug('Instance \"%s\" still exists.', instance_name)\n t1 = time.time()\n t = t1 - t0\n t_min = t / 60.0\n _LOGGER.info('Instance was deleted after %.1f s (%.1f m).', t, t_min)\n", "role": "user" }, { "content": "wait_for_instance_deletion", "role": "assistant" } ]
def x(filenameOrSamfile, indent=0): indent = ' ' * indent def x(sam): result = [] for i in range(sam.nreferences): result.append('%s%s (length %d)' % (indent, sam. get_reference_name(i), sam.lengths[i])) return '\n'.join(result) if isinstance(filenameOrSamfile, six.string_types): with samfile(filenameOrSamfile) as sam: return _references(sam) else: return _references(sam)
samReferencesToStr
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(filenameOrSamfile, indent=0):\n indent = ' ' * indent\n\n def x(sam):\n result = []\n for i in range(sam.nreferences):\n result.append('%s%s (length %d)' % (indent, sam.\n get_reference_name(i), sam.lengths[i]))\n return '\\n'.join(result)\n if isinstance(filenameOrSamfile, six.string_types):\n with samfile(filenameOrSamfile) as sam:\n return _references(sam)\n else:\n return _references(sam)\n", "role": "user" }, { "content": "samReferencesToStr", "role": "assistant" } ]
def x(sequence, quality, cigartuples): hardClipCount = cigarLength = 0 for operation, length in cigartuples: hardClipCount += operation == CHARD_CLIP cigarLength += length if operation in _CONSUMES_QUERY else 0 sequenceLength = len(sequence) if quality is not None: assert sequenceLength == len(quality) clipLeft = clipRight = 0 clippedSequence = sequence clippedQuality = quality if sequenceLength > cigarLength: alreadyClipped = False else: assert sequenceLength == cigarLength alreadyClipped = True if hardClipCount == 0: pass elif hardClipCount == 1: if cigartuples[0][0] == CHARD_CLIP: if not alreadyClipped: clipLeft = cigartuples[0][1] clippedSequence = sequence[clipLeft:] if quality is not None: clippedQuality = quality[clipLeft:] elif cigartuples[-1][0] == CHARD_CLIP: if not alreadyClipped: clipRight = cigartuples[-1][1] clippedSequence = sequence[:-clipRight] if quality is not None: clippedQuality = quality[:-clipRight] else: raise ValueError( 'Invalid CIGAR tuples (%s) contains hard-clipping operation that is neither at the start nor the end of the sequence.' % (cigartuples,)) elif hardClipCount == 2: assert cigartuples[0][0] == cigartuples[-1][0] == CHARD_CLIP if not alreadyClipped: clipLeft, clipRight = cigartuples[0][1], cigartuples[-1][1] clippedSequence = sequence[clipLeft:-clipRight] if quality is not None: clippedQuality = quality[clipLeft:-clipRight] else: raise ValueError( 'Invalid CIGAR tuples (%s) specifies hard-clipping %d times (2 is the maximum).' % (cigartuples, hardClipCount)) weClipped = bool(clipLeft or clipRight) if weClipped: assert not alreadyClipped if len(clippedSequence) + clipLeft + clipRight != sequenceLength: raise ValueError( 'Sequence %r (length %d) clipped to %r (length %d), but the difference between these two lengths (%d) is not equal to the sum (%d) of the left and right clip lengths (%d and %d respectively). CIGAR tuples: %s' % (sequence, len(sequence), clippedSequence, len( clippedSequence), abs(len(sequence) - len(clippedSequence)), clipLeft + clipRight, clipLeft, clipRight, cigartuples)) else: assert len(clippedSequence) == sequenceLength if quality is not None: assert len(clippedQuality) == sequenceLength return clippedSequence, clippedQuality, weClipped
hardClip
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(sequence, quality, cigartuples):\n hardClipCount = cigarLength = 0\n for operation, length in cigartuples:\n hardClipCount += operation == CHARD_CLIP\n cigarLength += length if operation in _CONSUMES_QUERY else 0\n sequenceLength = len(sequence)\n if quality is not None:\n assert sequenceLength == len(quality)\n clipLeft = clipRight = 0\n clippedSequence = sequence\n clippedQuality = quality\n if sequenceLength > cigarLength:\n alreadyClipped = False\n else:\n assert sequenceLength == cigarLength\n alreadyClipped = True\n if hardClipCount == 0:\n pass\n elif hardClipCount == 1:\n if cigartuples[0][0] == CHARD_CLIP:\n if not alreadyClipped:\n clipLeft = cigartuples[0][1]\n clippedSequence = sequence[clipLeft:]\n if quality is not None:\n clippedQuality = quality[clipLeft:]\n elif cigartuples[-1][0] == CHARD_CLIP:\n if not alreadyClipped:\n clipRight = cigartuples[-1][1]\n clippedSequence = sequence[:-clipRight]\n if quality is not None:\n clippedQuality = quality[:-clipRight]\n else:\n raise ValueError(\n 'Invalid CIGAR tuples (%s) contains hard-clipping operation that is neither at the start nor the end of the sequence.'\n % (cigartuples,))\n elif hardClipCount == 2:\n assert cigartuples[0][0] == cigartuples[-1][0] == CHARD_CLIP\n if not alreadyClipped:\n clipLeft, clipRight = cigartuples[0][1], cigartuples[-1][1]\n clippedSequence = sequence[clipLeft:-clipRight]\n if quality is not None:\n clippedQuality = quality[clipLeft:-clipRight]\n else:\n raise ValueError(\n 'Invalid CIGAR tuples (%s) specifies hard-clipping %d times (2 is the maximum).'\n % (cigartuples, hardClipCount))\n weClipped = bool(clipLeft or clipRight)\n if weClipped:\n assert not alreadyClipped\n if len(clippedSequence) + clipLeft + clipRight != sequenceLength:\n raise ValueError(\n 'Sequence %r (length %d) clipped to %r (length %d), but the difference between these two lengths (%d) is not equal to the sum (%d) of the left and right clip lengths (%d and %d respectively). CIGAR tuples: %s'\n % (sequence, len(sequence), clippedSequence, len(\n clippedSequence), abs(len(sequence) - len(clippedSequence)),\n clipLeft + clipRight, clipLeft, clipRight, cigartuples))\n else:\n assert len(clippedSequence) == sequenceLength\n if quality is not None:\n assert len(clippedQuality) == sequenceLength\n return clippedSequence, clippedQuality, weClipped\n", "role": "user" }, { "content": "hardClip", "role": "assistant" } ]
def x(parser, samfileIsPositionalArg=False): parser.add_argument('%ssamfile' % ('' if samfileIsPositionalArg else '--'), required=True, help='The SAM/BAM file to filter.') parser.add_argument('--referenceId', metavar='ID', nargs='+', action= 'append', help= 'A reference sequence id whose alignments should be kept (alignments against other references will be dropped). If omitted, alignments against all references will be kept. May be repeated.' ) parser.add_argument('--dropUnmapped', default=False, action= 'store_true', help='If given, unmapped matches will not be output.') parser.add_argument('--dropSecondary', default=False, action= 'store_true', help='If given, secondary matches will not be output.') parser.add_argument('--dropSupplementary', default=False, action= 'store_true', help= 'If given, supplementary matches will not be output.') parser.add_argument('--dropDuplicates', default=False, action= 'store_true', help= 'If given, matches flagged as optical or PCR duplicates will not be output.' ) parser.add_argument('--keepQCFailures', default=False, action= 'store_true', help= 'If given, reads that are considered quality control failures will be included in the output.' ) parser.add_argument('--minScore', type=float, help= 'If given, alignments with --scoreTag (default AS) values less than this value will not be output. If given, alignments that do not have a score will not be output.' ) parser.add_argument('--maxScore', type=float, help= 'If given, alignments with --scoreTag (default AS) values greater than this value will not be output. If given, alignments that do not have a score will not be output.' ) parser.add_argument('--scoreTag', default='AS', help= 'The alignment tag to extract for --minScore and --maxScore comparisons.' )
addFilteringOptions
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(parser, samfileIsPositionalArg=False):\n parser.add_argument('%ssamfile' % ('' if samfileIsPositionalArg else\n '--'), required=True, help='The SAM/BAM file to filter.')\n parser.add_argument('--referenceId', metavar='ID', nargs='+', action=\n 'append', help=\n 'A reference sequence id whose alignments should be kept (alignments against other references will be dropped). If omitted, alignments against all references will be kept. May be repeated.'\n )\n parser.add_argument('--dropUnmapped', default=False, action=\n 'store_true', help='If given, unmapped matches will not be output.')\n parser.add_argument('--dropSecondary', default=False, action=\n 'store_true', help='If given, secondary matches will not be output.')\n parser.add_argument('--dropSupplementary', default=False, action=\n 'store_true', help=\n 'If given, supplementary matches will not be output.')\n parser.add_argument('--dropDuplicates', default=False, action=\n 'store_true', help=\n 'If given, matches flagged as optical or PCR duplicates will not be output.'\n )\n parser.add_argument('--keepQCFailures', default=False, action=\n 'store_true', help=\n 'If given, reads that are considered quality control failures will be included in the output.'\n )\n parser.add_argument('--minScore', type=float, help=\n 'If given, alignments with --scoreTag (default AS) values less than this value will not be output. If given, alignments that do not have a score will not be output.'\n )\n parser.add_argument('--maxScore', type=float, help=\n 'If given, alignments with --scoreTag (default AS) values greater than this value will not be output. If given, alignments that do not have a score will not be output.'\n )\n parser.add_argument('--scoreTag', default='AS', help=\n 'The alignment tag to extract for --minScore and --maxScore comparisons.'\n )\n", "role": "user" }, { "content": "addFilteringOptions", "role": "assistant" } ]
def x(cls, args, filterRead=None, storeQueryIds=False): referenceIds = set(chain.from_iterable(args.referenceId) ) if args.referenceId else None return cls(args.samfile, filterRead=filterRead, referenceIds= referenceIds, storeQueryIds=storeQueryIds, dropUnmapped=args. dropUnmapped, dropSecondary=args.dropSecondary, dropSupplementary= args.dropSupplementary, dropDuplicates=args.dropDuplicates, keepQCFailures=args.keepQCFailures, minScore=args.minScore, maxScore=args.maxScore)
parseFilteringOptions
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, args, filterRead=None, storeQueryIds=False):\n referenceIds = set(chain.from_iterable(args.referenceId)\n ) if args.referenceId else None\n return cls(args.samfile, filterRead=filterRead, referenceIds=\n referenceIds, storeQueryIds=storeQueryIds, dropUnmapped=args.\n dropUnmapped, dropSecondary=args.dropSecondary, dropSupplementary=\n args.dropSupplementary, dropDuplicates=args.dropDuplicates,\n keepQCFailures=args.keepQCFailures, minScore=args.minScore,\n maxScore=args.maxScore)\n", "role": "user" }, { "content": "parseFilteringOptions", "role": "assistant" } ]
def x(self): referenceIds = self.referenceIds dropUnmapped = self.dropUnmapped dropSecondary = self.dropSecondary dropSupplementary = self.dropSupplementary dropDuplicates = self.dropDuplicates keepQCFailures = self.keepQCFailures storeQueryIds = self.storeQueryIds filterRead = self.filterRead minScore = self.minScore maxScore = self.maxScore scoreTag = self.scoreTag if storeQueryIds: self.queryIds = queryIds = set() lastAlignment = None count = 0 with samfile(self.filename) as samAlignment: for count, alignment in enumerate(samAlignment.fetch(), start=1): if storeQueryIds: queryIds.add(alignment.query_name) if minScore is not None or maxScore is not None: try: score = alignment.get_tag(scoreTag) except KeyError: continue else: if (minScore is not None and score < minScore or maxScore is not None and score > maxScore): continue if alignment.query_sequence is None: if lastAlignment is None: raise InvalidSAM( 'pysam produced an alignment (number %d) with no query sequence without previously giving an alignment with a sequence.' % count) assert alignment.cigartuples alignment.query_sequence, alignment.query_qualities, _ = ( _hardClip(lastAlignment.query_sequence, lastAlignment. query_qualities, alignment.cigartuples)) else: lastAlignment = alignment if alignment.cigartuples: (alignment.query_sequence, alignment.query_qualities, _ ) = (_hardClip(alignment.query_sequence, alignment. query_qualities, alignment.cigartuples)) if (filterRead is None or filterRead(Read(alignment.query_name, alignment.query_sequence, alignment.qual))) and not ( referenceIds and alignment.reference_name not in referenceIds or alignment.is_unmapped and dropUnmapped or alignment.is_secondary and dropSecondary or alignment. is_supplementary and dropSupplementary or alignment. is_duplicate and dropDuplicates or alignment.is_qcfail and not keepQCFailures): yield alignment self.alignmentCount = count
alignments
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n referenceIds = self.referenceIds\n dropUnmapped = self.dropUnmapped\n dropSecondary = self.dropSecondary\n dropSupplementary = self.dropSupplementary\n dropDuplicates = self.dropDuplicates\n keepQCFailures = self.keepQCFailures\n storeQueryIds = self.storeQueryIds\n filterRead = self.filterRead\n minScore = self.minScore\n maxScore = self.maxScore\n scoreTag = self.scoreTag\n if storeQueryIds:\n self.queryIds = queryIds = set()\n lastAlignment = None\n count = 0\n with samfile(self.filename) as samAlignment:\n for count, alignment in enumerate(samAlignment.fetch(), start=1):\n if storeQueryIds:\n queryIds.add(alignment.query_name)\n if minScore is not None or maxScore is not None:\n try:\n score = alignment.get_tag(scoreTag)\n except KeyError:\n continue\n else:\n if (minScore is not None and score < minScore or \n maxScore is not None and score > maxScore):\n continue\n if alignment.query_sequence is None:\n if lastAlignment is None:\n raise InvalidSAM(\n 'pysam produced an alignment (number %d) with no query sequence without previously giving an alignment with a sequence.'\n % count)\n assert alignment.cigartuples\n alignment.query_sequence, alignment.query_qualities, _ = (\n _hardClip(lastAlignment.query_sequence, lastAlignment.\n query_qualities, alignment.cigartuples))\n else:\n lastAlignment = alignment\n if alignment.cigartuples:\n (alignment.query_sequence, alignment.query_qualities, _\n ) = (_hardClip(alignment.query_sequence, alignment.\n query_qualities, alignment.cigartuples))\n if (filterRead is None or filterRead(Read(alignment.query_name,\n alignment.query_sequence, alignment.qual))) and not (\n referenceIds and alignment.reference_name not in\n referenceIds or alignment.is_unmapped and dropUnmapped or \n alignment.is_secondary and dropSecondary or alignment.\n is_supplementary and dropSupplementary or alignment.\n is_duplicate and dropDuplicates or alignment.is_qcfail and \n not keepQCFailures):\n yield alignment\n self.alignmentCount = count\n", "role": "user" }, { "content": "alignments", "role": "assistant" } ]
def x(self): result = {} with samfile(self.filename) as sam: if self.referenceIds: for referenceId in self.referenceIds: tid = sam.get_tid(referenceId) if tid == -1: raise UnknownReference( 'Reference %r is not present in the SAM/BAM file.' % referenceId) else: result[referenceId] = sam.lengths[tid] else: result = dict(zip(sam.references, sam.lengths)) return result
referenceLengths
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n result = {}\n with samfile(self.filename) as sam:\n if self.referenceIds:\n for referenceId in self.referenceIds:\n tid = sam.get_tid(referenceId)\n if tid == -1:\n raise UnknownReference(\n 'Reference %r is not present in the SAM/BAM file.' %\n referenceId)\n else:\n result[referenceId] = sam.lengths[tid]\n else:\n result = dict(zip(sam.references, sam.lengths))\n return result\n", "role": "user" }, { "content": "referenceLengths", "role": "assistant" } ]
def x(self, rcSuffix='', rcNeeded=False, padChar='-', queryInsertionChar= 'N', unknownQualityChar='!', allowDuplicateIds=False, addAlignment=False): referenceLength = self.referenceLength idCount = Counter() MATCH_OPERATIONS = {CMATCH, CEQUAL, CDIFF} for lineNumber, alignment in enumerate(self.samFilter.alignments(), start=1 ): query = alignment.query_sequence quality = ''.join(chr(q + 33) for q in alignment.query_qualities) if alignment.is_reverse: if rcNeeded: query = DNARead('id', query).reverseComplement().sequence quality = quality[::-1] if rcSuffix: alignment.query_name += rcSuffix if allowDuplicateIds: queryId = alignment.query_name else: count = idCount[alignment.query_name] idCount[alignment.query_name] += 1 queryId = alignment.query_name + ('' if count == 0 else '/%d' % count) referenceStart = alignment.reference_start atStart = True queryIndex = 0 referenceIndex = referenceStart alignedSequence = '' alignedQuality = '' for operation, length in alignment.cigartuples: if operation in MATCH_OPERATIONS: atStart = False alignedSequence += query[queryIndex:queryIndex + length] alignedQuality += quality[queryIndex:queryIndex + length] elif operation == CINS: atStart = False self.referenceInsertions[queryId].append((referenceIndex, query[queryIndex:queryIndex + length])) elif operation == CDEL: atStart = False alignedSequence += queryInsertionChar * length alignedQuality += unknownQualityChar * length elif operation == CREF_SKIP: atStart = False alignedSequence += queryInsertionChar * length alignedQuality += unknownQualityChar * length elif operation == CSOFT_CLIP: if atStart: unwantedLeft = length - referenceStart if unwantedLeft > 0: alignedSequence += query[queryIndex + unwantedLeft: queryIndex + length] alignedQuality += quality[queryIndex + unwantedLeft :queryIndex + length] referenceStart = 0 else: referenceStart -= length alignedSequence += query[queryIndex:queryIndex + length ] alignedQuality += quality[queryIndex:queryIndex + length] else: unwantedRight = referenceStart + len(alignedSequence ) + length - referenceLength if unwantedRight > 0: alignedSequence += query[queryIndex:queryIndex + length - unwantedRight] alignedQuality += quality[queryIndex:queryIndex + length - unwantedRight] else: alignedSequence += query[queryIndex:queryIndex + length ] alignedQuality += quality[queryIndex:queryIndex + length] elif operation == CHARD_CLIP: pass elif operation == CPAD: atStart = False else: raise ValueError('Unknown CIGAR operation:', operation) if operation in _CONSUMES_QUERY: queryIndex += length if operation in _CONSUMES_REFERENCE: referenceIndex += length if queryIndex != len(query): raise ValueError( 'Query %r not fully consumed when parsing CIGAR string. Query %r (len %d), final query index %d, CIGAR: %r' % (alignment.query_name, query, len(query), queryIndex, alignment.cigartuples)) padRightLength = referenceLength - (referenceStart + len( alignedSequence)) paddedSequence = (padChar * referenceStart + alignedSequence + padChar * padRightLength) paddedQuality = (unknownQualityChar * referenceStart + alignedQuality + unknownQualityChar * padRightLength) read = Read(queryId, paddedSequence, paddedQuality) if addAlignment: read.alignment = alignment yield read
queries
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, rcSuffix='', rcNeeded=False, padChar='-', queryInsertionChar=\n 'N', unknownQualityChar='!', allowDuplicateIds=False, addAlignment=False):\n referenceLength = self.referenceLength\n idCount = Counter()\n MATCH_OPERATIONS = {CMATCH, CEQUAL, CDIFF}\n for lineNumber, alignment in enumerate(self.samFilter.alignments(), start=1\n ):\n query = alignment.query_sequence\n quality = ''.join(chr(q + 33) for q in alignment.query_qualities)\n if alignment.is_reverse:\n if rcNeeded:\n query = DNARead('id', query).reverseComplement().sequence\n quality = quality[::-1]\n if rcSuffix:\n alignment.query_name += rcSuffix\n if allowDuplicateIds:\n queryId = alignment.query_name\n else:\n count = idCount[alignment.query_name]\n idCount[alignment.query_name] += 1\n queryId = alignment.query_name + ('' if count == 0 else '/%d' %\n count)\n referenceStart = alignment.reference_start\n atStart = True\n queryIndex = 0\n referenceIndex = referenceStart\n alignedSequence = ''\n alignedQuality = ''\n for operation, length in alignment.cigartuples:\n if operation in MATCH_OPERATIONS:\n atStart = False\n alignedSequence += query[queryIndex:queryIndex + length]\n alignedQuality += quality[queryIndex:queryIndex + length]\n elif operation == CINS:\n atStart = False\n self.referenceInsertions[queryId].append((referenceIndex,\n query[queryIndex:queryIndex + length]))\n elif operation == CDEL:\n atStart = False\n alignedSequence += queryInsertionChar * length\n alignedQuality += unknownQualityChar * length\n elif operation == CREF_SKIP:\n atStart = False\n alignedSequence += queryInsertionChar * length\n alignedQuality += unknownQualityChar * length\n elif operation == CSOFT_CLIP:\n if atStart:\n unwantedLeft = length - referenceStart\n if unwantedLeft > 0:\n alignedSequence += query[queryIndex + unwantedLeft:\n queryIndex + length]\n alignedQuality += quality[queryIndex + unwantedLeft\n :queryIndex + length]\n referenceStart = 0\n else:\n referenceStart -= length\n alignedSequence += query[queryIndex:queryIndex + length\n ]\n alignedQuality += quality[queryIndex:queryIndex +\n length]\n else:\n unwantedRight = referenceStart + len(alignedSequence\n ) + length - referenceLength\n if unwantedRight > 0:\n alignedSequence += query[queryIndex:queryIndex +\n length - unwantedRight]\n alignedQuality += quality[queryIndex:queryIndex +\n length - unwantedRight]\n else:\n alignedSequence += query[queryIndex:queryIndex + length\n ]\n alignedQuality += quality[queryIndex:queryIndex +\n length]\n elif operation == CHARD_CLIP:\n pass\n elif operation == CPAD:\n atStart = False\n else:\n raise ValueError('Unknown CIGAR operation:', operation)\n if operation in _CONSUMES_QUERY:\n queryIndex += length\n if operation in _CONSUMES_REFERENCE:\n referenceIndex += length\n if queryIndex != len(query):\n raise ValueError(\n 'Query %r not fully consumed when parsing CIGAR string. Query %r (len %d), final query index %d, CIGAR: %r'\n % (alignment.query_name, query, len(query), queryIndex,\n alignment.cigartuples))\n padRightLength = referenceLength - (referenceStart + len(\n alignedSequence))\n paddedSequence = (padChar * referenceStart + alignedSequence + \n padChar * padRightLength)\n paddedQuality = (unknownQualityChar * referenceStart +\n alignedQuality + unknownQualityChar * padRightLength)\n read = Read(queryId, paddedSequence, paddedQuality)\n if addAlignment:\n read.alignment = alignment\n yield read\n", "role": "user" }, { "content": "queries", "role": "assistant" } ]
def x(sel, ls_di, do): N = len(ls_di) @always_comb def x(): do.next = 0 for i in range(N): if i == sel: do.next = ls_di[i] return _mux
mux
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(sel, ls_di, do):\n N = len(ls_di)\n\n @always_comb\n def x():\n do.next = 0\n for i in range(N):\n if i == sel:\n do.next = ls_di[i]\n return _mux\n", "role": "user" }, { "content": "mux", "role": "assistant" } ]
def x(sel, di, ls_do): N = len(ls_do) @always_comb def x(): for i in range(N): ls_do[i].next = 0 if i == sel: ls_do[i].next = di return _demux
demux
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(sel, di, ls_do):\n N = len(ls_do)\n\n @always_comb\n def x():\n for i in range(N):\n ls_do[i].next = 0\n if i == sel:\n ls_do[i].next = di\n return _demux\n", "role": "user" }, { "content": "demux", "role": "assistant" } ]
def x(sel, lsls_di, ls_do): N = len(ls_do) lsls_in = [list(x) for x in zip(*lsls_di)] return [mux(sel, lsls_in[i], ls_do[i]) for i in range(N)]
ls_mux
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(sel, lsls_di, ls_do):\n N = len(ls_do)\n lsls_in = [list(x) for x in zip(*lsls_di)]\n return [mux(sel, lsls_in[i], ls_do[i]) for i in range(N)]\n", "role": "user" }, { "content": "ls_mux", "role": "assistant" } ]
def x(sel, ls_di, lsls_do): N = len(ls_di) lsls_out = [list(x) for x in zip(*lsls_do)] return [demux(sel, ls_di[i], lsls_out[i]) for i in range(N)]
ls_demux
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(sel, ls_di, lsls_do):\n N = len(ls_di)\n lsls_out = [list(x) for x in zip(*lsls_do)]\n return [demux(sel, ls_di[i], lsls_out[i]) for i in range(N)]\n", "role": "user" }, { "content": "ls_demux", "role": "assistant" } ]
def x(offset, bv_di, bv_do): LEN_I = len(bv_di) LEN_O = len(bv_do) assert LEN_I >= LEN_O, 'bitslice_select: expects len(bv_di) >= len(bv_do), but len(bv_di)={}, len(bv_do)'.format( LEN_I, LEN_O) OFFSET_MAX = LEN_I - LEN_O + 1 @always_comb def x(): bv_do.next = 0 for i in range(OFFSET_MAX): if i == offset: for b in range(LEN_O): bv_do.next[b] = bv_di[i + b] return _slice
bitslice_select
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(offset, bv_di, bv_do):\n LEN_I = len(bv_di)\n LEN_O = len(bv_do)\n assert LEN_I >= LEN_O, 'bitslice_select: expects len(bv_di) >= len(bv_do), but len(bv_di)={}, len(bv_do)'.format(\n LEN_I, LEN_O)\n OFFSET_MAX = LEN_I - LEN_O + 1\n\n @always_comb\n def x():\n bv_do.next = 0\n for i in range(OFFSET_MAX):\n if i == offset:\n for b in range(LEN_O):\n bv_do.next[b] = bv_di[i + b]\n return _slice\n", "role": "user" }, { "content": "bitslice_select", "role": "assistant" } ]
def x(offset, bv_di, bv_do): LEN_I = len(bv_di) LEN_O = len(bv_do) assert LEN_I % 8 == 0, 'byteslice_select: expects len(bv_di)=8*x, but len(bv_di)={} bits'.format( LEN_I) assert LEN_O % 8 == 0, 'byteslice_select: expects len(bv_do)=8*x, but len(bv_do)={} bits'.format( LEN_O) bit_offset = Signal(intbv(0)[len(offset) + 3:]) @always_comb def x(): bit_offset.next = offset << 3 _slice = bitslice_select(bit_offset, bv_di, bv_do) return _offset, _slice
byteslice_select
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(offset, bv_di, bv_do):\n LEN_I = len(bv_di)\n LEN_O = len(bv_do)\n assert LEN_I % 8 == 0, 'byteslice_select: expects len(bv_di)=8*x, but len(bv_di)={} bits'.format(\n LEN_I)\n assert LEN_O % 8 == 0, 'byteslice_select: expects len(bv_do)=8*x, but len(bv_do)={} bits'.format(\n LEN_O)\n bit_offset = Signal(intbv(0)[len(offset) + 3:])\n\n @always_comb\n def x():\n bit_offset.next = offset << 3\n _slice = bitslice_select(bit_offset, bv_di, bv_do)\n return _offset, _slice\n", "role": "user" }, { "content": "byteslice_select", "role": "assistant" } ]
def x(): file_mv = cli.file_mv desc = 'Find all runs (SRR..) associated with an SRA experiment (SRX...).' parser = cli.get_argument_parser(desc=desc) parser.add_argument('-e', '--experiment-file', type=str, required=True, metavar=file_mv, help= 'File with SRA experiment IDs (starting with "SRX").') parser.add_argument('-o', '--output-file', type=str, required=True, metavar=file_mv, help='The output file.') cli.add_reporting_args(parser) return parser
get_argument_parser
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n file_mv = cli.file_mv\n desc = 'Find all runs (SRR..) associated with an SRA experiment (SRX...).'\n parser = cli.get_argument_parser(desc=desc)\n parser.add_argument('-e', '--experiment-file', type=str, required=True,\n metavar=file_mv, help=\n 'File with SRA experiment IDs (starting with \"SRX\").')\n parser.add_argument('-o', '--output-file', type=str, required=True,\n metavar=file_mv, help='The output file.')\n cli.add_reporting_args(parser)\n return parser\n", "role": "user" }, { "content": "get_argument_parser", "role": "assistant" } ]
def x(args=None): if args is None: parser = get_argument_parser() args = parser.parse_args() experiment_file = args.experiment_file output_file = args.output_file host = 'ftp-trace.ncbi.nlm.nih.gov' user = 'anonymous' password = 'anonymous' experiments = misc.read_single(experiment_file) runs = [] with ftputil.FTPHost(host, user, password) as ftp_host: for exp in experiments: exp_dir = '/sra/sra-instant/reads/ByExp/sra/SRX/%s/%s/' % (exp[ :6], exp) ftp_host.chdir(exp_dir) run_folders = ftp_host.listdir(ftp_host.curdir) for folder in run_folders: files = ftp_host.listdir(folder) assert len(files) == 1 runs.append((exp, folder)) with open(output_file, 'wb') as ofh: writer = csv.writer(ofh, dialect='excel-tab', lineterminator=os. linesep, quoting=csv.QUOTE_NONE) for r in runs: writer.writerow(r) return 0
main
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(args=None):\n if args is None:\n parser = get_argument_parser()\n args = parser.parse_args()\n experiment_file = args.experiment_file\n output_file = args.output_file\n host = 'ftp-trace.ncbi.nlm.nih.gov'\n user = 'anonymous'\n password = 'anonymous'\n experiments = misc.read_single(experiment_file)\n runs = []\n with ftputil.FTPHost(host, user, password) as ftp_host:\n for exp in experiments:\n exp_dir = '/sra/sra-instant/reads/ByExp/sra/SRX/%s/%s/' % (exp[\n :6], exp)\n ftp_host.chdir(exp_dir)\n run_folders = ftp_host.listdir(ftp_host.curdir)\n for folder in run_folders:\n files = ftp_host.listdir(folder)\n assert len(files) == 1\n runs.append((exp, folder))\n with open(output_file, 'wb') as ofh:\n writer = csv.writer(ofh, dialect='excel-tab', lineterminator=os.\n linesep, quoting=csv.QUOTE_NONE)\n for r in runs:\n writer.writerow(r)\n return 0\n", "role": "user" }, { "content": "main", "role": "assistant" } ]
def x(cls, query_obj, search4, model_attrib): model = model_attrib.parent.class_ if isinstance(search4, str): query_obj = query_obj.join(model).filter(model_attrib.like(search4)) elif isinstance(search4, int): query_obj = query_obj.join(model).filter(model_attrib == search4) elif isinstance(search4, Iterable): query_obj = query_obj.join(model).filter(model_attrib.in_(search4)) return query_obj
one_to_many_query
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, query_obj, search4, model_attrib):\n model = model_attrib.parent.class_\n if isinstance(search4, str):\n query_obj = query_obj.join(model).filter(model_attrib.like(search4))\n elif isinstance(search4, int):\n query_obj = query_obj.join(model).filter(model_attrib == search4)\n elif isinstance(search4, Iterable):\n query_obj = query_obj.join(model).filter(model_attrib.in_(search4))\n return query_obj\n", "role": "user" }, { "content": "one_to_many_query", "role": "assistant" } ]
def x(self, name=None, identifier=None, entry_name=None, limit=None, as_df= False): q = self.session.query(models.Keyword) model_queries_config = (name, models.Keyword.name), (identifier, models .Keyword.identifier) q = self.get_model_queries(q, model_queries_config) q = self.get_many_to_many_queries(q, ((entry_name, models.Keyword. entries, models.Entry.name),)) return self._limit_and_df(q, limit, as_df)
keyword
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, name=None, identifier=None, entry_name=None, limit=None, as_df=\n False):\n q = self.session.query(models.Keyword)\n model_queries_config = (name, models.Keyword.name), (identifier, models\n .Keyword.identifier)\n q = self.get_model_queries(q, model_queries_config)\n q = self.get_many_to_many_queries(q, ((entry_name, models.Keyword.\n entries, models.Entry.name),))\n return self._limit_and_df(q, limit, as_df)\n", "role": "user" }, { "content": "keyword", "role": "assistant" } ]
def x(self, name=None, dataset=None, recommended_full_name=None, recommended_short_name=None, gene_name=None, taxid=None, accession=None, organism_host=None, feature_type=None, function_=None, ec_number=None, db_reference=None, alternative_name=None, disease_comment=None, disease_name=None, tissue_specificity=None, pmid=None, keyword=None, subcellular_location=None, tissue_in_reference=None, sequence=None, limit=None, as_df=False): q = self.session.query(models.Entry) model_queries_config = (dataset, models.Entry.dataset), (name, models. Entry.name), (recommended_full_name, models.Entry.recommended_full_name ), (recommended_short_name, models.Entry.recommended_short_name), ( gene_name, models.Entry.gene_name), (taxid, models.Entry.taxid) q = self.get_model_queries(q, model_queries_config) one_to_many_queries_config = (accession, models.Accession.accession), ( organism_host, models.OrganismHost.taxid), (feature_type, models. Feature.type_), (function_, models.Function.text), (ec_number, models.ECNumber.ec_number), (db_reference, models.DbReference. identifier), (alternative_name, models.AlternativeFullName.name), ( disease_comment, models.DiseaseComment.comment), (tissue_specificity, models.TissueSpecificity.comment), (sequence, models.Sequence.sequence) q = self.get_one_to_many_queries(q, one_to_many_queries_config) many_to_many_queries_config = (pmid, models.Entry.pmids, models.Pmid.pmid ), (keyword, models.Entry.keywords, models.Keyword.name), ( subcellular_location, models.Entry.subcellular_locations, models. SubcellularLocation.location), (tissue_in_reference, models.Entry. tissue_in_references, models.TissueInReference.tissue) q = self.get_many_to_many_queries(q, many_to_many_queries_config) if disease_name: q = q.join(models.Entry.disease_comments).join(models. DiseaseComment.disease) if isinstance(disease_name, str): q = q.filter(models.Disease.name.like(disease_name)) elif isinstance(disease_name, Iterable): q = q.filter(models.Disease.name.in_(disease_name)) return self._limit_and_df(q, limit, as_df)
entry
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, name=None, dataset=None, recommended_full_name=None,\n recommended_short_name=None, gene_name=None, taxid=None, accession=None,\n organism_host=None, feature_type=None, function_=None, ec_number=None,\n db_reference=None, alternative_name=None, disease_comment=None,\n disease_name=None, tissue_specificity=None, pmid=None, keyword=None,\n subcellular_location=None, tissue_in_reference=None, sequence=None,\n limit=None, as_df=False):\n q = self.session.query(models.Entry)\n model_queries_config = (dataset, models.Entry.dataset), (name, models.\n Entry.name), (recommended_full_name, models.Entry.recommended_full_name\n ), (recommended_short_name, models.Entry.recommended_short_name), (\n gene_name, models.Entry.gene_name), (taxid, models.Entry.taxid)\n q = self.get_model_queries(q, model_queries_config)\n one_to_many_queries_config = (accession, models.Accession.accession), (\n organism_host, models.OrganismHost.taxid), (feature_type, models.\n Feature.type_), (function_, models.Function.text), (ec_number,\n models.ECNumber.ec_number), (db_reference, models.DbReference.\n identifier), (alternative_name, models.AlternativeFullName.name), (\n disease_comment, models.DiseaseComment.comment), (tissue_specificity,\n models.TissueSpecificity.comment), (sequence, models.Sequence.sequence)\n q = self.get_one_to_many_queries(q, one_to_many_queries_config)\n many_to_many_queries_config = (pmid, models.Entry.pmids, models.Pmid.pmid\n ), (keyword, models.Entry.keywords, models.Keyword.name), (\n subcellular_location, models.Entry.subcellular_locations, models.\n SubcellularLocation.location), (tissue_in_reference, models.Entry.\n tissue_in_references, models.TissueInReference.tissue)\n q = self.get_many_to_many_queries(q, many_to_many_queries_config)\n if disease_name:\n q = q.join(models.Entry.disease_comments).join(models.\n DiseaseComment.disease)\n if isinstance(disease_name, str):\n q = q.filter(models.Disease.name.like(disease_name))\n elif isinstance(disease_name, Iterable):\n q = q.filter(models.Disease.name.in_(disease_name))\n return self._limit_and_df(q, limit, as_df)\n", "role": "user" }, { "content": "entry", "role": "assistant" } ]
def x(self, identifier=None, ref_id=None, ref_type=None, name=None, acronym =None, description=None, entry_name=None, limit=None, as_df=False): q = self.session.query(models.Disease) model_queries_config = (identifier, models.Disease.identifier), (ref_id, models.Disease.ref_id), (ref_type, models.Disease.ref_type), (name, models.Disease.name), (acronym, models.Disease.acronym), (description, models.Disease.description) q = self.get_model_queries(q, model_queries_config) if entry_name: q = q.session.query(models.Disease).join(models.DiseaseComment).join( models.Entry) if isinstance(entry_name, str): q = q.filter(models.Entry.name == entry_name) elif isinstance(entry_name, Iterable): q = q.filter(models.Entry.name.in_(entry_name)) return self._limit_and_df(q, limit, as_df)
disease
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, identifier=None, ref_id=None, ref_type=None, name=None, acronym\n =None, description=None, entry_name=None, limit=None, as_df=False):\n q = self.session.query(models.Disease)\n model_queries_config = (identifier, models.Disease.identifier), (ref_id,\n models.Disease.ref_id), (ref_type, models.Disease.ref_type), (name,\n models.Disease.name), (acronym, models.Disease.acronym), (description,\n models.Disease.description)\n q = self.get_model_queries(q, model_queries_config)\n if entry_name:\n q = q.session.query(models.Disease).join(models.DiseaseComment).join(\n models.Entry)\n if isinstance(entry_name, str):\n q = q.filter(models.Entry.name == entry_name)\n elif isinstance(entry_name, Iterable):\n q = q.filter(models.Entry.name.in_(entry_name))\n return self._limit_and_df(q, limit, as_df)\n", "role": "user" }, { "content": "disease", "role": "assistant" } ]
def x(self, comment=None, entry_name=None, limit=None, as_df=False): q = self.session.query(models.DiseaseComment) q = self.get_model_queries(q, ((comment, models.DiseaseComment.comment),)) q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),)) return self._limit_and_df(q, limit, as_df)
disease_comment
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, comment=None, entry_name=None, limit=None, as_df=False):\n q = self.session.query(models.DiseaseComment)\n q = self.get_model_queries(q, ((comment, models.DiseaseComment.comment),))\n q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),))\n return self._limit_and_df(q, limit, as_df)\n", "role": "user" }, { "content": "disease_comment", "role": "assistant" } ]
def x(self, type_=None, name=None, entry_name=None, limit=None, as_df=None): q = self.session.query(models.OtherGeneName) model_queries_config = (type_, models.OtherGeneName.type_), (name, models.OtherGeneName.name) q = self.get_model_queries(q, model_queries_config) q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),)) return self._limit_and_df(q, limit, as_df)
other_gene_name
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, type_=None, name=None, entry_name=None, limit=None, as_df=None):\n q = self.session.query(models.OtherGeneName)\n model_queries_config = (type_, models.OtherGeneName.type_), (name,\n models.OtherGeneName.name)\n q = self.get_model_queries(q, model_queries_config)\n q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),))\n return self._limit_and_df(q, limit, as_df)\n", "role": "user" }, { "content": "other_gene_name", "role": "assistant" } ]
def x(self, name=None, entry_name=None, limit=None, as_df=False): q = self.session.query(models.AlternativeFullName) model_queries_config = (name, models.AlternativeFullName.name), q = self.get_model_queries(q, model_queries_config) q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),)) return self._limit_and_df(q, limit, as_df)
alternative_full_name
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, name=None, entry_name=None, limit=None, as_df=False):\n q = self.session.query(models.AlternativeFullName)\n model_queries_config = (name, models.AlternativeFullName.name),\n q = self.get_model_queries(q, model_queries_config)\n q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),))\n return self._limit_and_df(q, limit, as_df)\n", "role": "user" }, { "content": "alternative_full_name", "role": "assistant" } ]
def x(self, name=None, entry_name=None, limit=None, as_df=False): q = self.session.query(models.AlternativeShortName) model_queries_config = (name, models.AlternativeShortName.name), q = self.get_model_queries(q, model_queries_config) q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),)) return self._limit_and_df(q, limit, as_df)
alternative_short_name
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, name=None, entry_name=None, limit=None, as_df=False):\n q = self.session.query(models.AlternativeShortName)\n model_queries_config = (name, models.AlternativeShortName.name),\n q = self.get_model_queries(q, model_queries_config)\n q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),))\n return self._limit_and_df(q, limit, as_df)\n", "role": "user" }, { "content": "alternative_short_name", "role": "assistant" } ]
def x(self, accession=None, entry_name=None, limit=None, as_df=False): q = self.session.query(models.Accession) model_queries_config = (accession, models.Accession.accession), q = self.get_model_queries(q, model_queries_config) q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),)) return self._limit_and_df(q, limit, as_df)
accession
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, accession=None, entry_name=None, limit=None, as_df=False):\n q = self.session.query(models.Accession)\n model_queries_config = (accession, models.Accession.accession),\n q = self.get_model_queries(q, model_queries_config)\n q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),))\n return self._limit_and_df(q, limit, as_df)\n", "role": "user" }, { "content": "accession", "role": "assistant" } ]
def x(self, pmid=None, entry_name=None, first=None, last=None, volume=None, name=None, date=None, title=None, limit=None, as_df=False): q = self.session.query(models.Pmid) model_queries_config = (pmid, models.Pmid.pmid), (last, models.Pmid.last ), (first, models.Pmid.first), (volume, models.Pmid.volume), (name, models.Pmid.name), (date, models.Pmid.date), (title, models.Pmid.title) q = self.get_model_queries(q, model_queries_config) q = self.get_many_to_many_queries(q, ((entry_name, models.Pmid.entries, models.Entry.name),)) return self._limit_and_df(q, limit, as_df)
pmid
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, pmid=None, entry_name=None, first=None, last=None, volume=None,\n name=None, date=None, title=None, limit=None, as_df=False):\n q = self.session.query(models.Pmid)\n model_queries_config = (pmid, models.Pmid.pmid), (last, models.Pmid.last\n ), (first, models.Pmid.first), (volume, models.Pmid.volume), (name,\n models.Pmid.name), (date, models.Pmid.date), (title, models.Pmid.title)\n q = self.get_model_queries(q, model_queries_config)\n q = self.get_many_to_many_queries(q, ((entry_name, models.Pmid.entries,\n models.Entry.name),))\n return self._limit_and_df(q, limit, as_df)\n", "role": "user" }, { "content": "pmid", "role": "assistant" } ]
def x(self, taxid=None, entry_name=None, limit=None, as_df=False): q = self.session.query(models.OrganismHost) q = self.get_model_queries(q, ((taxid, models.OrganismHost.taxid),)) q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),)) return self._limit_and_df(q, limit, as_df)
organism_host
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, taxid=None, entry_name=None, limit=None, as_df=False):\n q = self.session.query(models.OrganismHost)\n q = self.get_model_queries(q, ((taxid, models.OrganismHost.taxid),))\n q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),))\n return self._limit_and_df(q, limit, as_df)\n", "role": "user" }, { "content": "organism_host", "role": "assistant" } ]
def x(self, type_=None, identifier=None, entry_name=None, limit=None, as_df =False): q = self.session.query(models.DbReference) model_queries_config = (type_, models.DbReference.type_), (identifier, models.DbReference.identifier) q = self.get_model_queries(q, model_queries_config) q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),)) return self._limit_and_df(q, limit, as_df)
db_reference
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, type_=None, identifier=None, entry_name=None, limit=None, as_df\n =False):\n q = self.session.query(models.DbReference)\n model_queries_config = (type_, models.DbReference.type_), (identifier,\n models.DbReference.identifier)\n q = self.get_model_queries(q, model_queries_config)\n q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),))\n return self._limit_and_df(q, limit, as_df)\n", "role": "user" }, { "content": "db_reference", "role": "assistant" } ]
def x(self, type_=None, identifier=None, description=None, entry_name=None, limit=None, as_df=False): q = self.session.query(models.Feature) model_queries_config = (type_, models.Feature.type_), (identifier, models.Feature.identifier), (description, models.Feature.description) q = self.get_model_queries(q, model_queries_config) q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),)) return self._limit_and_df(q, limit, as_df)
feature
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, type_=None, identifier=None, description=None, entry_name=None,\n limit=None, as_df=False):\n q = self.session.query(models.Feature)\n model_queries_config = (type_, models.Feature.type_), (identifier,\n models.Feature.identifier), (description, models.Feature.description)\n q = self.get_model_queries(q, model_queries_config)\n q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),))\n return self._limit_and_df(q, limit, as_df)\n", "role": "user" }, { "content": "feature", "role": "assistant" } ]
def x(self, text=None, entry_name=None, limit=None, as_df=False): q = self.session.query(models.Function) model_queries_config = (text, models.Function.text), q = self.get_model_queries(q, model_queries_config) q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),)) return self._limit_and_df(q, limit, as_df)
function
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, text=None, entry_name=None, limit=None, as_df=False):\n q = self.session.query(models.Function)\n model_queries_config = (text, models.Function.text),\n q = self.get_model_queries(q, model_queries_config)\n q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),))\n return self._limit_and_df(q, limit, as_df)\n", "role": "user" }, { "content": "function", "role": "assistant" } ]
def x(self, ec_number=None, entry_name=None, limit=None, as_df=False): q = self.session.query(models.ECNumber) q = self.get_model_queries(q, ((ec_number, models.ECNumber.ec_number),)) q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),)) return self._limit_and_df(q, limit, as_df)
ec_number
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, ec_number=None, entry_name=None, limit=None, as_df=False):\n q = self.session.query(models.ECNumber)\n q = self.get_model_queries(q, ((ec_number, models.ECNumber.ec_number),))\n q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),))\n return self._limit_and_df(q, limit, as_df)\n", "role": "user" }, { "content": "ec_number", "role": "assistant" } ]
def x(self, location=None, entry_name=None, limit=None, as_df=False): q = self.session.query(models.SubcellularLocation) q = self.get_model_queries(q, ((location, models.SubcellularLocation. location),)) q = self.get_many_to_many_queries(q, ((entry_name, models. SubcellularLocation.entries, models.Entry.name),)) return self._limit_and_df(q, limit, as_df)
subcellular_location
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, location=None, entry_name=None, limit=None, as_df=False):\n q = self.session.query(models.SubcellularLocation)\n q = self.get_model_queries(q, ((location, models.SubcellularLocation.\n location),))\n q = self.get_many_to_many_queries(q, ((entry_name, models.\n SubcellularLocation.entries, models.Entry.name),))\n return self._limit_and_df(q, limit, as_df)\n", "role": "user" }, { "content": "subcellular_location", "role": "assistant" } ]
def x(self, comment=None, entry_name=None, limit=None, as_df=False): q = self.session.query(models.TissueSpecificity) q = self.get_model_queries(q, ((comment, models.TissueSpecificity. comment),)) q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),)) return self._limit_and_df(q, limit, as_df)
tissue_specificity
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, comment=None, entry_name=None, limit=None, as_df=False):\n q = self.session.query(models.TissueSpecificity)\n q = self.get_model_queries(q, ((comment, models.TissueSpecificity.\n comment),))\n q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),))\n return self._limit_and_df(q, limit, as_df)\n", "role": "user" }, { "content": "tissue_specificity", "role": "assistant" } ]
def x(self, tissue=None, entry_name=None, limit=None, as_df=False): q = self.session.query(models.TissueInReference) model_queries_config = (tissue, models.TissueInReference.tissue), q = self.get_model_queries(q, model_queries_config) q = self.get_many_to_many_queries(q, ((entry_name, models. TissueInReference.entries, models.Entry.name),)) return self._limit_and_df(q, limit, as_df)
tissue_in_reference
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, tissue=None, entry_name=None, limit=None, as_df=False):\n q = self.session.query(models.TissueInReference)\n model_queries_config = (tissue, models.TissueInReference.tissue),\n q = self.get_model_queries(q, model_queries_config)\n q = self.get_many_to_many_queries(q, ((entry_name, models.\n TissueInReference.entries, models.Entry.name),))\n return self._limit_and_df(q, limit, as_df)\n", "role": "user" }, { "content": "tissue_in_reference", "role": "assistant" } ]
def x(self): q = self.session.query(distinct(models.DbReference.type_)) return [x[0] for x in q.all()]
dbreference_types
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n q = self.session.query(distinct(models.DbReference.type_))\n return [x[0] for x in q.all()]\n", "role": "user" }, { "content": "dbreference_types", "role": "assistant" } ]
def x(self): r = self.session.query(distinct(models.Entry.taxid)).all() return [x[0] for x in r]
taxids
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n r = self.session.query(distinct(models.Entry.taxid)).all()\n return [x[0] for x in r]\n", "role": "user" }, { "content": "taxids", "role": "assistant" } ]
def x(self): r = self.session.query(distinct(models.Entry.dataset)).all() return [x[0] for x in r]
datasets
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n r = self.session.query(distinct(models.Entry.dataset)).all()\n return [x[0] for x in r]\n", "role": "user" }, { "content": "datasets", "role": "assistant" } ]
def x(self): r = self.session.query(distinct(models.Feature.type_)).all() return [x[0] for x in r]
feature_types
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n r = self.session.query(distinct(models.Feature.type_)).all()\n return [x[0] for x in r]\n", "role": "user" }, { "content": "feature_types", "role": "assistant" } ]
def x(self): return [x[0] for x in self.session.query(models.SubcellularLocation. location).all()]
subcellular_locations
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return [x[0] for x in self.session.query(models.SubcellularLocation.\n location).all()]\n", "role": "user" }, { "content": "subcellular_locations", "role": "assistant" } ]
def x(self): return [x[0] for x in self.session.query(models.TissueInReference. tissue).all()]
tissues_in_references
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return [x[0] for x in self.session.query(models.TissueInReference.\n tissue).all()]\n", "role": "user" }, { "content": "tissues_in_references", "role": "assistant" } ]
def x(self): return [x[0] for x in self.session.query(models.Keyword.name).all()]
keywords
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return [x[0] for x in self.session.query(models.Keyword.name).all()]\n", "role": "user" }, { "content": "keywords", "role": "assistant" } ]
def x(self): return [x[0] for x in self.session.query(models.Disease.name).all()]
diseases
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return [x[0] for x in self.session.query(models.Disease.name).all()]\n", "role": "user" }, { "content": "diseases", "role": "assistant" } ]
def x(self): return [x for x in self.session.query(models.Version).all()]
version
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return [x for x in self.session.query(models.Version).all()]\n", "role": "user" }, { "content": "version", "role": "assistant" } ]
def x(self, force=False): if not force: try: self.api.call_rc('ixPortTakeOwnership {}'.format(self.uri)) except Exception as _: raise TgnError( 'Failed to take ownership for port {} current owner is {}'. format(self, self.owner)) else: self.api.call_rc('ixPortTakeOwnership {} force'.format(self.uri))
reserve
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, force=False):\n if not force:\n try:\n self.api.call_rc('ixPortTakeOwnership {}'.format(self.uri))\n except Exception as _:\n raise TgnError(\n 'Failed to take ownership for port {} current owner is {}'.\n format(self, self.owner))\n else:\n self.api.call_rc('ixPortTakeOwnership {} force'.format(self.uri))\n", "role": "user" }, { "content": "reserve", "role": "assistant" } ]
def x(self): self.ix_command('write') stream_warnings = self.streamRegion.generateWarningList() warnings_list = self.api.call('join ' + ' {' + stream_warnings + '} ' + ' LiStSeP').split('LiStSeP') if self.streamRegion.generateWarningList( ) else [] for warning in warnings_list: if warning: raise StreamWarningsError(warning)
write
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n self.ix_command('write')\n stream_warnings = self.streamRegion.generateWarningList()\n warnings_list = self.api.call('join ' + ' {' + stream_warnings + '} ' +\n ' LiStSeP').split('LiStSeP') if self.streamRegion.generateWarningList(\n ) else []\n for warning in warnings_list:\n if warning:\n raise StreamWarningsError(warning)\n", "role": "user" }, { "content": "write", "role": "assistant" } ]
def x(self, config_file_name): config_file_name = config_file_name.replace('\\', '/') ext = path.splitext(config_file_name)[-1].lower() if ext == '.prt': self.api.call_rc('port import "{}" {}'.format(config_file_name, self.uri)) elif ext == '.str': self.reset() self.api.call_rc('stream import "{}" {}'.format(config_file_name, self.uri)) else: raise ValueError('Configuration file type {} not supported.'.format (ext)) self.write() self.discover()
load_config
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, config_file_name):\n config_file_name = config_file_name.replace('\\\\', '/')\n ext = path.splitext(config_file_name)[-1].lower()\n if ext == '.prt':\n self.api.call_rc('port import \"{}\" {}'.format(config_file_name,\n self.uri))\n elif ext == '.str':\n self.reset()\n self.api.call_rc('stream import \"{}\" {}'.format(config_file_name,\n self.uri))\n else:\n raise ValueError('Configuration file type {} not supported.'.format\n (ext))\n self.write()\n self.discover()\n", "role": "user" }, { "content": "load_config", "role": "assistant" } ]
def x(self, config_file_name): config_file_name = config_file_name.replace('\\', '/') ext = path.splitext(config_file_name)[-1].lower() if ext == '.prt': self.api.call_rc('port export "{}" {}'.format(config_file_name, self.uri)) elif ext == '.str': self.api.call_rc('stream export "{}" {}'.format(config_file_name, self.uri)) else: raise ValueError('Configuration file type {} not supported.'.format (ext))
save_config
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, config_file_name):\n config_file_name = config_file_name.replace('\\\\', '/')\n ext = path.splitext(config_file_name)[-1].lower()\n if ext == '.prt':\n self.api.call_rc('port export \"{}\" {}'.format(config_file_name,\n self.uri))\n elif ext == '.str':\n self.api.call_rc('stream export \"{}\" {}'.format(config_file_name,\n self.uri))\n else:\n raise ValueError('Configuration file type {} not supported.'.format\n (ext))\n", "role": "user" }, { "content": "save_config", "role": "assistant" } ]
def x(self, blocking=False): self.session.start_transmit(blocking, False, self)
start_transmit
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, blocking=False):\n self.session.start_transmit(blocking, False, self)\n", "role": "user" }, { "content": "start_transmit", "role": "assistant" } ]
def x(self, cap_file_name=None, cap_file_format=IxeCapFileFormat.mem): return self.session.stop_capture(cap_file_name, cap_file_format, self)[self ]
stop_capture
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, cap_file_name=None, cap_file_format=IxeCapFileFormat.mem):\n return self.session.stop_capture(cap_file_name, cap_file_format, self)[self\n ]\n", "role": "user" }, { "content": "stop_capture", "role": "assistant" } ]
def x(self, *frame_nums): frames = [] for frame_num in frame_nums: if self.captureBuffer.getframe(frame_num) == '0': frames.append(self.captureBuffer.frame) else: frames.append(None) return frames
get_cap_frames
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, *frame_nums):\n frames = []\n for frame_num in frame_nums:\n if self.captureBuffer.getframe(frame_num) == '0':\n frames.append(self.captureBuffer.frame)\n else:\n frames.append(None)\n return frames\n", "role": "user" }, { "content": "get_cap_frames", "role": "assistant" } ]
def x(self): stat = IxeStat(self) stat.ix_set_default() stat.enableValidStats = True stat.ix_set() stat.write()
clear_port_stats
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n stat = IxeStat(self)\n stat.ix_set_default()\n stat.enableValidStats = True\n stat.ix_set()\n stat.write()\n", "role": "user" }, { "content": "clear_port_stats", "role": "assistant" } ]
def x(self, mode=IxePhyMode.ignore): if isinstance(mode, IxePhyMode): if mode.value: self.api.call_rc('port setPhyMode {} {}'.format(mode.value, self.uri)) else: self.api.call_rc('port setPhyMode {} {}'.format(mode, self.uri))
set_phy_mode
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, mode=IxePhyMode.ignore):\n if isinstance(mode, IxePhyMode):\n if mode.value:\n self.api.call_rc('port setPhyMode {} {}'.format(mode.value,\n self.uri))\n else:\n self.api.call_rc('port setPhyMode {} {}'.format(mode, self.uri))\n", "role": "user" }, { "content": "set_phy_mode", "role": "assistant" } ]
def x(self, mode): self.api.call_rc('port setTransmitMode {} {}'.format(mode, self.uri))
set_transmit_mode
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, mode):\n self.api.call_rc('port setTransmitMode {} {}'.format(mode, self.uri))\n", "role": "user" }, { "content": "set_transmit_mode", "role": "assistant" } ]
def x(image): sizes = [] seen_widths = [] for size in settings_sizes(): img = get_thumbnail_shim(image, size) if img.width in seen_widths: continue seen_widths.append(img.width) sizes.append({'url': img.url, 'width': img.width, 'height': img.height} ) return sizes
get_image_sizes
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(image):\n sizes = []\n seen_widths = []\n for size in settings_sizes():\n img = get_thumbnail_shim(image, size)\n if img.width in seen_widths:\n continue\n seen_widths.append(img.width)\n sizes.append({'url': img.url, 'width': img.width, 'height': img.height}\n )\n return sizes\n", "role": "user" }, { "content": "get_image_sizes", "role": "assistant" } ]
def x(src, container='div', classes='', inner_classes='', alt='', background_image=False, no_css=False, aria_hidden=False): context = {'container': container, 'classes': classes, 'aspect_padding_bottom': aspect_ratio_percent(src), 'alt': alt, 'background_image': background_image, 'no_css': no_css, 'inner_classes': inner_classes, 'aria_hidden': aria_hidden} if not get_thumbnail_engine(): context['image'] = src return context sizes = get_image_sizes(src) context['image_sizes'] = sizes context['image'] = sizes[0] context['image_sizes_json'] = json.dumps(sizes) srcset_items = ['{} {}w'.format(size['url'], size['width']) for size in sizes] context['image_sizes_srcset'] = ', '.join(srcset_items) return context
get_template_context
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(src, container='div', classes='', inner_classes='', alt='',\n background_image=False, no_css=False, aria_hidden=False):\n context = {'container': container, 'classes': classes,\n 'aspect_padding_bottom': aspect_ratio_percent(src), 'alt': alt,\n 'background_image': background_image, 'no_css': no_css,\n 'inner_classes': inner_classes, 'aria_hidden': aria_hidden}\n if not get_thumbnail_engine():\n context['image'] = src\n return context\n sizes = get_image_sizes(src)\n context['image_sizes'] = sizes\n context['image'] = sizes[0]\n context['image_sizes_json'] = json.dumps(sizes)\n srcset_items = ['{} {}w'.format(size['url'], size['width']) for size in\n sizes]\n context['image_sizes_srcset'] = ', '.join(srcset_items)\n return context\n", "role": "user" }, { "content": "get_template_context", "role": "assistant" } ]
def x(self, images, reference=None): images = check_images(images) reference = check_reference(images, reference) def x(item): key, image = item return asarray([key, self._get(image, reference)]) transformations = images.map(func, with_keys=True).toarray() if images.shape[0] == 1: transformations = [transformations] algorithm = self.__class__.__name__ return RegistrationModel(dict(transformations), algorithm=algorithm)
fit
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, images, reference=None):\n images = check_images(images)\n reference = check_reference(images, reference)\n\n def x(item):\n key, image = item\n return asarray([key, self._get(image, reference)])\n transformations = images.map(func, with_keys=True).toarray()\n if images.shape[0] == 1:\n transformations = [transformations]\n algorithm = self.__class__.__name__\n return RegistrationModel(dict(transformations), algorithm=algorithm)\n", "role": "user" }, { "content": "fit", "role": "assistant" } ]
def x(self, images, reference=None): images = check_images(images) check_reference(images, reference) def x(image): t = self._get(image, reference) return t.apply(image) return images.map(func)
fit_and_transform
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, images, reference=None):\n images = check_images(images)\n check_reference(images, reference)\n\n def x(image):\n t = self._get(image, reference)\n return t.apply(image)\n return images.map(func)\n", "role": "user" }, { "content": "fit_and_transform", "role": "assistant" } ]
def x(entities, path_patterns, strict=False): if isinstance(path_patterns, string_types): path_patterns = [path_patterns] for pattern in path_patterns: if strict: defined = re.findall('\\{(.*?)(?:<[^>]+>)?\\}', pattern) if set(entities.keys()) - set(defined): continue new_path = pattern optional_patterns = re.findall('\\[(.*?)\\]', pattern) for optional_pattern in optional_patterns: optional_chunk = replace_entities(entities, optional_pattern) or '' new_path = new_path.replace('[%s]' % optional_pattern, optional_chunk) new_path = replace_entities(entities, new_path) if new_path: return new_path return None
build_path
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(entities, path_patterns, strict=False):\n if isinstance(path_patterns, string_types):\n path_patterns = [path_patterns]\n for pattern in path_patterns:\n if strict:\n defined = re.findall('\\\\{(.*?)(?:<[^>]+>)?\\\\}', pattern)\n if set(entities.keys()) - set(defined):\n continue\n new_path = pattern\n optional_patterns = re.findall('\\\\[(.*?)\\\\]', pattern)\n for optional_pattern in optional_patterns:\n optional_chunk = replace_entities(entities, optional_pattern) or ''\n new_path = new_path.replace('[%s]' % optional_pattern,\n optional_chunk)\n new_path = replace_entities(entities, new_path)\n if new_path:\n return new_path\n return None\n", "role": "user" }, { "content": "build_path", "role": "assistant" } ]
def x(environ): settings = {} for name, value in environ.items(): if not name.startswith('DJANGO_'): continue name = name.replace('DJANGO_', '', 1) if _ignore_setting(name): continue try: settings[name] = ast.literal_eval(value) except (SyntaxError, ValueError) as err: LOGGER.warn('Unable to parse setting %s=%s (%s)', name, value, err) return settings
get_settings_from_environment
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(environ):\n settings = {}\n for name, value in environ.items():\n if not name.startswith('DJANGO_'):\n continue\n name = name.replace('DJANGO_', '', 1)\n if _ignore_setting(name):\n continue\n try:\n settings[name] = ast.literal_eval(value)\n except (SyntaxError, ValueError) as err:\n LOGGER.warn('Unable to parse setting %s=%s (%s)', name, value, err)\n return settings\n", "role": "user" }, { "content": "get_settings_from_environment", "role": "assistant" } ]
def x(data, digest_size=DEFAULT_DIGEST_SIZE): if digest_size % 3 != 0: raise ValueError('digest_size must be a multiple of 3') if not 0 <= digest_size <= 63: raise ValueError('digest_size must be between 0 and 63 (bytes)') sha512 = Digest(hashlib.sha512(data.encode(ENC)).digest()) return sha512[:digest_size]
vmc_digest
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(data, digest_size=DEFAULT_DIGEST_SIZE):\n if digest_size % 3 != 0:\n raise ValueError('digest_size must be a multiple of 3')\n if not 0 <= digest_size <= 63:\n raise ValueError('digest_size must be between 0 and 63 (bytes)')\n sha512 = Digest(hashlib.sha512(data.encode(ENC)).digest())\n return sha512[:digest_size]\n", "role": "user" }, { "content": "vmc_digest", "role": "assistant" } ]
def x(matrix, top): assert isinstance(matrix, ExpMatrix) assert isinstance(top, (int, np.integer)) if top >= matrix.p: logger.warning( 'Variance filter has no effect ("top" parameter is >= number of genes).' ) return matrix.copy() var = np.var(matrix.X, axis=1, ddof=1) total_var = np.sum(var) a = np.argsort(var) a = a[::-1] sel = np.zeros(matrix.p, dtype=np.bool_) sel[a[:top]] = True lost_p = matrix.p - top lost_var = total_var - np.sum(var[sel]) logger.info( 'Selected the %d most variable genes (excluded %.1f%% of genes, representing %.1f%% of total variance).' , top, 100 * (lost_p / float(matrix.p)), 100 * (lost_var / total_var)) matrix = matrix.loc[sel] return matrix
filter_variance
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(matrix, top):\n assert isinstance(matrix, ExpMatrix)\n assert isinstance(top, (int, np.integer))\n if top >= matrix.p:\n logger.warning(\n 'Variance filter has no effect (\"top\" parameter is >= number of genes).'\n )\n return matrix.copy()\n var = np.var(matrix.X, axis=1, ddof=1)\n total_var = np.sum(var)\n a = np.argsort(var)\n a = a[::-1]\n sel = np.zeros(matrix.p, dtype=np.bool_)\n sel[a[:top]] = True\n lost_p = matrix.p - top\n lost_var = total_var - np.sum(var[sel])\n logger.info(\n 'Selected the %d most variable genes (excluded %.1f%% of genes, representing %.1f%% of total variance).'\n , top, 100 * (lost_p / float(matrix.p)), 100 * (lost_var / total_var))\n matrix = matrix.loc[sel]\n return matrix\n", "role": "user" }, { "content": "filter_variance", "role": "assistant" } ]
def x(matrix, top): assert isinstance(matrix, ExpMatrix) assert isinstance(top, int) if top >= matrix.p: logger.warning( 'Gene expression filter with `top` parameter that is >= the number of genes!' ) top = matrix.p a = np.argsort(np.mean(matrix.X, axis=1)) a = a[::-1] sel = np.zeros(matrix.p, dtype=np.bool_) sel[a[:top]] = True matrix = matrix.loc[sel] return matrix
filter_mean
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(matrix, top):\n assert isinstance(matrix, ExpMatrix)\n assert isinstance(top, int)\n if top >= matrix.p:\n logger.warning(\n 'Gene expression filter with `top` parameter that is >= the number of genes!'\n )\n top = matrix.p\n a = np.argsort(np.mean(matrix.X, axis=1))\n a = a[::-1]\n sel = np.zeros(matrix.p, dtype=np.bool_)\n sel[a[:top]] = True\n matrix = matrix.loc[sel]\n return matrix\n", "role": "user" }, { "content": "filter_mean", "role": "assistant" } ]
def x(matrix, top, percentile=50): assert isinstance(matrix, ExpMatrix) assert isinstance(top, int) assert isinstance(percentile, (int, float)) if top >= matrix.p: logger.warning( 'Gene expression filter with `top` parameter that is >= the number of genes!' ) top = matrix.p a = np.argsort(np.percentile(matrix.X, percentile, axis=1)) a = a[::-1] sel = np.zeros(matrix.p, dtype=np.bool_) sel[a[:top]] = True matrix = matrix.loc[sel] return matrix
filter_percentile
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(matrix, top, percentile=50):\n assert isinstance(matrix, ExpMatrix)\n assert isinstance(top, int)\n assert isinstance(percentile, (int, float))\n if top >= matrix.p:\n logger.warning(\n 'Gene expression filter with `top` parameter that is >= the number of genes!'\n )\n top = matrix.p\n a = np.argsort(np.percentile(matrix.X, percentile, axis=1))\n a = a[::-1]\n sel = np.zeros(matrix.p, dtype=np.bool_)\n sel[a[:top]] = True\n matrix = matrix.loc[sel]\n return matrix\n", "role": "user" }, { "content": "filter_percentile", "role": "assistant" } ]
def x(chrom): try: c = int(chrom) except: if chrom in ['X', 'Y']: return chrom elif chrom == 'MT': return '_MT' else: return '__' + chrom else: return '%02d' % c
transform_chrom
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(chrom):\n try:\n c = int(chrom)\n except:\n if chrom in ['X', 'Y']:\n return chrom\n elif chrom == 'MT':\n return '_MT'\n else:\n return '__' + chrom\n else:\n return '%02d' % c\n", "role": "user" }, { "content": "transform_chrom", "role": "assistant" } ]
def x(fasta_file, fancy_sort=True): chromlen = [] with gzip.open(fasta_file, 'rt', encoding='ascii') as fh: fasta = SeqIO.parse(fh, 'fasta') for i, f in enumerate(fasta): chromlen.append((f.id, len(f.seq))) _LOGGER.info('Processed chromosome "%s"...', f.id) chromlen = pd.Series(OrderedDict(chromlen)) chromlen.index.name = 'Chromosome' chromlen.name = 'Length' if fancy_sort: chrom_for_sorting = chromlen.index.to_series().apply(_transform_chrom) a = chrom_for_sorting.argsort(kind='mergesort') chromlen = chromlen.iloc[a] return chromlen
get_chromosome_lengths
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(fasta_file, fancy_sort=True):\n chromlen = []\n with gzip.open(fasta_file, 'rt', encoding='ascii') as fh:\n fasta = SeqIO.parse(fh, 'fasta')\n for i, f in enumerate(fasta):\n chromlen.append((f.id, len(f.seq)))\n _LOGGER.info('Processed chromosome \"%s\"...', f.id)\n chromlen = pd.Series(OrderedDict(chromlen))\n chromlen.index.name = 'Chromosome'\n chromlen.name = 'Length'\n if fancy_sort:\n chrom_for_sorting = chromlen.index.to_series().apply(_transform_chrom)\n a = chrom_for_sorting.argsort(kind='mergesort')\n chromlen = chromlen.iloc[a]\n return chromlen\n", "role": "user" }, { "content": "get_chromosome_lengths", "role": "assistant" } ]
def x(schema): def x(schema): if isinstance(schema, dict): if 'allOf' in schema: for x in schema['allOf']: sub_schema = x sub_schema.pop('title', None) schema = _merge_dicts(schema, sub_schema) schema.pop('allOf') schema = traverse(schema) elif 'properties' in schema: for x in schema.get('properties', []): schema['properties'][x] = traverse(schema['properties'][x]) elif 'items' in schema: schema['items'] = traverse(schema['items']) return schema return traverse(schema)
resolve_schema
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(schema):\n\n def x(schema):\n if isinstance(schema, dict):\n if 'allOf' in schema:\n for x in schema['allOf']:\n sub_schema = x\n sub_schema.pop('title', None)\n schema = _merge_dicts(schema, sub_schema)\n schema.pop('allOf')\n schema = traverse(schema)\n elif 'properties' in schema:\n for x in schema.get('properties', []):\n schema['properties'][x] = traverse(schema['properties'][x])\n elif 'items' in schema:\n schema['items'] = traverse(schema['items'])\n return schema\n return traverse(schema)\n", "role": "user" }, { "content": "resolve_schema", "role": "assistant" } ]
def x(first, second): new = deepcopy(first) for k, v in second.items(): if isinstance(v, dict) and v: ret = _merge_dicts(new.get(k, dict()), v) new[k] = ret else: new[k] = second[k] return new
merge_dicts
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(first, second):\n new = deepcopy(first)\n for k, v in second.items():\n if isinstance(v, dict) and v:\n ret = _merge_dicts(new.get(k, dict()), v)\n new[k] = ret\n else:\n new[k] = second[k]\n return new\n", "role": "user" }, { "content": "merge_dicts", "role": "assistant" } ]