source_code stringlengths 45 7.24k | function_name stringlengths 1 111 | messages list |
|---|---|---|
def x(key, project):
cred = get_storage_credentials(key)
return storage.Client(project=project, credentials=cred)
| get_client | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(key, project):\n cred = get_storage_credentials(key)\n return storage.Client(project=project, credentials=cred)\n",
"role": "user"
},
{
"content": "get_client",
"role": "assistant"
}
] |
def x(client, bucket, prefix=''):
bucket = client.get_bucket(bucket)
files = list(bucket.list_blobs(prefix=prefix))
return files
| get_files | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(client, bucket, prefix=''):\n bucket = client.get_bucket(bucket)\n files = list(bucket.list_blobs(prefix=prefix))\n return files\n",
"role": "user"
},
{
"content": "get_files",
"role": "assistant"
}
] |
def x(client, bucket, remote_path, local_path, overwrite=False):
bucket = client.get_bucket(bucket)
if not overwrite and os.path.isfile(local_path):
raise OSError('File already exists!')
with open(local_path, 'wb') as ofh:
bucket.get_blob(remote_path).download_to_file(ofh)
| download_file | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(client, bucket, remote_path, local_path, overwrite=False):\n bucket = client.get_bucket(bucket)\n if not overwrite and os.path.isfile(local_path):\n raise OSError('File already exists!')\n with open(local_path, 'wb') as ofh:\n bucket.get_blob(remote_path).download_to_file(ofh)\n",
"role": "user"
},
{
"content": "download_file",
"role": "assistant"
}
] |
def x(client, bucket, remote_path):
bucket = client.get_bucket(bucket)
bucket.delete_blob(remote_path)
| delete_file | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(client, bucket, remote_path):\n bucket = client.get_bucket(bucket)\n bucket.delete_blob(remote_path)\n",
"role": "user"
},
{
"content": "delete_file",
"role": "assistant"
}
] |
def x(client, bucket, local_path, remote_path, overwrite=False):
bucket = client.get_bucket(bucket)
blob = storage.Blob(remote_path, bucket)
if not overwrite and blob.exists():
raise Conflict('File/object already exists on the bucket!')
blob.upload_from_filename(local_path)
| upload_file | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(client, bucket, local_path, remote_path, overwrite=False):\n bucket = client.get_bucket(bucket)\n blob = storage.Blob(remote_path, bucket)\n if not overwrite and blob.exists():\n raise Conflict('File/object already exists on the bucket!')\n blob.upload_from_filename(local_path)\n",
"role": "user"
},
{
"content": "upload_file",
"role": "assistant"
}
] |
def x(self, readAlignments):
if self.limit is not None and self.count == self.limit:
return False
if self.maxAlignmentsPerRead is not None and len(readAlignments
) > self.maxAlignmentsPerRead:
return False
if self.readIdRegex and self.readIdRegex.search(readAlignments.read.id
) is None:
return False
if self.titleFilter:
wantedAlignments = []
for alignment in readAlignments:
if self.titleFilter.accept(alignment.subjectTitle
) != TitleFilter.REJECT:
wantedAlignments.append(alignment)
if wantedAlignments:
readAlignments[:] = wantedAlignments
else:
return False
minSequenceLen = self.minSequenceLen
maxSequenceLen = self.maxSequenceLen
if minSequenceLen is not None or maxSequenceLen is not None:
wantedAlignments = []
for alignment in readAlignments:
length = alignment.subjectLength
if not (minSequenceLen is not None and length < minSequenceLen or
maxSequenceLen is not None and length > self.maxSequenceLen):
wantedAlignments.append(alignment)
if wantedAlignments:
readAlignments[:] = wantedAlignments
else:
return False
if self.taxonomy is not None:
wantedAlignments = []
for alignment in readAlignments:
lineage = self.lineageFetcher.lineage(alignment.subjectTitle)
if lineage:
for taxonomyIdAndScientificName in lineage:
if self.taxonomy in taxonomyIdAndScientificName:
wantedAlignments.append(alignment)
else:
wantedAlignments.append(alignment)
if wantedAlignments:
readAlignments[:] = wantedAlignments
else:
return False
if self.oneAlignmentPerRead and readAlignments:
readAlignments[:] = [bestAlignment(readAlignments)]
if self.maxHspsPerHit is not None:
for alignment in readAlignments:
hsps = alignment.hsps
if len(hsps) > self.maxHspsPerHit:
alignment.hsps = hsps[:self.maxHspsPerHit]
if self.scoreCutoff is not None:
wantedAlignments = []
for alignment in readAlignments:
hsps = alignment.hsps
wantedHsps = []
for hsp in hsps:
if hsp.betterThan(self.scoreCutoff):
wantedHsps.append(hsp)
if wantedHsps:
alignment.hsps = wantedHsps
wantedAlignments.append(alignment)
if wantedAlignments:
readAlignments[:] = wantedAlignments
else:
return False
minStart = self.minStart
maxStop = self.maxStop
if minStart is not None or maxStop is not None:
wantedAlignments = []
for alignment in readAlignments:
hsps = alignment.hsps
wantedHsps = []
for hsp in hsps:
if not (minStart is not None and hsp.readStartInSubject <
minStart or maxStop is not None and hsp.
readEndInSubject > maxStop):
wantedHsps.append(hsp)
if wantedHsps:
alignment.hsps = wantedHsps
wantedAlignments.append(alignment)
if wantedAlignments:
readAlignments[:] = wantedAlignments
else:
return False
self.count += 1
return readAlignments
| filter | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, readAlignments):\n if self.limit is not None and self.count == self.limit:\n return False\n if self.maxAlignmentsPerRead is not None and len(readAlignments\n ) > self.maxAlignmentsPerRead:\n return False\n if self.readIdRegex and self.readIdRegex.search(readAlignments.read.id\n ) is None:\n return False\n if self.titleFilter:\n wantedAlignments = []\n for alignment in readAlignments:\n if self.titleFilter.accept(alignment.subjectTitle\n ) != TitleFilter.REJECT:\n wantedAlignments.append(alignment)\n if wantedAlignments:\n readAlignments[:] = wantedAlignments\n else:\n return False\n minSequenceLen = self.minSequenceLen\n maxSequenceLen = self.maxSequenceLen\n if minSequenceLen is not None or maxSequenceLen is not None:\n wantedAlignments = []\n for alignment in readAlignments:\n length = alignment.subjectLength\n if not (minSequenceLen is not None and length < minSequenceLen or\n maxSequenceLen is not None and length > self.maxSequenceLen):\n wantedAlignments.append(alignment)\n if wantedAlignments:\n readAlignments[:] = wantedAlignments\n else:\n return False\n if self.taxonomy is not None:\n wantedAlignments = []\n for alignment in readAlignments:\n lineage = self.lineageFetcher.lineage(alignment.subjectTitle)\n if lineage:\n for taxonomyIdAndScientificName in lineage:\n if self.taxonomy in taxonomyIdAndScientificName:\n wantedAlignments.append(alignment)\n else:\n wantedAlignments.append(alignment)\n if wantedAlignments:\n readAlignments[:] = wantedAlignments\n else:\n return False\n if self.oneAlignmentPerRead and readAlignments:\n readAlignments[:] = [bestAlignment(readAlignments)]\n if self.maxHspsPerHit is not None:\n for alignment in readAlignments:\n hsps = alignment.hsps\n if len(hsps) > self.maxHspsPerHit:\n alignment.hsps = hsps[:self.maxHspsPerHit]\n if self.scoreCutoff is not None:\n wantedAlignments = []\n for alignment in readAlignments:\n hsps = alignment.hsps\n wantedHsps = []\n for hsp in hsps:\n if hsp.betterThan(self.scoreCutoff):\n wantedHsps.append(hsp)\n if wantedHsps:\n alignment.hsps = wantedHsps\n wantedAlignments.append(alignment)\n if wantedAlignments:\n readAlignments[:] = wantedAlignments\n else:\n return False\n minStart = self.minStart\n maxStop = self.maxStop\n if minStart is not None or maxStop is not None:\n wantedAlignments = []\n for alignment in readAlignments:\n hsps = alignment.hsps\n wantedHsps = []\n for hsp in hsps:\n if not (minStart is not None and hsp.readStartInSubject <\n minStart or maxStop is not None and hsp.\n readEndInSubject > maxStop):\n wantedHsps.append(hsp)\n if wantedHsps:\n alignment.hsps = wantedHsps\n wantedAlignments.append(alignment)\n if wantedAlignments:\n readAlignments[:] = wantedAlignments\n else:\n return False\n self.count += 1\n return readAlignments\n",
"role": "user"
},
{
"content": "filter",
"role": "assistant"
}
] |
def x(self):
for readAlignments in self:
for alignment in readAlignments:
for hsp in alignment.hsps:
yield hsp
| hsps | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n for readAlignments in self:\n for alignment in readAlignments:\n for hsp in alignment.hsps:\n yield hsp\n",
"role": "user"
},
{
"content": "hsps",
"role": "assistant"
}
] |
def x(self, **kwargs):
self._filters.append(ReadsAlignmentsFilter(**kwargs).filter)
return self
| filter | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, **kwargs):\n self._filters.append(ReadsAlignmentsFilter(**kwargs).filter)\n return self\n",
"role": "user"
},
{
"content": "filter",
"role": "assistant"
}
] |
def x(species, release=None, ftp=None):
assert isinstance(species, (str, _oldstr)) or isinstance(species, Iterable)
if release is not None:
assert isinstance(release, int)
if ftp is not None:
assert isinstance(ftp, ftplib.FTP)
close_connection = False
ftp_server = 'ftp.ensembl.org'
ftp_user = 'anonymous'
if ftp is None:
ftp = ftplib.FTP(ftp_server)
ftp.login(ftp_user)
close_connection = True
if release is None:
release = util.get_latest_release(ftp=ftp)
species_data = OrderedDict()
if isinstance(species, (str, _oldstr)):
species_list = [species]
else:
species_list = species
for spec in species_list:
species_dir = '/pub/release-%d/gtf/%s' % (release, spec.lower())
data = []
ftp.dir(species_dir, data.append)
gtf_file = []
for d in data:
i = d.rindex(' ')
fn = d[i + 1:]
if fn.endswith('.%d.gtf.gz' % release):
gtf_file.append(fn)
assert len(gtf_file) == 1
gtf_file = gtf_file[0]
_LOGGER.debug('GTF file: %s', gtf_file)
checksum_url = '/'.join([species_dir, 'CHECKSUMS'])
file_checksums = util.get_file_checksums(checksum_url, ftp=ftp)
gtf_checksum = file_checksums[gtf_file]
_LOGGER.debug('GTF file checksum: %d', gtf_checksum)
gtf_url = 'ftp://%s%s/%s' % (ftp_server, species_dir, gtf_file)
species_data[spec] = gtf_url, gtf_checksum
if close_connection:
ftp.close()
return species_data
| get_annotation_urls_and_checksums | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(species, release=None, ftp=None):\n assert isinstance(species, (str, _oldstr)) or isinstance(species, Iterable)\n if release is not None:\n assert isinstance(release, int)\n if ftp is not None:\n assert isinstance(ftp, ftplib.FTP)\n close_connection = False\n ftp_server = 'ftp.ensembl.org'\n ftp_user = 'anonymous'\n if ftp is None:\n ftp = ftplib.FTP(ftp_server)\n ftp.login(ftp_user)\n close_connection = True\n if release is None:\n release = util.get_latest_release(ftp=ftp)\n species_data = OrderedDict()\n if isinstance(species, (str, _oldstr)):\n species_list = [species]\n else:\n species_list = species\n for spec in species_list:\n species_dir = '/pub/release-%d/gtf/%s' % (release, spec.lower())\n data = []\n ftp.dir(species_dir, data.append)\n gtf_file = []\n for d in data:\n i = d.rindex(' ')\n fn = d[i + 1:]\n if fn.endswith('.%d.gtf.gz' % release):\n gtf_file.append(fn)\n assert len(gtf_file) == 1\n gtf_file = gtf_file[0]\n _LOGGER.debug('GTF file: %s', gtf_file)\n checksum_url = '/'.join([species_dir, 'CHECKSUMS'])\n file_checksums = util.get_file_checksums(checksum_url, ftp=ftp)\n gtf_checksum = file_checksums[gtf_file]\n _LOGGER.debug('GTF file checksum: %d', gtf_checksum)\n gtf_url = 'ftp://%s%s/%s' % (ftp_server, species_dir, gtf_file)\n species_data[spec] = gtf_url, gtf_checksum\n if close_connection:\n ftp.close()\n return species_data\n",
"role": "user"
},
{
"content": "get_annotation_urls_and_checksums",
"role": "assistant"
}
] |
def x(path_or_buffer, include_polymorphic_pseudogenes=True,
remove_duplicates=True, **kwargs):
valid_biotypes = set(['protein_coding'])
if include_polymorphic_pseudogenes:
valid_biotypes.add('polymorphic_pseudogene')
df = get_genes(path_or_buffer, valid_biotypes, remove_duplicates=
remove_duplicates, **kwargs)
return df
| get_protein_coding_genes | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(path_or_buffer, include_polymorphic_pseudogenes=True,\n remove_duplicates=True, **kwargs):\n valid_biotypes = set(['protein_coding'])\n if include_polymorphic_pseudogenes:\n valid_biotypes.add('polymorphic_pseudogene')\n df = get_genes(path_or_buffer, valid_biotypes, remove_duplicates=\n remove_duplicates, **kwargs)\n return df\n",
"role": "user"
},
{
"content": "get_protein_coding_genes",
"role": "assistant"
}
] |
def x(path_or_buffer, remove_duplicates=True, **kwargs):
valid_biotypes = set(['lincRNA'])
df = get_genes(path_or_buffer, valid_biotypes, remove_duplicates=
remove_duplicates, **kwargs)
return df
| get_linc_rna_genes | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(path_or_buffer, remove_duplicates=True, **kwargs):\n valid_biotypes = set(['lincRNA'])\n df = get_genes(path_or_buffer, valid_biotypes, remove_duplicates=\n remove_duplicates, **kwargs)\n return df\n",
"role": "user"
},
{
"content": "get_linc_rna_genes",
"role": "assistant"
}
] |
def x(levels=0, dimensions=None, sequential=True, string=True, max_int=255,
max_pkt_len=150, max_dim_size=3):
MAX_INT = max_int
if string:
MAX_INT = min(255, max_int)
MAX_DIM_SIZE = max_dim_size
MAX_PKT_LEN = max_pkt_len
def x():
next_i.i = (next_i.i + 1) % (MAX_INT + 1)
return next_i.i
next_i.i = 0
def x(length):
if sequential:
pld = [next_i() for _ in xrange(length)]
else:
pld = [random.randint(0, MAX_INT) for _ in xrange(length)]
if string:
pld = str(bytearray(pld))
return pld
def x(level, pld):
if level > 0:
if pld == None:
pld = random.randint(1, MAX_DIM_SIZE)
if isinstance(pld, int):
pld = pld * [None]
if isinstance(pld, list):
for i in range(len(pld)):
pld[i] = next_level(level - 1, pld[i])
return pld
else:
raise TypeError('Expected None, int or list, got {}: {}'.
format(type(pld), pld))
elif level == 0:
if pld == None:
pld = random.randint(1, MAX_PKT_LEN)
if isinstance(pld, int):
return payload(pld)
else:
raise TypeError('Expected None or int, got {}: {}'.format(
type(pld), pld))
else:
raise ValueError('Expected int>=0, got {}'.format(level))
pld = copy.deepcopy(dimensions)
pld = next_level(levels, pld)
return pld
| payload_generator | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(levels=0, dimensions=None, sequential=True, string=True, max_int=255,\n max_pkt_len=150, max_dim_size=3):\n MAX_INT = max_int\n if string:\n MAX_INT = min(255, max_int)\n MAX_DIM_SIZE = max_dim_size\n MAX_PKT_LEN = max_pkt_len\n\n def x():\n next_i.i = (next_i.i + 1) % (MAX_INT + 1)\n return next_i.i\n next_i.i = 0\n\n def x(length):\n if sequential:\n pld = [next_i() for _ in xrange(length)]\n else:\n pld = [random.randint(0, MAX_INT) for _ in xrange(length)]\n if string:\n pld = str(bytearray(pld))\n return pld\n\n def x(level, pld):\n if level > 0:\n if pld == None:\n pld = random.randint(1, MAX_DIM_SIZE)\n if isinstance(pld, int):\n pld = pld * [None]\n if isinstance(pld, list):\n for i in range(len(pld)):\n pld[i] = next_level(level - 1, pld[i])\n return pld\n else:\n raise TypeError('Expected None, int or list, got {}: {}'.\n format(type(pld), pld))\n elif level == 0:\n if pld == None:\n pld = random.randint(1, MAX_PKT_LEN)\n if isinstance(pld, int):\n return payload(pld)\n else:\n raise TypeError('Expected None or int, got {}: {}'.format(\n type(pld), pld))\n else:\n raise ValueError('Expected int>=0, got {}'.format(level))\n pld = copy.deepcopy(dimensions)\n pld = next_level(levels, pld)\n return pld\n",
"role": "user"
},
{
"content": "payload_generator",
"role": "assistant"
}
] |
def x(title, db='nucleotide'):
titleId = title.split(' ', 1)[0]
try:
gi = titleId.split('|')[1]
except IndexError:
gi = str(titleId)
try:
client = Entrez.efetch(db=db, rettype='gb', retmode='text', id=gi)
except URLError:
return None
else:
record = SeqIO.read(client, 'gb')
client.close()
return record
| getSequence | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(title, db='nucleotide'):\n titleId = title.split(' ', 1)[0]\n try:\n gi = titleId.split('|')[1]\n except IndexError:\n gi = str(titleId)\n try:\n client = Entrez.efetch(db=db, rettype='gb', retmode='text', id=gi)\n except URLError:\n return None\n else:\n record = SeqIO.read(client, 'gb')\n client.close()\n return record\n",
"role": "user"
},
{
"content": "getSequence",
"role": "assistant"
}
] |
def x(self):
lowest_layer = None
for layer in self.cloud_layers:
if layer.coverage not in [CloudLayer.BROKEN, CloudLayer.OVERCAST]:
continue
if lowest_layer is None:
lowest_layer = layer
continue
if layer.height > lowest_layer.height:
continue
if (layer.height < lowest_layer.height or lowest_layer.
get_coverage_percentage() < layer.get_coverage_percentage()):
lowest_layer = layer
return lowest_layer
| get_ceiling_cloud_layer | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n lowest_layer = None\n for layer in self.cloud_layers:\n if layer.coverage not in [CloudLayer.BROKEN, CloudLayer.OVERCAST]:\n continue\n if lowest_layer is None:\n lowest_layer = layer\n continue\n if layer.height > lowest_layer.height:\n continue\n if (layer.height < lowest_layer.height or lowest_layer.\n get_coverage_percentage() < layer.get_coverage_percentage()):\n lowest_layer = layer\n return lowest_layer\n",
"role": "user"
},
{
"content": "get_ceiling_cloud_layer",
"role": "assistant"
}
] |
def x(self):
self.raw_text = self.xml_data.find('raw_text').text
self.station = WeatherStation(self.xml_data.find('station_id').text)
self.station.latitude = float(self.xml_data.find('latitude').text)
self.station.longitude = float(self.xml_data.find('longitude').text)
self.station.elevation = float(self.xml_data.find('elevation_m').text
) * 3.28084
| parse_xml_data | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n self.raw_text = self.xml_data.find('raw_text').text\n self.station = WeatherStation(self.xml_data.find('station_id').text)\n self.station.latitude = float(self.xml_data.find('latitude').text)\n self.station.longitude = float(self.xml_data.find('longitude').text)\n self.station.elevation = float(self.xml_data.find('elevation_m').text\n ) * 3.28084\n",
"role": "user"
},
{
"content": "parse_xml_data",
"role": "assistant"
}
] |
def x(self, mock_response=None):
if mock_response is not None:
body = mock_response
else:
api_url = self.get_api_url()
body = urlopen(api_url).read()
xml_root = ElementTree.fromstring(body)
xml_warnings = xml_root.find('warnings')
if len(xml_warnings.attrib) != 0:
print('Data warnings found: %s' % xml_warnings.attrib)
xml_errors = xml_root.find('errors')
if len(xml_errors.attrib) != 0:
raise Exception('Data errors found: %s' % xml_errors.attrib)
self.xml_data = xml_root.find('data')
| download_data | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, mock_response=None):\n if mock_response is not None:\n body = mock_response\n else:\n api_url = self.get_api_url()\n body = urlopen(api_url).read()\n xml_root = ElementTree.fromstring(body)\n xml_warnings = xml_root.find('warnings')\n if len(xml_warnings.attrib) != 0:\n print('Data warnings found: %s' % xml_warnings.attrib)\n xml_errors = xml_root.find('errors')\n if len(xml_errors.attrib) != 0:\n raise Exception('Data errors found: %s' % xml_errors.attrib)\n self.xml_data = xml_root.find('data')\n",
"role": "user"
},
{
"content": "download_data",
"role": "assistant"
}
] |
def x(file_handle, file_type):
base_counts = defaultdict(int)
read_number = 0
total_length = 0
length_list = []
records = SeqIO.parse(file_handle, file_type)
for record in records:
total_length += len(record)
read_number += 1
length_list.append(len(record))
for base in record:
base_counts[base] += 1
result = {'read_number': read_number, 'total_length': total_length,
'average_length': total_length / read_number if read_number > 0 else
0, 'max_length': max(length_list) if length_list else 0,
'min_length': min(length_list) if length_list else 0,
'median_length': median(length_list) if length_list else 0,
'base_counts': base_counts}
return result
| summarizeReads | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(file_handle, file_type):\n base_counts = defaultdict(int)\n read_number = 0\n total_length = 0\n length_list = []\n records = SeqIO.parse(file_handle, file_type)\n for record in records:\n total_length += len(record)\n read_number += 1\n length_list.append(len(record))\n for base in record:\n base_counts[base] += 1\n result = {'read_number': read_number, 'total_length': total_length,\n 'average_length': total_length / read_number if read_number > 0 else\n 0, 'max_length': max(length_list) if length_list else 0,\n 'min_length': min(length_list) if length_list else 0,\n 'median_length': median(length_list) if length_list else 0,\n 'base_counts': base_counts}\n return result\n",
"role": "user"
},
{
"content": "summarizeReads",
"role": "assistant"
}
] |
def x(read, categories, defaultCategory=None, suppressedCategory='...',
minLength=1):
result = []
append = result.append
get = categories.get
first = True
currentCategory = None
currentCount = 0
suppressing = False
suppressedCount = 0
if minLength < 1:
raise ValueError('minLength must be at least 1')
for base in read.sequence:
thisCategory = get(base, defaultCategory)
if first:
first = False
currentCategory = thisCategory
currentCount += 1
elif thisCategory == currentCategory:
currentCount += 1
else:
if currentCount < minLength:
if suppressing:
suppressedCount += currentCount
else:
suppressedCount = currentCount
suppressing = True
else:
if suppressing:
append((suppressedCategory, suppressedCount))
suppressedCount = 0
suppressing = False
append((currentCategory, currentCount))
currentCategory = thisCategory
currentCount = 1
if suppressing:
append((suppressedCategory, suppressedCount + currentCount))
elif currentCount >= minLength:
append((currentCategory, currentCount))
elif currentCount:
append((suppressedCategory, currentCount))
return result
| sequenceCategoryLengths | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(read, categories, defaultCategory=None, suppressedCategory='...',\n minLength=1):\n result = []\n append = result.append\n get = categories.get\n first = True\n currentCategory = None\n currentCount = 0\n suppressing = False\n suppressedCount = 0\n if minLength < 1:\n raise ValueError('minLength must be at least 1')\n for base in read.sequence:\n thisCategory = get(base, defaultCategory)\n if first:\n first = False\n currentCategory = thisCategory\n currentCount += 1\n elif thisCategory == currentCategory:\n currentCount += 1\n else:\n if currentCount < minLength:\n if suppressing:\n suppressedCount += currentCount\n else:\n suppressedCount = currentCount\n suppressing = True\n else:\n if suppressing:\n append((suppressedCategory, suppressedCount))\n suppressedCount = 0\n suppressing = False\n append((currentCategory, currentCount))\n currentCategory = thisCategory\n currentCount = 1\n if suppressing:\n append((suppressedCategory, suppressedCount + currentCount))\n elif currentCount >= minLength:\n append((currentCategory, currentCount))\n elif currentCount:\n append((suppressedCategory, currentCount))\n return result\n",
"role": "user"
},
{
"content": "sequenceCategoryLengths",
"role": "assistant"
}
] |
def x(title, target):
targetLen = len(target)
result = []
for word in title.split():
if len(word) >= targetLen:
offset = word.lower().find(target.lower())
if offset > -1:
result.append(word[:offset + targetLen])
break
result.append(word)
return ' '.join(result)
| simplifyTitle | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(title, target):\n targetLen = len(target)\n result = []\n for word in title.split():\n if len(word) >= targetLen:\n offset = word.lower().find(target.lower())\n if offset > -1:\n result.append(word[:offset + targetLen])\n break\n result.append(word)\n return ' '.join(result)\n",
"role": "user"
},
{
"content": "simplifyTitle",
"role": "assistant"
}
] |
def x(self, name=None, hdl=None, analyze_cmd=None, elaborate_cmd=None,
simulate_cmd=None):
if not isinstance(name, str) or name.strip() == '':
raise ValueError('Invalid _simulator name')
if hdl.lower() not in ('vhdl', 'verilog'):
raise ValueError('Invalid hdl {}'.format(hdl))
if not isinstance(analyze_cmd, str) or analyze_cmd.strip() == '':
raise ValueError('Invalid analyzer command')
if elaborate_cmd is not None:
if not isinstance(elaborate_cmd, str) or elaborate_cmd.strip() == '':
raise ValueError('Invalid elaborate_cmd command')
if not isinstance(simulate_cmd, str) or simulate_cmd.strip() == '':
raise ValueError('Invalid _simulator command')
self.sim_reg[name] = hdl.lower(), analyze_cmd, elaborate_cmd, simulate_cmd
| registerSimulator | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, name=None, hdl=None, analyze_cmd=None, elaborate_cmd=None,\n simulate_cmd=None):\n if not isinstance(name, str) or name.strip() == '':\n raise ValueError('Invalid _simulator name')\n if hdl.lower() not in ('vhdl', 'verilog'):\n raise ValueError('Invalid hdl {}'.format(hdl))\n if not isinstance(analyze_cmd, str) or analyze_cmd.strip() == '':\n raise ValueError('Invalid analyzer command')\n if elaborate_cmd is not None:\n if not isinstance(elaborate_cmd, str) or elaborate_cmd.strip() == '':\n raise ValueError('Invalid elaborate_cmd command')\n if not isinstance(simulate_cmd, str) or simulate_cmd.strip() == '':\n raise ValueError('Invalid _simulator command')\n self.sim_reg[name] = hdl.lower(), analyze_cmd, elaborate_cmd, simulate_cmd\n",
"role": "user"
},
{
"content": "registerSimulator",
"role": "assistant"
}
] |
def x(self, func, **kwargs):
vals = {}
vals['topname'] = func.func_name
vals['unitname'] = func.func_name.lower()
hdlsim = self._simulator
if not hdlsim:
raise ValueError('No _simulator specified')
if not self.sim_reg.has_key(hdlsim):
raise ValueError('Simulator {} is not registered'.format(hdlsim))
hdl, analyze_cmd, elaborate_cmd, simulate_cmd = self.sim_reg[hdlsim]
if hdl == 'verilog':
toVerilog(func, **kwargs)
if self._trace:
self._enableTracesVerilog('./tb_{topname}.v'.format(**vals))
elif hdl == 'vhdl':
toVHDL(func, **kwargs)
os.system(analyze_cmd.format(**vals))
if elaborate_cmd:
os.system(elaborate_cmd.format(**vals))
return Cosimulation(simulate_cmd.format(**vals), **kwargs)
| getCosimulation | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, func, **kwargs):\n vals = {}\n vals['topname'] = func.func_name\n vals['unitname'] = func.func_name.lower()\n hdlsim = self._simulator\n if not hdlsim:\n raise ValueError('No _simulator specified')\n if not self.sim_reg.has_key(hdlsim):\n raise ValueError('Simulator {} is not registered'.format(hdlsim))\n hdl, analyze_cmd, elaborate_cmd, simulate_cmd = self.sim_reg[hdlsim]\n if hdl == 'verilog':\n toVerilog(func, **kwargs)\n if self._trace:\n self._enableTracesVerilog('./tb_{topname}.v'.format(**vals))\n elif hdl == 'vhdl':\n toVHDL(func, **kwargs)\n os.system(analyze_cmd.format(**vals))\n if elaborate_cmd:\n os.system(elaborate_cmd.format(**vals))\n return Cosimulation(simulate_cmd.format(**vals), **kwargs)\n",
"role": "user"
},
{
"content": "getCosimulation",
"role": "assistant"
}
] |
def x(self, verilogFile):
fname, _ = os.path.splitext(verilogFile)
inserted = False
for _, line in enumerate(fileinput.input(verilogFile, inplace=1)):
sys.stdout.write(line)
if line.startswith('end') and not inserted:
sys.stdout.write('\n\n')
sys.stdout.write('initial begin\n')
sys.stdout.write(' $dumpfile("{}_cosim.vcd");\n'.format(fname))
sys.stdout.write(' $dumpvars(0, dut);\n')
sys.stdout.write('end\n\n')
inserted = True
| enableTracesVerilog | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, verilogFile):\n fname, _ = os.path.splitext(verilogFile)\n inserted = False\n for _, line in enumerate(fileinput.input(verilogFile, inplace=1)):\n sys.stdout.write(line)\n if line.startswith('end') and not inserted:\n sys.stdout.write('\\n\\n')\n sys.stdout.write('initial begin\\n')\n sys.stdout.write(' $dumpfile(\"{}_cosim.vcd\");\\n'.format(fname))\n sys.stdout.write(' $dumpvars(0, dut);\\n')\n sys.stdout.write('end\\n\\n')\n inserted = True\n",
"role": "user"
},
{
"content": "enableTracesVerilog",
"role": "assistant"
}
] |
def x(self, func, **kwargs):
if self._simulator == 'myhdl':
if not self._trace:
sim_dut = func(**kwargs)
else:
sim_dut = traceSignals(func, **kwargs)
else:
sim_dut = self._getCosimulation(func, **kwargs)
return sim_dut
| getDut | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, func, **kwargs):\n if self._simulator == 'myhdl':\n if not self._trace:\n sim_dut = func(**kwargs)\n else:\n sim_dut = traceSignals(func, **kwargs)\n else:\n sim_dut = self._getCosimulation(func, **kwargs)\n return sim_dut\n",
"role": "user"
},
{
"content": "getDut",
"role": "assistant"
}
] |
def x(self):
for _file in self._files:
with asHandle(_file) as fp:
for sequenceId, sequence, quality in FastqGeneralIterator(fp):
yield self.readClass(sequenceId, sequence, quality)
| iter | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n for _file in self._files:\n with asHandle(_file) as fp:\n for sequenceId, sequence, quality in FastqGeneralIterator(fp):\n yield self.readClass(sequenceId, sequence, quality)\n",
"role": "user"
},
{
"content": "iter",
"role": "assistant"
}
] |
def x(params):
assert params['fastaFile'][-1
] == 'a', 'You must specify a file in fasta-format that ends in .fasta'
fileCount = count = seqCount = 0
outfp = None
with open(params['fastaFile']) as infp:
for seq in SeqIO.parse(infp, 'fasta'):
seqCount += 1
if count == params['seqsPerJob']:
outfp.close()
count = 0
if count == 0:
outfp = open('%d.fasta' % fileCount, 'w')
fileCount += 1
count += 1
outfp.write('>%s\n%s\n' % (seq.description, str(seq.seq)))
outfp.close()
return fileCount, seqCount
| splitFASTA | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(params):\n assert params['fastaFile'][-1\n ] == 'a', 'You must specify a file in fasta-format that ends in .fasta'\n fileCount = count = seqCount = 0\n outfp = None\n with open(params['fastaFile']) as infp:\n for seq in SeqIO.parse(infp, 'fasta'):\n seqCount += 1\n if count == params['seqsPerJob']:\n outfp.close()\n count = 0\n if count == 0:\n outfp = open('%d.fasta' % fileCount, 'w')\n fileCount += 1\n count += 1\n outfp.write('>%s\\n%s\\n' % (seq.description, str(seq.seq)))\n outfp.close()\n return fileCount, seqCount\n",
"role": "user"
},
{
"content": "splitFASTA",
"role": "assistant"
}
] |
def x(self, task):
if not self._valid_name(task.name):
raise ValueError(task.name)
self._tasks[task.name] = task
incomplete_dependencies = set()
for dependency in task.dependencies:
if not self._valid_name(dependency) or dependency in self._failed:
self._cascade_failure(task.name)
break
if dependency not in self._completed:
incomplete_dependencies.add(dependency)
else:
try:
self._graph.add(task.name, incomplete_dependencies)
except ValueError:
self._cascade_failure(task.name)
| add_task | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, task):\n if not self._valid_name(task.name):\n raise ValueError(task.name)\n self._tasks[task.name] = task\n incomplete_dependencies = set()\n for dependency in task.dependencies:\n if not self._valid_name(dependency) or dependency in self._failed:\n self._cascade_failure(task.name)\n break\n if dependency not in self._completed:\n incomplete_dependencies.add(dependency)\n else:\n try:\n self._graph.add(task.name, incomplete_dependencies)\n except ValueError:\n self._cascade_failure(task.name)\n",
"role": "user"
},
{
"content": "add_task",
"role": "assistant"
}
] |
def x(self, name=None):
if name is None:
for possibility in self._graph.roots:
if possibility not in self._running:
name = possibility
break
else:
return None
elif name not in self._graph.roots or name in self._running:
raise ValueError(name)
self._running.add(name)
return self._tasks[name]
| start_task | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, name=None):\n if name is None:\n for possibility in self._graph.roots:\n if possibility not in self._running:\n name = possibility\n break\n else:\n return None\n elif name not in self._graph.roots or name in self._running:\n raise ValueError(name)\n self._running.add(name)\n return self._tasks[name]\n",
"role": "user"
},
{
"content": "start_task",
"role": "assistant"
}
] |
def x(self, name, success=True):
self._running.remove(name)
if success:
self._completed.add(name)
self._graph.remove(name, strategy=Strategy.orphan)
else:
self._cascade_failure(name)
| end_task | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, name, success=True):\n self._running.remove(name)\n if success:\n self._completed.add(name)\n self._graph.remove(name, strategy=Strategy.orphan)\n else:\n self._cascade_failure(name)\n",
"role": "user"
},
{
"content": "end_task",
"role": "assistant"
}
] |
def x(self):
self._failed.update(self._graph.nodes)
self._graph = Graph()
self._running = set()
| fail_remaining | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n self._failed.update(self._graph.nodes)\n self._graph = Graph()\n self._running = set()\n",
"role": "user"
},
{
"content": "fail_remaining",
"role": "assistant"
}
] |
def x(self, name):
if name in self._graph:
self._failed.update(self._graph.remove(name, strategy=Strategy.remove))
else:
self._failed.add(name)
| cascade_failure | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, name):\n if name in self._graph:\n self._failed.update(self._graph.remove(name, strategy=Strategy.remove))\n else:\n self._failed.add(name)\n",
"role": "user"
},
{
"content": "cascade_failure",
"role": "assistant"
}
] |
def x(gi, ranges):
record = getSequence(gi)
if record is None:
print("Looks like you're offline.")
sys.exit(3)
else:
printed = set()
if ranges:
for start, end in ranges:
for index, feature in enumerate(record.features):
if start < int(feature.location.end) and end > int(feature
.location.start) and index not in printed:
print(feature)
printed.add(index)
else:
for feature in record.features:
print(feature)
| main | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(gi, ranges):\n record = getSequence(gi)\n if record is None:\n print(\"Looks like you're offline.\")\n sys.exit(3)\n else:\n printed = set()\n if ranges:\n for start, end in ranges:\n for index, feature in enumerate(record.features):\n if start < int(feature.location.end) and end > int(feature\n .location.start) and index not in printed:\n print(feature)\n printed.add(index)\n else:\n for feature in record.features:\n print(feature)\n",
"role": "user"
},
{
"content": "main",
"role": "assistant"
}
] |
def x(credentials, project, zone, name, startup_script=None,
startup_script_url=None, metadata=None, machine_type='f1-micro', tags=
None, disk_size_gb=10, wait_until_done=False):
if startup_script is not None and startup_script_url is not None:
raise ValueError(
'Cannot specify a startup script string and URL at the same time!')
access_token = credentials.get_access_token()
if metadata is None:
metadata = {}
meta_items = [{'key': k, 'value': v} for k, v in metadata.items()]
if tags is None:
tags = []
if startup_script is not None:
meta_items.insert(0, {'key': 'startup-script', 'value': startup_script}
)
elif startup_script_url is not None:
meta_items.insert(0, {'key': 'startup-script-url', 'value':
startup_script_url})
payload = {'name': name, 'zone': 'projects/%s/zones/%s' % (project,
zone), 'machineType': 'projects/%s/zones/%s/machineTypes/%s' % (
project, zone, machine_type), 'metadata': {'items': meta_items},
'tags': {'items': tags}, 'disks': [{'type': 'PERSISTENT', 'boot':
True, 'mode': 'READ_WRITE', 'autoDelete': True, 'deviceName': name,
'initializeParams': {'sourceImage':
'projects/ubuntu-os-cloud/global/images/ubuntu-1604-xenial-v20170815a',
'diskType': 'projects/%s/zones/%s/diskTypes/pd-standard' % (project,
zone), 'diskSizeGb': str(disk_size_gb)}}], 'canIpForward': False,
'networkInterfaces': [{'network':
'projects/%s/global/networks/default' % project, 'subnetwork':
'projects/%s/regions/%s/subnetworks/default' % (project, zone[:-2]),
'accessConfigs': [{'name': 'External NAT', 'type': 'ONE_TO_ONE_NAT'
}]}], 'description': '', 'scheduling': {'preemptible': False,
'onHostMaintenance': 'MIGRATE', 'automaticRestart': True},
'serviceAccounts': [{'email': 'default', 'scopes': [
'https://www.googleapis.com/auth/compute',
'https://www.googleapis.com/auth/devstorage.read_write',
'https://www.googleapis.com/auth/logging.write',
'https://www.googleapis.com/auth/monitoring.write',
'https://www.googleapis.com/auth/servicecontrol',
'https://www.googleapis.com/auth/service.management.readonly',
'https://www.googleapis.com/auth/trace.append']}]}
headers = {'Authorization': 'Bearer %s' % access_token.access_token}
_LOGGER.debug('Access token: %s' % access_token.access_token)
_LOGGER.debug('Payload: %s', json.dumps(payload, sort_keys=True, indent=4))
r = requests.post(
'https://www.googleapis.com/compute/v1/projects/%s/zones/%s/instances'
% (project, zone), headers=headers, json=payload)
r.raise_for_status()
op_name = r.json()['name']
_LOGGER.info('Submitted request to create intsance (HTTP code: %d).', r
.status_code)
if wait_until_done:
_LOGGER.info('Waiting until operation is done...')
wait_for_zone_op(access_token, project, zone, op_name)
return op_name
| create_instance | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(credentials, project, zone, name, startup_script=None,\n startup_script_url=None, metadata=None, machine_type='f1-micro', tags=\n None, disk_size_gb=10, wait_until_done=False):\n if startup_script is not None and startup_script_url is not None:\n raise ValueError(\n 'Cannot specify a startup script string and URL at the same time!')\n access_token = credentials.get_access_token()\n if metadata is None:\n metadata = {}\n meta_items = [{'key': k, 'value': v} for k, v in metadata.items()]\n if tags is None:\n tags = []\n if startup_script is not None:\n meta_items.insert(0, {'key': 'startup-script', 'value': startup_script}\n )\n elif startup_script_url is not None:\n meta_items.insert(0, {'key': 'startup-script-url', 'value':\n startup_script_url})\n payload = {'name': name, 'zone': 'projects/%s/zones/%s' % (project,\n zone), 'machineType': 'projects/%s/zones/%s/machineTypes/%s' % (\n project, zone, machine_type), 'metadata': {'items': meta_items},\n 'tags': {'items': tags}, 'disks': [{'type': 'PERSISTENT', 'boot': \n True, 'mode': 'READ_WRITE', 'autoDelete': True, 'deviceName': name,\n 'initializeParams': {'sourceImage':\n 'projects/ubuntu-os-cloud/global/images/ubuntu-1604-xenial-v20170815a',\n 'diskType': 'projects/%s/zones/%s/diskTypes/pd-standard' % (project,\n zone), 'diskSizeGb': str(disk_size_gb)}}], 'canIpForward': False,\n 'networkInterfaces': [{'network': \n 'projects/%s/global/networks/default' % project, 'subnetwork': \n 'projects/%s/regions/%s/subnetworks/default' % (project, zone[:-2]),\n 'accessConfigs': [{'name': 'External NAT', 'type': 'ONE_TO_ONE_NAT'\n }]}], 'description': '', 'scheduling': {'preemptible': False,\n 'onHostMaintenance': 'MIGRATE', 'automaticRestart': True},\n 'serviceAccounts': [{'email': 'default', 'scopes': [\n 'https://www.googleapis.com/auth/compute',\n 'https://www.googleapis.com/auth/devstorage.read_write',\n 'https://www.googleapis.com/auth/logging.write',\n 'https://www.googleapis.com/auth/monitoring.write',\n 'https://www.googleapis.com/auth/servicecontrol',\n 'https://www.googleapis.com/auth/service.management.readonly',\n 'https://www.googleapis.com/auth/trace.append']}]}\n headers = {'Authorization': 'Bearer %s' % access_token.access_token}\n _LOGGER.debug('Access token: %s' % access_token.access_token)\n _LOGGER.debug('Payload: %s', json.dumps(payload, sort_keys=True, indent=4))\n r = requests.post(\n 'https://www.googleapis.com/compute/v1/projects/%s/zones/%s/instances'\n % (project, zone), headers=headers, json=payload)\n r.raise_for_status()\n op_name = r.json()['name']\n _LOGGER.info('Submitted request to create intsance (HTTP code: %d).', r\n .status_code)\n if wait_until_done:\n _LOGGER.info('Waiting until operation is done...')\n wait_for_zone_op(access_token, project, zone, op_name)\n return op_name\n",
"role": "user"
},
{
"content": "create_instance",
"role": "assistant"
}
] |
def x(credentials, project, zone, name, wait_until_done=False):
access_token = credentials.get_access_token()
headers = {'Authorization': 'Bearer %s' % access_token.access_token}
r = requests.delete(
'https://www.googleapis.com/compute/v1/projects/%s/zones/%s/instances/%s'
% (project, zone, name), headers=headers)
r.raise_for_status()
op_name = r.json()['name']
_LOGGER.info('Submitted request to create intsance (HTTP code: %d).', r
.status_code)
if wait_until_done:
_LOGGER.info('Waiting until operation is done...')
wait_for_zone_op(access_token, project, zone, op_name)
return op_name
| delete_instance | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(credentials, project, zone, name, wait_until_done=False):\n access_token = credentials.get_access_token()\n headers = {'Authorization': 'Bearer %s' % access_token.access_token}\n r = requests.delete(\n 'https://www.googleapis.com/compute/v1/projects/%s/zones/%s/instances/%s'\n % (project, zone, name), headers=headers)\n r.raise_for_status()\n op_name = r.json()['name']\n _LOGGER.info('Submitted request to create intsance (HTTP code: %d).', r\n .status_code)\n if wait_until_done:\n _LOGGER.info('Waiting until operation is done...')\n wait_for_zone_op(access_token, project, zone, op_name)\n return op_name\n",
"role": "user"
},
{
"content": "delete_instance",
"role": "assistant"
}
] |
def x(credentials, project, zone, instance_name, interval_seconds=5):
t0 = time.time()
access_token = credentials.get_access_token()
headers = {'Authorization': 'Bearer %s' % access_token.access_token}
r = requests.get(
'https://www.googleapis.com/compute/v1/projects/%s/zones/%s/instances/%s'
% (project, zone, instance_name), headers=headers)
if r.status_code == 404:
raise AssertionError('Instance "%s" does not exist!' % instance_name)
r.raise_for_status()
_LOGGER.debug('Instance "%s" exists.', instance_name)
while True:
time.sleep(interval_seconds)
access_token = credentials.get_access_token()
headers = {'Authorization': 'Bearer %s' % access_token.access_token}
r = requests.get(
'https://www.googleapis.com/compute/v1/projects/%s/zones/%s/instances/%s'
% (project, zone, instance_name), headers=headers)
if r.status_code == 404:
break
r.raise_for_status()
_LOGGER.debug('Instance "%s" still exists.', instance_name)
t1 = time.time()
t = t1 - t0
t_min = t / 60.0
_LOGGER.info('Instance was deleted after %.1f s (%.1f m).', t, t_min)
| wait_for_instance_deletion | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(credentials, project, zone, instance_name, interval_seconds=5):\n t0 = time.time()\n access_token = credentials.get_access_token()\n headers = {'Authorization': 'Bearer %s' % access_token.access_token}\n r = requests.get(\n 'https://www.googleapis.com/compute/v1/projects/%s/zones/%s/instances/%s'\n % (project, zone, instance_name), headers=headers)\n if r.status_code == 404:\n raise AssertionError('Instance \"%s\" does not exist!' % instance_name)\n r.raise_for_status()\n _LOGGER.debug('Instance \"%s\" exists.', instance_name)\n while True:\n time.sleep(interval_seconds)\n access_token = credentials.get_access_token()\n headers = {'Authorization': 'Bearer %s' % access_token.access_token}\n r = requests.get(\n 'https://www.googleapis.com/compute/v1/projects/%s/zones/%s/instances/%s'\n % (project, zone, instance_name), headers=headers)\n if r.status_code == 404:\n break\n r.raise_for_status()\n _LOGGER.debug('Instance \"%s\" still exists.', instance_name)\n t1 = time.time()\n t = t1 - t0\n t_min = t / 60.0\n _LOGGER.info('Instance was deleted after %.1f s (%.1f m).', t, t_min)\n",
"role": "user"
},
{
"content": "wait_for_instance_deletion",
"role": "assistant"
}
] |
def x(filenameOrSamfile, indent=0):
indent = ' ' * indent
def x(sam):
result = []
for i in range(sam.nreferences):
result.append('%s%s (length %d)' % (indent, sam.
get_reference_name(i), sam.lengths[i]))
return '\n'.join(result)
if isinstance(filenameOrSamfile, six.string_types):
with samfile(filenameOrSamfile) as sam:
return _references(sam)
else:
return _references(sam)
| samReferencesToStr | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(filenameOrSamfile, indent=0):\n indent = ' ' * indent\n\n def x(sam):\n result = []\n for i in range(sam.nreferences):\n result.append('%s%s (length %d)' % (indent, sam.\n get_reference_name(i), sam.lengths[i]))\n return '\\n'.join(result)\n if isinstance(filenameOrSamfile, six.string_types):\n with samfile(filenameOrSamfile) as sam:\n return _references(sam)\n else:\n return _references(sam)\n",
"role": "user"
},
{
"content": "samReferencesToStr",
"role": "assistant"
}
] |
def x(sequence, quality, cigartuples):
hardClipCount = cigarLength = 0
for operation, length in cigartuples:
hardClipCount += operation == CHARD_CLIP
cigarLength += length if operation in _CONSUMES_QUERY else 0
sequenceLength = len(sequence)
if quality is not None:
assert sequenceLength == len(quality)
clipLeft = clipRight = 0
clippedSequence = sequence
clippedQuality = quality
if sequenceLength > cigarLength:
alreadyClipped = False
else:
assert sequenceLength == cigarLength
alreadyClipped = True
if hardClipCount == 0:
pass
elif hardClipCount == 1:
if cigartuples[0][0] == CHARD_CLIP:
if not alreadyClipped:
clipLeft = cigartuples[0][1]
clippedSequence = sequence[clipLeft:]
if quality is not None:
clippedQuality = quality[clipLeft:]
elif cigartuples[-1][0] == CHARD_CLIP:
if not alreadyClipped:
clipRight = cigartuples[-1][1]
clippedSequence = sequence[:-clipRight]
if quality is not None:
clippedQuality = quality[:-clipRight]
else:
raise ValueError(
'Invalid CIGAR tuples (%s) contains hard-clipping operation that is neither at the start nor the end of the sequence.'
% (cigartuples,))
elif hardClipCount == 2:
assert cigartuples[0][0] == cigartuples[-1][0] == CHARD_CLIP
if not alreadyClipped:
clipLeft, clipRight = cigartuples[0][1], cigartuples[-1][1]
clippedSequence = sequence[clipLeft:-clipRight]
if quality is not None:
clippedQuality = quality[clipLeft:-clipRight]
else:
raise ValueError(
'Invalid CIGAR tuples (%s) specifies hard-clipping %d times (2 is the maximum).'
% (cigartuples, hardClipCount))
weClipped = bool(clipLeft or clipRight)
if weClipped:
assert not alreadyClipped
if len(clippedSequence) + clipLeft + clipRight != sequenceLength:
raise ValueError(
'Sequence %r (length %d) clipped to %r (length %d), but the difference between these two lengths (%d) is not equal to the sum (%d) of the left and right clip lengths (%d and %d respectively). CIGAR tuples: %s'
% (sequence, len(sequence), clippedSequence, len(
clippedSequence), abs(len(sequence) - len(clippedSequence)),
clipLeft + clipRight, clipLeft, clipRight, cigartuples))
else:
assert len(clippedSequence) == sequenceLength
if quality is not None:
assert len(clippedQuality) == sequenceLength
return clippedSequence, clippedQuality, weClipped
| hardClip | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(sequence, quality, cigartuples):\n hardClipCount = cigarLength = 0\n for operation, length in cigartuples:\n hardClipCount += operation == CHARD_CLIP\n cigarLength += length if operation in _CONSUMES_QUERY else 0\n sequenceLength = len(sequence)\n if quality is not None:\n assert sequenceLength == len(quality)\n clipLeft = clipRight = 0\n clippedSequence = sequence\n clippedQuality = quality\n if sequenceLength > cigarLength:\n alreadyClipped = False\n else:\n assert sequenceLength == cigarLength\n alreadyClipped = True\n if hardClipCount == 0:\n pass\n elif hardClipCount == 1:\n if cigartuples[0][0] == CHARD_CLIP:\n if not alreadyClipped:\n clipLeft = cigartuples[0][1]\n clippedSequence = sequence[clipLeft:]\n if quality is not None:\n clippedQuality = quality[clipLeft:]\n elif cigartuples[-1][0] == CHARD_CLIP:\n if not alreadyClipped:\n clipRight = cigartuples[-1][1]\n clippedSequence = sequence[:-clipRight]\n if quality is not None:\n clippedQuality = quality[:-clipRight]\n else:\n raise ValueError(\n 'Invalid CIGAR tuples (%s) contains hard-clipping operation that is neither at the start nor the end of the sequence.'\n % (cigartuples,))\n elif hardClipCount == 2:\n assert cigartuples[0][0] == cigartuples[-1][0] == CHARD_CLIP\n if not alreadyClipped:\n clipLeft, clipRight = cigartuples[0][1], cigartuples[-1][1]\n clippedSequence = sequence[clipLeft:-clipRight]\n if quality is not None:\n clippedQuality = quality[clipLeft:-clipRight]\n else:\n raise ValueError(\n 'Invalid CIGAR tuples (%s) specifies hard-clipping %d times (2 is the maximum).'\n % (cigartuples, hardClipCount))\n weClipped = bool(clipLeft or clipRight)\n if weClipped:\n assert not alreadyClipped\n if len(clippedSequence) + clipLeft + clipRight != sequenceLength:\n raise ValueError(\n 'Sequence %r (length %d) clipped to %r (length %d), but the difference between these two lengths (%d) is not equal to the sum (%d) of the left and right clip lengths (%d and %d respectively). CIGAR tuples: %s'\n % (sequence, len(sequence), clippedSequence, len(\n clippedSequence), abs(len(sequence) - len(clippedSequence)),\n clipLeft + clipRight, clipLeft, clipRight, cigartuples))\n else:\n assert len(clippedSequence) == sequenceLength\n if quality is not None:\n assert len(clippedQuality) == sequenceLength\n return clippedSequence, clippedQuality, weClipped\n",
"role": "user"
},
{
"content": "hardClip",
"role": "assistant"
}
] |
def x(parser, samfileIsPositionalArg=False):
parser.add_argument('%ssamfile' % ('' if samfileIsPositionalArg else
'--'), required=True, help='The SAM/BAM file to filter.')
parser.add_argument('--referenceId', metavar='ID', nargs='+', action=
'append', help=
'A reference sequence id whose alignments should be kept (alignments against other references will be dropped). If omitted, alignments against all references will be kept. May be repeated.'
)
parser.add_argument('--dropUnmapped', default=False, action=
'store_true', help='If given, unmapped matches will not be output.')
parser.add_argument('--dropSecondary', default=False, action=
'store_true', help='If given, secondary matches will not be output.')
parser.add_argument('--dropSupplementary', default=False, action=
'store_true', help=
'If given, supplementary matches will not be output.')
parser.add_argument('--dropDuplicates', default=False, action=
'store_true', help=
'If given, matches flagged as optical or PCR duplicates will not be output.'
)
parser.add_argument('--keepQCFailures', default=False, action=
'store_true', help=
'If given, reads that are considered quality control failures will be included in the output.'
)
parser.add_argument('--minScore', type=float, help=
'If given, alignments with --scoreTag (default AS) values less than this value will not be output. If given, alignments that do not have a score will not be output.'
)
parser.add_argument('--maxScore', type=float, help=
'If given, alignments with --scoreTag (default AS) values greater than this value will not be output. If given, alignments that do not have a score will not be output.'
)
parser.add_argument('--scoreTag', default='AS', help=
'The alignment tag to extract for --minScore and --maxScore comparisons.'
)
| addFilteringOptions | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(parser, samfileIsPositionalArg=False):\n parser.add_argument('%ssamfile' % ('' if samfileIsPositionalArg else\n '--'), required=True, help='The SAM/BAM file to filter.')\n parser.add_argument('--referenceId', metavar='ID', nargs='+', action=\n 'append', help=\n 'A reference sequence id whose alignments should be kept (alignments against other references will be dropped). If omitted, alignments against all references will be kept. May be repeated.'\n )\n parser.add_argument('--dropUnmapped', default=False, action=\n 'store_true', help='If given, unmapped matches will not be output.')\n parser.add_argument('--dropSecondary', default=False, action=\n 'store_true', help='If given, secondary matches will not be output.')\n parser.add_argument('--dropSupplementary', default=False, action=\n 'store_true', help=\n 'If given, supplementary matches will not be output.')\n parser.add_argument('--dropDuplicates', default=False, action=\n 'store_true', help=\n 'If given, matches flagged as optical or PCR duplicates will not be output.'\n )\n parser.add_argument('--keepQCFailures', default=False, action=\n 'store_true', help=\n 'If given, reads that are considered quality control failures will be included in the output.'\n )\n parser.add_argument('--minScore', type=float, help=\n 'If given, alignments with --scoreTag (default AS) values less than this value will not be output. If given, alignments that do not have a score will not be output.'\n )\n parser.add_argument('--maxScore', type=float, help=\n 'If given, alignments with --scoreTag (default AS) values greater than this value will not be output. If given, alignments that do not have a score will not be output.'\n )\n parser.add_argument('--scoreTag', default='AS', help=\n 'The alignment tag to extract for --minScore and --maxScore comparisons.'\n )\n",
"role": "user"
},
{
"content": "addFilteringOptions",
"role": "assistant"
}
] |
def x(cls, args, filterRead=None, storeQueryIds=False):
referenceIds = set(chain.from_iterable(args.referenceId)
) if args.referenceId else None
return cls(args.samfile, filterRead=filterRead, referenceIds=
referenceIds, storeQueryIds=storeQueryIds, dropUnmapped=args.
dropUnmapped, dropSecondary=args.dropSecondary, dropSupplementary=
args.dropSupplementary, dropDuplicates=args.dropDuplicates,
keepQCFailures=args.keepQCFailures, minScore=args.minScore,
maxScore=args.maxScore)
| parseFilteringOptions | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, args, filterRead=None, storeQueryIds=False):\n referenceIds = set(chain.from_iterable(args.referenceId)\n ) if args.referenceId else None\n return cls(args.samfile, filterRead=filterRead, referenceIds=\n referenceIds, storeQueryIds=storeQueryIds, dropUnmapped=args.\n dropUnmapped, dropSecondary=args.dropSecondary, dropSupplementary=\n args.dropSupplementary, dropDuplicates=args.dropDuplicates,\n keepQCFailures=args.keepQCFailures, minScore=args.minScore,\n maxScore=args.maxScore)\n",
"role": "user"
},
{
"content": "parseFilteringOptions",
"role": "assistant"
}
] |
def x(self):
referenceIds = self.referenceIds
dropUnmapped = self.dropUnmapped
dropSecondary = self.dropSecondary
dropSupplementary = self.dropSupplementary
dropDuplicates = self.dropDuplicates
keepQCFailures = self.keepQCFailures
storeQueryIds = self.storeQueryIds
filterRead = self.filterRead
minScore = self.minScore
maxScore = self.maxScore
scoreTag = self.scoreTag
if storeQueryIds:
self.queryIds = queryIds = set()
lastAlignment = None
count = 0
with samfile(self.filename) as samAlignment:
for count, alignment in enumerate(samAlignment.fetch(), start=1):
if storeQueryIds:
queryIds.add(alignment.query_name)
if minScore is not None or maxScore is not None:
try:
score = alignment.get_tag(scoreTag)
except KeyError:
continue
else:
if (minScore is not None and score < minScore or
maxScore is not None and score > maxScore):
continue
if alignment.query_sequence is None:
if lastAlignment is None:
raise InvalidSAM(
'pysam produced an alignment (number %d) with no query sequence without previously giving an alignment with a sequence.'
% count)
assert alignment.cigartuples
alignment.query_sequence, alignment.query_qualities, _ = (
_hardClip(lastAlignment.query_sequence, lastAlignment.
query_qualities, alignment.cigartuples))
else:
lastAlignment = alignment
if alignment.cigartuples:
(alignment.query_sequence, alignment.query_qualities, _
) = (_hardClip(alignment.query_sequence, alignment.
query_qualities, alignment.cigartuples))
if (filterRead is None or filterRead(Read(alignment.query_name,
alignment.query_sequence, alignment.qual))) and not (
referenceIds and alignment.reference_name not in
referenceIds or alignment.is_unmapped and dropUnmapped or
alignment.is_secondary and dropSecondary or alignment.
is_supplementary and dropSupplementary or alignment.
is_duplicate and dropDuplicates or alignment.is_qcfail and
not keepQCFailures):
yield alignment
self.alignmentCount = count
| alignments | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n referenceIds = self.referenceIds\n dropUnmapped = self.dropUnmapped\n dropSecondary = self.dropSecondary\n dropSupplementary = self.dropSupplementary\n dropDuplicates = self.dropDuplicates\n keepQCFailures = self.keepQCFailures\n storeQueryIds = self.storeQueryIds\n filterRead = self.filterRead\n minScore = self.minScore\n maxScore = self.maxScore\n scoreTag = self.scoreTag\n if storeQueryIds:\n self.queryIds = queryIds = set()\n lastAlignment = None\n count = 0\n with samfile(self.filename) as samAlignment:\n for count, alignment in enumerate(samAlignment.fetch(), start=1):\n if storeQueryIds:\n queryIds.add(alignment.query_name)\n if minScore is not None or maxScore is not None:\n try:\n score = alignment.get_tag(scoreTag)\n except KeyError:\n continue\n else:\n if (minScore is not None and score < minScore or \n maxScore is not None and score > maxScore):\n continue\n if alignment.query_sequence is None:\n if lastAlignment is None:\n raise InvalidSAM(\n 'pysam produced an alignment (number %d) with no query sequence without previously giving an alignment with a sequence.'\n % count)\n assert alignment.cigartuples\n alignment.query_sequence, alignment.query_qualities, _ = (\n _hardClip(lastAlignment.query_sequence, lastAlignment.\n query_qualities, alignment.cigartuples))\n else:\n lastAlignment = alignment\n if alignment.cigartuples:\n (alignment.query_sequence, alignment.query_qualities, _\n ) = (_hardClip(alignment.query_sequence, alignment.\n query_qualities, alignment.cigartuples))\n if (filterRead is None or filterRead(Read(alignment.query_name,\n alignment.query_sequence, alignment.qual))) and not (\n referenceIds and alignment.reference_name not in\n referenceIds or alignment.is_unmapped and dropUnmapped or \n alignment.is_secondary and dropSecondary or alignment.\n is_supplementary and dropSupplementary or alignment.\n is_duplicate and dropDuplicates or alignment.is_qcfail and \n not keepQCFailures):\n yield alignment\n self.alignmentCount = count\n",
"role": "user"
},
{
"content": "alignments",
"role": "assistant"
}
] |
def x(self):
result = {}
with samfile(self.filename) as sam:
if self.referenceIds:
for referenceId in self.referenceIds:
tid = sam.get_tid(referenceId)
if tid == -1:
raise UnknownReference(
'Reference %r is not present in the SAM/BAM file.' %
referenceId)
else:
result[referenceId] = sam.lengths[tid]
else:
result = dict(zip(sam.references, sam.lengths))
return result
| referenceLengths | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n result = {}\n with samfile(self.filename) as sam:\n if self.referenceIds:\n for referenceId in self.referenceIds:\n tid = sam.get_tid(referenceId)\n if tid == -1:\n raise UnknownReference(\n 'Reference %r is not present in the SAM/BAM file.' %\n referenceId)\n else:\n result[referenceId] = sam.lengths[tid]\n else:\n result = dict(zip(sam.references, sam.lengths))\n return result\n",
"role": "user"
},
{
"content": "referenceLengths",
"role": "assistant"
}
] |
def x(self, rcSuffix='', rcNeeded=False, padChar='-', queryInsertionChar=
'N', unknownQualityChar='!', allowDuplicateIds=False, addAlignment=False):
referenceLength = self.referenceLength
idCount = Counter()
MATCH_OPERATIONS = {CMATCH, CEQUAL, CDIFF}
for lineNumber, alignment in enumerate(self.samFilter.alignments(), start=1
):
query = alignment.query_sequence
quality = ''.join(chr(q + 33) for q in alignment.query_qualities)
if alignment.is_reverse:
if rcNeeded:
query = DNARead('id', query).reverseComplement().sequence
quality = quality[::-1]
if rcSuffix:
alignment.query_name += rcSuffix
if allowDuplicateIds:
queryId = alignment.query_name
else:
count = idCount[alignment.query_name]
idCount[alignment.query_name] += 1
queryId = alignment.query_name + ('' if count == 0 else '/%d' %
count)
referenceStart = alignment.reference_start
atStart = True
queryIndex = 0
referenceIndex = referenceStart
alignedSequence = ''
alignedQuality = ''
for operation, length in alignment.cigartuples:
if operation in MATCH_OPERATIONS:
atStart = False
alignedSequence += query[queryIndex:queryIndex + length]
alignedQuality += quality[queryIndex:queryIndex + length]
elif operation == CINS:
atStart = False
self.referenceInsertions[queryId].append((referenceIndex,
query[queryIndex:queryIndex + length]))
elif operation == CDEL:
atStart = False
alignedSequence += queryInsertionChar * length
alignedQuality += unknownQualityChar * length
elif operation == CREF_SKIP:
atStart = False
alignedSequence += queryInsertionChar * length
alignedQuality += unknownQualityChar * length
elif operation == CSOFT_CLIP:
if atStart:
unwantedLeft = length - referenceStart
if unwantedLeft > 0:
alignedSequence += query[queryIndex + unwantedLeft:
queryIndex + length]
alignedQuality += quality[queryIndex + unwantedLeft
:queryIndex + length]
referenceStart = 0
else:
referenceStart -= length
alignedSequence += query[queryIndex:queryIndex + length
]
alignedQuality += quality[queryIndex:queryIndex +
length]
else:
unwantedRight = referenceStart + len(alignedSequence
) + length - referenceLength
if unwantedRight > 0:
alignedSequence += query[queryIndex:queryIndex +
length - unwantedRight]
alignedQuality += quality[queryIndex:queryIndex +
length - unwantedRight]
else:
alignedSequence += query[queryIndex:queryIndex + length
]
alignedQuality += quality[queryIndex:queryIndex +
length]
elif operation == CHARD_CLIP:
pass
elif operation == CPAD:
atStart = False
else:
raise ValueError('Unknown CIGAR operation:', operation)
if operation in _CONSUMES_QUERY:
queryIndex += length
if operation in _CONSUMES_REFERENCE:
referenceIndex += length
if queryIndex != len(query):
raise ValueError(
'Query %r not fully consumed when parsing CIGAR string. Query %r (len %d), final query index %d, CIGAR: %r'
% (alignment.query_name, query, len(query), queryIndex,
alignment.cigartuples))
padRightLength = referenceLength - (referenceStart + len(
alignedSequence))
paddedSequence = (padChar * referenceStart + alignedSequence +
padChar * padRightLength)
paddedQuality = (unknownQualityChar * referenceStart +
alignedQuality + unknownQualityChar * padRightLength)
read = Read(queryId, paddedSequence, paddedQuality)
if addAlignment:
read.alignment = alignment
yield read
| queries | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, rcSuffix='', rcNeeded=False, padChar='-', queryInsertionChar=\n 'N', unknownQualityChar='!', allowDuplicateIds=False, addAlignment=False):\n referenceLength = self.referenceLength\n idCount = Counter()\n MATCH_OPERATIONS = {CMATCH, CEQUAL, CDIFF}\n for lineNumber, alignment in enumerate(self.samFilter.alignments(), start=1\n ):\n query = alignment.query_sequence\n quality = ''.join(chr(q + 33) for q in alignment.query_qualities)\n if alignment.is_reverse:\n if rcNeeded:\n query = DNARead('id', query).reverseComplement().sequence\n quality = quality[::-1]\n if rcSuffix:\n alignment.query_name += rcSuffix\n if allowDuplicateIds:\n queryId = alignment.query_name\n else:\n count = idCount[alignment.query_name]\n idCount[alignment.query_name] += 1\n queryId = alignment.query_name + ('' if count == 0 else '/%d' %\n count)\n referenceStart = alignment.reference_start\n atStart = True\n queryIndex = 0\n referenceIndex = referenceStart\n alignedSequence = ''\n alignedQuality = ''\n for operation, length in alignment.cigartuples:\n if operation in MATCH_OPERATIONS:\n atStart = False\n alignedSequence += query[queryIndex:queryIndex + length]\n alignedQuality += quality[queryIndex:queryIndex + length]\n elif operation == CINS:\n atStart = False\n self.referenceInsertions[queryId].append((referenceIndex,\n query[queryIndex:queryIndex + length]))\n elif operation == CDEL:\n atStart = False\n alignedSequence += queryInsertionChar * length\n alignedQuality += unknownQualityChar * length\n elif operation == CREF_SKIP:\n atStart = False\n alignedSequence += queryInsertionChar * length\n alignedQuality += unknownQualityChar * length\n elif operation == CSOFT_CLIP:\n if atStart:\n unwantedLeft = length - referenceStart\n if unwantedLeft > 0:\n alignedSequence += query[queryIndex + unwantedLeft:\n queryIndex + length]\n alignedQuality += quality[queryIndex + unwantedLeft\n :queryIndex + length]\n referenceStart = 0\n else:\n referenceStart -= length\n alignedSequence += query[queryIndex:queryIndex + length\n ]\n alignedQuality += quality[queryIndex:queryIndex +\n length]\n else:\n unwantedRight = referenceStart + len(alignedSequence\n ) + length - referenceLength\n if unwantedRight > 0:\n alignedSequence += query[queryIndex:queryIndex +\n length - unwantedRight]\n alignedQuality += quality[queryIndex:queryIndex +\n length - unwantedRight]\n else:\n alignedSequence += query[queryIndex:queryIndex + length\n ]\n alignedQuality += quality[queryIndex:queryIndex +\n length]\n elif operation == CHARD_CLIP:\n pass\n elif operation == CPAD:\n atStart = False\n else:\n raise ValueError('Unknown CIGAR operation:', operation)\n if operation in _CONSUMES_QUERY:\n queryIndex += length\n if operation in _CONSUMES_REFERENCE:\n referenceIndex += length\n if queryIndex != len(query):\n raise ValueError(\n 'Query %r not fully consumed when parsing CIGAR string. Query %r (len %d), final query index %d, CIGAR: %r'\n % (alignment.query_name, query, len(query), queryIndex,\n alignment.cigartuples))\n padRightLength = referenceLength - (referenceStart + len(\n alignedSequence))\n paddedSequence = (padChar * referenceStart + alignedSequence + \n padChar * padRightLength)\n paddedQuality = (unknownQualityChar * referenceStart +\n alignedQuality + unknownQualityChar * padRightLength)\n read = Read(queryId, paddedSequence, paddedQuality)\n if addAlignment:\n read.alignment = alignment\n yield read\n",
"role": "user"
},
{
"content": "queries",
"role": "assistant"
}
] |
def x(sel, ls_di, do):
N = len(ls_di)
@always_comb
def x():
do.next = 0
for i in range(N):
if i == sel:
do.next = ls_di[i]
return _mux
| mux | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(sel, ls_di, do):\n N = len(ls_di)\n\n @always_comb\n def x():\n do.next = 0\n for i in range(N):\n if i == sel:\n do.next = ls_di[i]\n return _mux\n",
"role": "user"
},
{
"content": "mux",
"role": "assistant"
}
] |
def x(sel, di, ls_do):
N = len(ls_do)
@always_comb
def x():
for i in range(N):
ls_do[i].next = 0
if i == sel:
ls_do[i].next = di
return _demux
| demux | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(sel, di, ls_do):\n N = len(ls_do)\n\n @always_comb\n def x():\n for i in range(N):\n ls_do[i].next = 0\n if i == sel:\n ls_do[i].next = di\n return _demux\n",
"role": "user"
},
{
"content": "demux",
"role": "assistant"
}
] |
def x(sel, lsls_di, ls_do):
N = len(ls_do)
lsls_in = [list(x) for x in zip(*lsls_di)]
return [mux(sel, lsls_in[i], ls_do[i]) for i in range(N)]
| ls_mux | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(sel, lsls_di, ls_do):\n N = len(ls_do)\n lsls_in = [list(x) for x in zip(*lsls_di)]\n return [mux(sel, lsls_in[i], ls_do[i]) for i in range(N)]\n",
"role": "user"
},
{
"content": "ls_mux",
"role": "assistant"
}
] |
def x(sel, ls_di, lsls_do):
N = len(ls_di)
lsls_out = [list(x) for x in zip(*lsls_do)]
return [demux(sel, ls_di[i], lsls_out[i]) for i in range(N)]
| ls_demux | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(sel, ls_di, lsls_do):\n N = len(ls_di)\n lsls_out = [list(x) for x in zip(*lsls_do)]\n return [demux(sel, ls_di[i], lsls_out[i]) for i in range(N)]\n",
"role": "user"
},
{
"content": "ls_demux",
"role": "assistant"
}
] |
def x(offset, bv_di, bv_do):
LEN_I = len(bv_di)
LEN_O = len(bv_do)
assert LEN_I >= LEN_O, 'bitslice_select: expects len(bv_di) >= len(bv_do), but len(bv_di)={}, len(bv_do)'.format(
LEN_I, LEN_O)
OFFSET_MAX = LEN_I - LEN_O + 1
@always_comb
def x():
bv_do.next = 0
for i in range(OFFSET_MAX):
if i == offset:
for b in range(LEN_O):
bv_do.next[b] = bv_di[i + b]
return _slice
| bitslice_select | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(offset, bv_di, bv_do):\n LEN_I = len(bv_di)\n LEN_O = len(bv_do)\n assert LEN_I >= LEN_O, 'bitslice_select: expects len(bv_di) >= len(bv_do), but len(bv_di)={}, len(bv_do)'.format(\n LEN_I, LEN_O)\n OFFSET_MAX = LEN_I - LEN_O + 1\n\n @always_comb\n def x():\n bv_do.next = 0\n for i in range(OFFSET_MAX):\n if i == offset:\n for b in range(LEN_O):\n bv_do.next[b] = bv_di[i + b]\n return _slice\n",
"role": "user"
},
{
"content": "bitslice_select",
"role": "assistant"
}
] |
def x(offset, bv_di, bv_do):
LEN_I = len(bv_di)
LEN_O = len(bv_do)
assert LEN_I % 8 == 0, 'byteslice_select: expects len(bv_di)=8*x, but len(bv_di)={} bits'.format(
LEN_I)
assert LEN_O % 8 == 0, 'byteslice_select: expects len(bv_do)=8*x, but len(bv_do)={} bits'.format(
LEN_O)
bit_offset = Signal(intbv(0)[len(offset) + 3:])
@always_comb
def x():
bit_offset.next = offset << 3
_slice = bitslice_select(bit_offset, bv_di, bv_do)
return _offset, _slice
| byteslice_select | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(offset, bv_di, bv_do):\n LEN_I = len(bv_di)\n LEN_O = len(bv_do)\n assert LEN_I % 8 == 0, 'byteslice_select: expects len(bv_di)=8*x, but len(bv_di)={} bits'.format(\n LEN_I)\n assert LEN_O % 8 == 0, 'byteslice_select: expects len(bv_do)=8*x, but len(bv_do)={} bits'.format(\n LEN_O)\n bit_offset = Signal(intbv(0)[len(offset) + 3:])\n\n @always_comb\n def x():\n bit_offset.next = offset << 3\n _slice = bitslice_select(bit_offset, bv_di, bv_do)\n return _offset, _slice\n",
"role": "user"
},
{
"content": "byteslice_select",
"role": "assistant"
}
] |
def x():
file_mv = cli.file_mv
desc = 'Find all runs (SRR..) associated with an SRA experiment (SRX...).'
parser = cli.get_argument_parser(desc=desc)
parser.add_argument('-e', '--experiment-file', type=str, required=True,
metavar=file_mv, help=
'File with SRA experiment IDs (starting with "SRX").')
parser.add_argument('-o', '--output-file', type=str, required=True,
metavar=file_mv, help='The output file.')
cli.add_reporting_args(parser)
return parser
| get_argument_parser | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n file_mv = cli.file_mv\n desc = 'Find all runs (SRR..) associated with an SRA experiment (SRX...).'\n parser = cli.get_argument_parser(desc=desc)\n parser.add_argument('-e', '--experiment-file', type=str, required=True,\n metavar=file_mv, help=\n 'File with SRA experiment IDs (starting with \"SRX\").')\n parser.add_argument('-o', '--output-file', type=str, required=True,\n metavar=file_mv, help='The output file.')\n cli.add_reporting_args(parser)\n return parser\n",
"role": "user"
},
{
"content": "get_argument_parser",
"role": "assistant"
}
] |
def x(args=None):
if args is None:
parser = get_argument_parser()
args = parser.parse_args()
experiment_file = args.experiment_file
output_file = args.output_file
host = 'ftp-trace.ncbi.nlm.nih.gov'
user = 'anonymous'
password = 'anonymous'
experiments = misc.read_single(experiment_file)
runs = []
with ftputil.FTPHost(host, user, password) as ftp_host:
for exp in experiments:
exp_dir = '/sra/sra-instant/reads/ByExp/sra/SRX/%s/%s/' % (exp[
:6], exp)
ftp_host.chdir(exp_dir)
run_folders = ftp_host.listdir(ftp_host.curdir)
for folder in run_folders:
files = ftp_host.listdir(folder)
assert len(files) == 1
runs.append((exp, folder))
with open(output_file, 'wb') as ofh:
writer = csv.writer(ofh, dialect='excel-tab', lineterminator=os.
linesep, quoting=csv.QUOTE_NONE)
for r in runs:
writer.writerow(r)
return 0
| main | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(args=None):\n if args is None:\n parser = get_argument_parser()\n args = parser.parse_args()\n experiment_file = args.experiment_file\n output_file = args.output_file\n host = 'ftp-trace.ncbi.nlm.nih.gov'\n user = 'anonymous'\n password = 'anonymous'\n experiments = misc.read_single(experiment_file)\n runs = []\n with ftputil.FTPHost(host, user, password) as ftp_host:\n for exp in experiments:\n exp_dir = '/sra/sra-instant/reads/ByExp/sra/SRX/%s/%s/' % (exp[\n :6], exp)\n ftp_host.chdir(exp_dir)\n run_folders = ftp_host.listdir(ftp_host.curdir)\n for folder in run_folders:\n files = ftp_host.listdir(folder)\n assert len(files) == 1\n runs.append((exp, folder))\n with open(output_file, 'wb') as ofh:\n writer = csv.writer(ofh, dialect='excel-tab', lineterminator=os.\n linesep, quoting=csv.QUOTE_NONE)\n for r in runs:\n writer.writerow(r)\n return 0\n",
"role": "user"
},
{
"content": "main",
"role": "assistant"
}
] |
def x(cls, query_obj, search4, model_attrib):
model = model_attrib.parent.class_
if isinstance(search4, str):
query_obj = query_obj.join(model).filter(model_attrib.like(search4))
elif isinstance(search4, int):
query_obj = query_obj.join(model).filter(model_attrib == search4)
elif isinstance(search4, Iterable):
query_obj = query_obj.join(model).filter(model_attrib.in_(search4))
return query_obj
| one_to_many_query | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, query_obj, search4, model_attrib):\n model = model_attrib.parent.class_\n if isinstance(search4, str):\n query_obj = query_obj.join(model).filter(model_attrib.like(search4))\n elif isinstance(search4, int):\n query_obj = query_obj.join(model).filter(model_attrib == search4)\n elif isinstance(search4, Iterable):\n query_obj = query_obj.join(model).filter(model_attrib.in_(search4))\n return query_obj\n",
"role": "user"
},
{
"content": "one_to_many_query",
"role": "assistant"
}
] |
def x(self, name=None, identifier=None, entry_name=None, limit=None, as_df=
False):
q = self.session.query(models.Keyword)
model_queries_config = (name, models.Keyword.name), (identifier, models
.Keyword.identifier)
q = self.get_model_queries(q, model_queries_config)
q = self.get_many_to_many_queries(q, ((entry_name, models.Keyword.
entries, models.Entry.name),))
return self._limit_and_df(q, limit, as_df)
| keyword | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, name=None, identifier=None, entry_name=None, limit=None, as_df=\n False):\n q = self.session.query(models.Keyword)\n model_queries_config = (name, models.Keyword.name), (identifier, models\n .Keyword.identifier)\n q = self.get_model_queries(q, model_queries_config)\n q = self.get_many_to_many_queries(q, ((entry_name, models.Keyword.\n entries, models.Entry.name),))\n return self._limit_and_df(q, limit, as_df)\n",
"role": "user"
},
{
"content": "keyword",
"role": "assistant"
}
] |
def x(self, name=None, dataset=None, recommended_full_name=None,
recommended_short_name=None, gene_name=None, taxid=None, accession=None,
organism_host=None, feature_type=None, function_=None, ec_number=None,
db_reference=None, alternative_name=None, disease_comment=None,
disease_name=None, tissue_specificity=None, pmid=None, keyword=None,
subcellular_location=None, tissue_in_reference=None, sequence=None,
limit=None, as_df=False):
q = self.session.query(models.Entry)
model_queries_config = (dataset, models.Entry.dataset), (name, models.
Entry.name), (recommended_full_name, models.Entry.recommended_full_name
), (recommended_short_name, models.Entry.recommended_short_name), (
gene_name, models.Entry.gene_name), (taxid, models.Entry.taxid)
q = self.get_model_queries(q, model_queries_config)
one_to_many_queries_config = (accession, models.Accession.accession), (
organism_host, models.OrganismHost.taxid), (feature_type, models.
Feature.type_), (function_, models.Function.text), (ec_number,
models.ECNumber.ec_number), (db_reference, models.DbReference.
identifier), (alternative_name, models.AlternativeFullName.name), (
disease_comment, models.DiseaseComment.comment), (tissue_specificity,
models.TissueSpecificity.comment), (sequence, models.Sequence.sequence)
q = self.get_one_to_many_queries(q, one_to_many_queries_config)
many_to_many_queries_config = (pmid, models.Entry.pmids, models.Pmid.pmid
), (keyword, models.Entry.keywords, models.Keyword.name), (
subcellular_location, models.Entry.subcellular_locations, models.
SubcellularLocation.location), (tissue_in_reference, models.Entry.
tissue_in_references, models.TissueInReference.tissue)
q = self.get_many_to_many_queries(q, many_to_many_queries_config)
if disease_name:
q = q.join(models.Entry.disease_comments).join(models.
DiseaseComment.disease)
if isinstance(disease_name, str):
q = q.filter(models.Disease.name.like(disease_name))
elif isinstance(disease_name, Iterable):
q = q.filter(models.Disease.name.in_(disease_name))
return self._limit_and_df(q, limit, as_df)
| entry | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, name=None, dataset=None, recommended_full_name=None,\n recommended_short_name=None, gene_name=None, taxid=None, accession=None,\n organism_host=None, feature_type=None, function_=None, ec_number=None,\n db_reference=None, alternative_name=None, disease_comment=None,\n disease_name=None, tissue_specificity=None, pmid=None, keyword=None,\n subcellular_location=None, tissue_in_reference=None, sequence=None,\n limit=None, as_df=False):\n q = self.session.query(models.Entry)\n model_queries_config = (dataset, models.Entry.dataset), (name, models.\n Entry.name), (recommended_full_name, models.Entry.recommended_full_name\n ), (recommended_short_name, models.Entry.recommended_short_name), (\n gene_name, models.Entry.gene_name), (taxid, models.Entry.taxid)\n q = self.get_model_queries(q, model_queries_config)\n one_to_many_queries_config = (accession, models.Accession.accession), (\n organism_host, models.OrganismHost.taxid), (feature_type, models.\n Feature.type_), (function_, models.Function.text), (ec_number,\n models.ECNumber.ec_number), (db_reference, models.DbReference.\n identifier), (alternative_name, models.AlternativeFullName.name), (\n disease_comment, models.DiseaseComment.comment), (tissue_specificity,\n models.TissueSpecificity.comment), (sequence, models.Sequence.sequence)\n q = self.get_one_to_many_queries(q, one_to_many_queries_config)\n many_to_many_queries_config = (pmid, models.Entry.pmids, models.Pmid.pmid\n ), (keyword, models.Entry.keywords, models.Keyword.name), (\n subcellular_location, models.Entry.subcellular_locations, models.\n SubcellularLocation.location), (tissue_in_reference, models.Entry.\n tissue_in_references, models.TissueInReference.tissue)\n q = self.get_many_to_many_queries(q, many_to_many_queries_config)\n if disease_name:\n q = q.join(models.Entry.disease_comments).join(models.\n DiseaseComment.disease)\n if isinstance(disease_name, str):\n q = q.filter(models.Disease.name.like(disease_name))\n elif isinstance(disease_name, Iterable):\n q = q.filter(models.Disease.name.in_(disease_name))\n return self._limit_and_df(q, limit, as_df)\n",
"role": "user"
},
{
"content": "entry",
"role": "assistant"
}
] |
def x(self, identifier=None, ref_id=None, ref_type=None, name=None, acronym
=None, description=None, entry_name=None, limit=None, as_df=False):
q = self.session.query(models.Disease)
model_queries_config = (identifier, models.Disease.identifier), (ref_id,
models.Disease.ref_id), (ref_type, models.Disease.ref_type), (name,
models.Disease.name), (acronym, models.Disease.acronym), (description,
models.Disease.description)
q = self.get_model_queries(q, model_queries_config)
if entry_name:
q = q.session.query(models.Disease).join(models.DiseaseComment).join(
models.Entry)
if isinstance(entry_name, str):
q = q.filter(models.Entry.name == entry_name)
elif isinstance(entry_name, Iterable):
q = q.filter(models.Entry.name.in_(entry_name))
return self._limit_and_df(q, limit, as_df)
| disease | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, identifier=None, ref_id=None, ref_type=None, name=None, acronym\n =None, description=None, entry_name=None, limit=None, as_df=False):\n q = self.session.query(models.Disease)\n model_queries_config = (identifier, models.Disease.identifier), (ref_id,\n models.Disease.ref_id), (ref_type, models.Disease.ref_type), (name,\n models.Disease.name), (acronym, models.Disease.acronym), (description,\n models.Disease.description)\n q = self.get_model_queries(q, model_queries_config)\n if entry_name:\n q = q.session.query(models.Disease).join(models.DiseaseComment).join(\n models.Entry)\n if isinstance(entry_name, str):\n q = q.filter(models.Entry.name == entry_name)\n elif isinstance(entry_name, Iterable):\n q = q.filter(models.Entry.name.in_(entry_name))\n return self._limit_and_df(q, limit, as_df)\n",
"role": "user"
},
{
"content": "disease",
"role": "assistant"
}
] |
def x(self, comment=None, entry_name=None, limit=None, as_df=False):
q = self.session.query(models.DiseaseComment)
q = self.get_model_queries(q, ((comment, models.DiseaseComment.comment),))
q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),))
return self._limit_and_df(q, limit, as_df)
| disease_comment | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, comment=None, entry_name=None, limit=None, as_df=False):\n q = self.session.query(models.DiseaseComment)\n q = self.get_model_queries(q, ((comment, models.DiseaseComment.comment),))\n q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),))\n return self._limit_and_df(q, limit, as_df)\n",
"role": "user"
},
{
"content": "disease_comment",
"role": "assistant"
}
] |
def x(self, type_=None, name=None, entry_name=None, limit=None, as_df=None):
q = self.session.query(models.OtherGeneName)
model_queries_config = (type_, models.OtherGeneName.type_), (name,
models.OtherGeneName.name)
q = self.get_model_queries(q, model_queries_config)
q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),))
return self._limit_and_df(q, limit, as_df)
| other_gene_name | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, type_=None, name=None, entry_name=None, limit=None, as_df=None):\n q = self.session.query(models.OtherGeneName)\n model_queries_config = (type_, models.OtherGeneName.type_), (name,\n models.OtherGeneName.name)\n q = self.get_model_queries(q, model_queries_config)\n q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),))\n return self._limit_and_df(q, limit, as_df)\n",
"role": "user"
},
{
"content": "other_gene_name",
"role": "assistant"
}
] |
def x(self, name=None, entry_name=None, limit=None, as_df=False):
q = self.session.query(models.AlternativeFullName)
model_queries_config = (name, models.AlternativeFullName.name),
q = self.get_model_queries(q, model_queries_config)
q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),))
return self._limit_and_df(q, limit, as_df)
| alternative_full_name | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, name=None, entry_name=None, limit=None, as_df=False):\n q = self.session.query(models.AlternativeFullName)\n model_queries_config = (name, models.AlternativeFullName.name),\n q = self.get_model_queries(q, model_queries_config)\n q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),))\n return self._limit_and_df(q, limit, as_df)\n",
"role": "user"
},
{
"content": "alternative_full_name",
"role": "assistant"
}
] |
def x(self, name=None, entry_name=None, limit=None, as_df=False):
q = self.session.query(models.AlternativeShortName)
model_queries_config = (name, models.AlternativeShortName.name),
q = self.get_model_queries(q, model_queries_config)
q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),))
return self._limit_and_df(q, limit, as_df)
| alternative_short_name | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, name=None, entry_name=None, limit=None, as_df=False):\n q = self.session.query(models.AlternativeShortName)\n model_queries_config = (name, models.AlternativeShortName.name),\n q = self.get_model_queries(q, model_queries_config)\n q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),))\n return self._limit_and_df(q, limit, as_df)\n",
"role": "user"
},
{
"content": "alternative_short_name",
"role": "assistant"
}
] |
def x(self, accession=None, entry_name=None, limit=None, as_df=False):
q = self.session.query(models.Accession)
model_queries_config = (accession, models.Accession.accession),
q = self.get_model_queries(q, model_queries_config)
q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),))
return self._limit_and_df(q, limit, as_df)
| accession | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, accession=None, entry_name=None, limit=None, as_df=False):\n q = self.session.query(models.Accession)\n model_queries_config = (accession, models.Accession.accession),\n q = self.get_model_queries(q, model_queries_config)\n q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),))\n return self._limit_and_df(q, limit, as_df)\n",
"role": "user"
},
{
"content": "accession",
"role": "assistant"
}
] |
def x(self, pmid=None, entry_name=None, first=None, last=None, volume=None,
name=None, date=None, title=None, limit=None, as_df=False):
q = self.session.query(models.Pmid)
model_queries_config = (pmid, models.Pmid.pmid), (last, models.Pmid.last
), (first, models.Pmid.first), (volume, models.Pmid.volume), (name,
models.Pmid.name), (date, models.Pmid.date), (title, models.Pmid.title)
q = self.get_model_queries(q, model_queries_config)
q = self.get_many_to_many_queries(q, ((entry_name, models.Pmid.entries,
models.Entry.name),))
return self._limit_and_df(q, limit, as_df)
| pmid | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, pmid=None, entry_name=None, first=None, last=None, volume=None,\n name=None, date=None, title=None, limit=None, as_df=False):\n q = self.session.query(models.Pmid)\n model_queries_config = (pmid, models.Pmid.pmid), (last, models.Pmid.last\n ), (first, models.Pmid.first), (volume, models.Pmid.volume), (name,\n models.Pmid.name), (date, models.Pmid.date), (title, models.Pmid.title)\n q = self.get_model_queries(q, model_queries_config)\n q = self.get_many_to_many_queries(q, ((entry_name, models.Pmid.entries,\n models.Entry.name),))\n return self._limit_and_df(q, limit, as_df)\n",
"role": "user"
},
{
"content": "pmid",
"role": "assistant"
}
] |
def x(self, taxid=None, entry_name=None, limit=None, as_df=False):
q = self.session.query(models.OrganismHost)
q = self.get_model_queries(q, ((taxid, models.OrganismHost.taxid),))
q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),))
return self._limit_and_df(q, limit, as_df)
| organism_host | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, taxid=None, entry_name=None, limit=None, as_df=False):\n q = self.session.query(models.OrganismHost)\n q = self.get_model_queries(q, ((taxid, models.OrganismHost.taxid),))\n q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),))\n return self._limit_and_df(q, limit, as_df)\n",
"role": "user"
},
{
"content": "organism_host",
"role": "assistant"
}
] |
def x(self, type_=None, identifier=None, entry_name=None, limit=None, as_df
=False):
q = self.session.query(models.DbReference)
model_queries_config = (type_, models.DbReference.type_), (identifier,
models.DbReference.identifier)
q = self.get_model_queries(q, model_queries_config)
q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),))
return self._limit_and_df(q, limit, as_df)
| db_reference | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, type_=None, identifier=None, entry_name=None, limit=None, as_df\n =False):\n q = self.session.query(models.DbReference)\n model_queries_config = (type_, models.DbReference.type_), (identifier,\n models.DbReference.identifier)\n q = self.get_model_queries(q, model_queries_config)\n q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),))\n return self._limit_and_df(q, limit, as_df)\n",
"role": "user"
},
{
"content": "db_reference",
"role": "assistant"
}
] |
def x(self, type_=None, identifier=None, description=None, entry_name=None,
limit=None, as_df=False):
q = self.session.query(models.Feature)
model_queries_config = (type_, models.Feature.type_), (identifier,
models.Feature.identifier), (description, models.Feature.description)
q = self.get_model_queries(q, model_queries_config)
q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),))
return self._limit_and_df(q, limit, as_df)
| feature | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, type_=None, identifier=None, description=None, entry_name=None,\n limit=None, as_df=False):\n q = self.session.query(models.Feature)\n model_queries_config = (type_, models.Feature.type_), (identifier,\n models.Feature.identifier), (description, models.Feature.description)\n q = self.get_model_queries(q, model_queries_config)\n q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),))\n return self._limit_and_df(q, limit, as_df)\n",
"role": "user"
},
{
"content": "feature",
"role": "assistant"
}
] |
def x(self, text=None, entry_name=None, limit=None, as_df=False):
q = self.session.query(models.Function)
model_queries_config = (text, models.Function.text),
q = self.get_model_queries(q, model_queries_config)
q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),))
return self._limit_and_df(q, limit, as_df)
| function | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, text=None, entry_name=None, limit=None, as_df=False):\n q = self.session.query(models.Function)\n model_queries_config = (text, models.Function.text),\n q = self.get_model_queries(q, model_queries_config)\n q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),))\n return self._limit_and_df(q, limit, as_df)\n",
"role": "user"
},
{
"content": "function",
"role": "assistant"
}
] |
def x(self, ec_number=None, entry_name=None, limit=None, as_df=False):
q = self.session.query(models.ECNumber)
q = self.get_model_queries(q, ((ec_number, models.ECNumber.ec_number),))
q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),))
return self._limit_and_df(q, limit, as_df)
| ec_number | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, ec_number=None, entry_name=None, limit=None, as_df=False):\n q = self.session.query(models.ECNumber)\n q = self.get_model_queries(q, ((ec_number, models.ECNumber.ec_number),))\n q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),))\n return self._limit_and_df(q, limit, as_df)\n",
"role": "user"
},
{
"content": "ec_number",
"role": "assistant"
}
] |
def x(self, location=None, entry_name=None, limit=None, as_df=False):
q = self.session.query(models.SubcellularLocation)
q = self.get_model_queries(q, ((location, models.SubcellularLocation.
location),))
q = self.get_many_to_many_queries(q, ((entry_name, models.
SubcellularLocation.entries, models.Entry.name),))
return self._limit_and_df(q, limit, as_df)
| subcellular_location | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, location=None, entry_name=None, limit=None, as_df=False):\n q = self.session.query(models.SubcellularLocation)\n q = self.get_model_queries(q, ((location, models.SubcellularLocation.\n location),))\n q = self.get_many_to_many_queries(q, ((entry_name, models.\n SubcellularLocation.entries, models.Entry.name),))\n return self._limit_and_df(q, limit, as_df)\n",
"role": "user"
},
{
"content": "subcellular_location",
"role": "assistant"
}
] |
def x(self, comment=None, entry_name=None, limit=None, as_df=False):
q = self.session.query(models.TissueSpecificity)
q = self.get_model_queries(q, ((comment, models.TissueSpecificity.
comment),))
q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),))
return self._limit_and_df(q, limit, as_df)
| tissue_specificity | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, comment=None, entry_name=None, limit=None, as_df=False):\n q = self.session.query(models.TissueSpecificity)\n q = self.get_model_queries(q, ((comment, models.TissueSpecificity.\n comment),))\n q = self.get_one_to_many_queries(q, ((entry_name, models.Entry.name),))\n return self._limit_and_df(q, limit, as_df)\n",
"role": "user"
},
{
"content": "tissue_specificity",
"role": "assistant"
}
] |
def x(self, tissue=None, entry_name=None, limit=None, as_df=False):
q = self.session.query(models.TissueInReference)
model_queries_config = (tissue, models.TissueInReference.tissue),
q = self.get_model_queries(q, model_queries_config)
q = self.get_many_to_many_queries(q, ((entry_name, models.
TissueInReference.entries, models.Entry.name),))
return self._limit_and_df(q, limit, as_df)
| tissue_in_reference | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, tissue=None, entry_name=None, limit=None, as_df=False):\n q = self.session.query(models.TissueInReference)\n model_queries_config = (tissue, models.TissueInReference.tissue),\n q = self.get_model_queries(q, model_queries_config)\n q = self.get_many_to_many_queries(q, ((entry_name, models.\n TissueInReference.entries, models.Entry.name),))\n return self._limit_and_df(q, limit, as_df)\n",
"role": "user"
},
{
"content": "tissue_in_reference",
"role": "assistant"
}
] |
def x(self):
q = self.session.query(distinct(models.DbReference.type_))
return [x[0] for x in q.all()]
| dbreference_types | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n q = self.session.query(distinct(models.DbReference.type_))\n return [x[0] for x in q.all()]\n",
"role": "user"
},
{
"content": "dbreference_types",
"role": "assistant"
}
] |
def x(self):
r = self.session.query(distinct(models.Entry.taxid)).all()
return [x[0] for x in r]
| taxids | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n r = self.session.query(distinct(models.Entry.taxid)).all()\n return [x[0] for x in r]\n",
"role": "user"
},
{
"content": "taxids",
"role": "assistant"
}
] |
def x(self):
r = self.session.query(distinct(models.Entry.dataset)).all()
return [x[0] for x in r]
| datasets | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n r = self.session.query(distinct(models.Entry.dataset)).all()\n return [x[0] for x in r]\n",
"role": "user"
},
{
"content": "datasets",
"role": "assistant"
}
] |
def x(self):
r = self.session.query(distinct(models.Feature.type_)).all()
return [x[0] for x in r]
| feature_types | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n r = self.session.query(distinct(models.Feature.type_)).all()\n return [x[0] for x in r]\n",
"role": "user"
},
{
"content": "feature_types",
"role": "assistant"
}
] |
def x(self):
return [x[0] for x in self.session.query(models.SubcellularLocation.
location).all()]
| subcellular_locations | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n return [x[0] for x in self.session.query(models.SubcellularLocation.\n location).all()]\n",
"role": "user"
},
{
"content": "subcellular_locations",
"role": "assistant"
}
] |
def x(self):
return [x[0] for x in self.session.query(models.TissueInReference.
tissue).all()]
| tissues_in_references | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n return [x[0] for x in self.session.query(models.TissueInReference.\n tissue).all()]\n",
"role": "user"
},
{
"content": "tissues_in_references",
"role": "assistant"
}
] |
def x(self):
return [x[0] for x in self.session.query(models.Keyword.name).all()]
| keywords | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n return [x[0] for x in self.session.query(models.Keyword.name).all()]\n",
"role": "user"
},
{
"content": "keywords",
"role": "assistant"
}
] |
def x(self):
return [x[0] for x in self.session.query(models.Disease.name).all()]
| diseases | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n return [x[0] for x in self.session.query(models.Disease.name).all()]\n",
"role": "user"
},
{
"content": "diseases",
"role": "assistant"
}
] |
def x(self):
return [x for x in self.session.query(models.Version).all()]
| version | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n return [x for x in self.session.query(models.Version).all()]\n",
"role": "user"
},
{
"content": "version",
"role": "assistant"
}
] |
def x(self, force=False):
if not force:
try:
self.api.call_rc('ixPortTakeOwnership {}'.format(self.uri))
except Exception as _:
raise TgnError(
'Failed to take ownership for port {} current owner is {}'.
format(self, self.owner))
else:
self.api.call_rc('ixPortTakeOwnership {} force'.format(self.uri))
| reserve | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, force=False):\n if not force:\n try:\n self.api.call_rc('ixPortTakeOwnership {}'.format(self.uri))\n except Exception as _:\n raise TgnError(\n 'Failed to take ownership for port {} current owner is {}'.\n format(self, self.owner))\n else:\n self.api.call_rc('ixPortTakeOwnership {} force'.format(self.uri))\n",
"role": "user"
},
{
"content": "reserve",
"role": "assistant"
}
] |
def x(self):
self.ix_command('write')
stream_warnings = self.streamRegion.generateWarningList()
warnings_list = self.api.call('join ' + ' {' + stream_warnings + '} ' +
' LiStSeP').split('LiStSeP') if self.streamRegion.generateWarningList(
) else []
for warning in warnings_list:
if warning:
raise StreamWarningsError(warning)
| write | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n self.ix_command('write')\n stream_warnings = self.streamRegion.generateWarningList()\n warnings_list = self.api.call('join ' + ' {' + stream_warnings + '} ' +\n ' LiStSeP').split('LiStSeP') if self.streamRegion.generateWarningList(\n ) else []\n for warning in warnings_list:\n if warning:\n raise StreamWarningsError(warning)\n",
"role": "user"
},
{
"content": "write",
"role": "assistant"
}
] |
def x(self, config_file_name):
config_file_name = config_file_name.replace('\\', '/')
ext = path.splitext(config_file_name)[-1].lower()
if ext == '.prt':
self.api.call_rc('port import "{}" {}'.format(config_file_name,
self.uri))
elif ext == '.str':
self.reset()
self.api.call_rc('stream import "{}" {}'.format(config_file_name,
self.uri))
else:
raise ValueError('Configuration file type {} not supported.'.format
(ext))
self.write()
self.discover()
| load_config | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, config_file_name):\n config_file_name = config_file_name.replace('\\\\', '/')\n ext = path.splitext(config_file_name)[-1].lower()\n if ext == '.prt':\n self.api.call_rc('port import \"{}\" {}'.format(config_file_name,\n self.uri))\n elif ext == '.str':\n self.reset()\n self.api.call_rc('stream import \"{}\" {}'.format(config_file_name,\n self.uri))\n else:\n raise ValueError('Configuration file type {} not supported.'.format\n (ext))\n self.write()\n self.discover()\n",
"role": "user"
},
{
"content": "load_config",
"role": "assistant"
}
] |
def x(self, config_file_name):
config_file_name = config_file_name.replace('\\', '/')
ext = path.splitext(config_file_name)[-1].lower()
if ext == '.prt':
self.api.call_rc('port export "{}" {}'.format(config_file_name,
self.uri))
elif ext == '.str':
self.api.call_rc('stream export "{}" {}'.format(config_file_name,
self.uri))
else:
raise ValueError('Configuration file type {} not supported.'.format
(ext))
| save_config | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, config_file_name):\n config_file_name = config_file_name.replace('\\\\', '/')\n ext = path.splitext(config_file_name)[-1].lower()\n if ext == '.prt':\n self.api.call_rc('port export \"{}\" {}'.format(config_file_name,\n self.uri))\n elif ext == '.str':\n self.api.call_rc('stream export \"{}\" {}'.format(config_file_name,\n self.uri))\n else:\n raise ValueError('Configuration file type {} not supported.'.format\n (ext))\n",
"role": "user"
},
{
"content": "save_config",
"role": "assistant"
}
] |
def x(self, blocking=False):
self.session.start_transmit(blocking, False, self)
| start_transmit | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, blocking=False):\n self.session.start_transmit(blocking, False, self)\n",
"role": "user"
},
{
"content": "start_transmit",
"role": "assistant"
}
] |
def x(self, cap_file_name=None, cap_file_format=IxeCapFileFormat.mem):
return self.session.stop_capture(cap_file_name, cap_file_format, self)[self
]
| stop_capture | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, cap_file_name=None, cap_file_format=IxeCapFileFormat.mem):\n return self.session.stop_capture(cap_file_name, cap_file_format, self)[self\n ]\n",
"role": "user"
},
{
"content": "stop_capture",
"role": "assistant"
}
] |
def x(self, *frame_nums):
frames = []
for frame_num in frame_nums:
if self.captureBuffer.getframe(frame_num) == '0':
frames.append(self.captureBuffer.frame)
else:
frames.append(None)
return frames
| get_cap_frames | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, *frame_nums):\n frames = []\n for frame_num in frame_nums:\n if self.captureBuffer.getframe(frame_num) == '0':\n frames.append(self.captureBuffer.frame)\n else:\n frames.append(None)\n return frames\n",
"role": "user"
},
{
"content": "get_cap_frames",
"role": "assistant"
}
] |
def x(self):
stat = IxeStat(self)
stat.ix_set_default()
stat.enableValidStats = True
stat.ix_set()
stat.write()
| clear_port_stats | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n stat = IxeStat(self)\n stat.ix_set_default()\n stat.enableValidStats = True\n stat.ix_set()\n stat.write()\n",
"role": "user"
},
{
"content": "clear_port_stats",
"role": "assistant"
}
] |
def x(self, mode=IxePhyMode.ignore):
if isinstance(mode, IxePhyMode):
if mode.value:
self.api.call_rc('port setPhyMode {} {}'.format(mode.value,
self.uri))
else:
self.api.call_rc('port setPhyMode {} {}'.format(mode, self.uri))
| set_phy_mode | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, mode=IxePhyMode.ignore):\n if isinstance(mode, IxePhyMode):\n if mode.value:\n self.api.call_rc('port setPhyMode {} {}'.format(mode.value,\n self.uri))\n else:\n self.api.call_rc('port setPhyMode {} {}'.format(mode, self.uri))\n",
"role": "user"
},
{
"content": "set_phy_mode",
"role": "assistant"
}
] |
def x(self, mode):
self.api.call_rc('port setTransmitMode {} {}'.format(mode, self.uri))
| set_transmit_mode | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, mode):\n self.api.call_rc('port setTransmitMode {} {}'.format(mode, self.uri))\n",
"role": "user"
},
{
"content": "set_transmit_mode",
"role": "assistant"
}
] |
def x(image):
sizes = []
seen_widths = []
for size in settings_sizes():
img = get_thumbnail_shim(image, size)
if img.width in seen_widths:
continue
seen_widths.append(img.width)
sizes.append({'url': img.url, 'width': img.width, 'height': img.height}
)
return sizes
| get_image_sizes | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(image):\n sizes = []\n seen_widths = []\n for size in settings_sizes():\n img = get_thumbnail_shim(image, size)\n if img.width in seen_widths:\n continue\n seen_widths.append(img.width)\n sizes.append({'url': img.url, 'width': img.width, 'height': img.height}\n )\n return sizes\n",
"role": "user"
},
{
"content": "get_image_sizes",
"role": "assistant"
}
] |
def x(src, container='div', classes='', inner_classes='', alt='',
background_image=False, no_css=False, aria_hidden=False):
context = {'container': container, 'classes': classes,
'aspect_padding_bottom': aspect_ratio_percent(src), 'alt': alt,
'background_image': background_image, 'no_css': no_css,
'inner_classes': inner_classes, 'aria_hidden': aria_hidden}
if not get_thumbnail_engine():
context['image'] = src
return context
sizes = get_image_sizes(src)
context['image_sizes'] = sizes
context['image'] = sizes[0]
context['image_sizes_json'] = json.dumps(sizes)
srcset_items = ['{} {}w'.format(size['url'], size['width']) for size in
sizes]
context['image_sizes_srcset'] = ', '.join(srcset_items)
return context
| get_template_context | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(src, container='div', classes='', inner_classes='', alt='',\n background_image=False, no_css=False, aria_hidden=False):\n context = {'container': container, 'classes': classes,\n 'aspect_padding_bottom': aspect_ratio_percent(src), 'alt': alt,\n 'background_image': background_image, 'no_css': no_css,\n 'inner_classes': inner_classes, 'aria_hidden': aria_hidden}\n if not get_thumbnail_engine():\n context['image'] = src\n return context\n sizes = get_image_sizes(src)\n context['image_sizes'] = sizes\n context['image'] = sizes[0]\n context['image_sizes_json'] = json.dumps(sizes)\n srcset_items = ['{} {}w'.format(size['url'], size['width']) for size in\n sizes]\n context['image_sizes_srcset'] = ', '.join(srcset_items)\n return context\n",
"role": "user"
},
{
"content": "get_template_context",
"role": "assistant"
}
] |
def x(self, images, reference=None):
images = check_images(images)
reference = check_reference(images, reference)
def x(item):
key, image = item
return asarray([key, self._get(image, reference)])
transformations = images.map(func, with_keys=True).toarray()
if images.shape[0] == 1:
transformations = [transformations]
algorithm = self.__class__.__name__
return RegistrationModel(dict(transformations), algorithm=algorithm)
| fit | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, images, reference=None):\n images = check_images(images)\n reference = check_reference(images, reference)\n\n def x(item):\n key, image = item\n return asarray([key, self._get(image, reference)])\n transformations = images.map(func, with_keys=True).toarray()\n if images.shape[0] == 1:\n transformations = [transformations]\n algorithm = self.__class__.__name__\n return RegistrationModel(dict(transformations), algorithm=algorithm)\n",
"role": "user"
},
{
"content": "fit",
"role": "assistant"
}
] |
def x(self, images, reference=None):
images = check_images(images)
check_reference(images, reference)
def x(image):
t = self._get(image, reference)
return t.apply(image)
return images.map(func)
| fit_and_transform | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, images, reference=None):\n images = check_images(images)\n check_reference(images, reference)\n\n def x(image):\n t = self._get(image, reference)\n return t.apply(image)\n return images.map(func)\n",
"role": "user"
},
{
"content": "fit_and_transform",
"role": "assistant"
}
] |
def x(entities, path_patterns, strict=False):
if isinstance(path_patterns, string_types):
path_patterns = [path_patterns]
for pattern in path_patterns:
if strict:
defined = re.findall('\\{(.*?)(?:<[^>]+>)?\\}', pattern)
if set(entities.keys()) - set(defined):
continue
new_path = pattern
optional_patterns = re.findall('\\[(.*?)\\]', pattern)
for optional_pattern in optional_patterns:
optional_chunk = replace_entities(entities, optional_pattern) or ''
new_path = new_path.replace('[%s]' % optional_pattern,
optional_chunk)
new_path = replace_entities(entities, new_path)
if new_path:
return new_path
return None
| build_path | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(entities, path_patterns, strict=False):\n if isinstance(path_patterns, string_types):\n path_patterns = [path_patterns]\n for pattern in path_patterns:\n if strict:\n defined = re.findall('\\\\{(.*?)(?:<[^>]+>)?\\\\}', pattern)\n if set(entities.keys()) - set(defined):\n continue\n new_path = pattern\n optional_patterns = re.findall('\\\\[(.*?)\\\\]', pattern)\n for optional_pattern in optional_patterns:\n optional_chunk = replace_entities(entities, optional_pattern) or ''\n new_path = new_path.replace('[%s]' % optional_pattern,\n optional_chunk)\n new_path = replace_entities(entities, new_path)\n if new_path:\n return new_path\n return None\n",
"role": "user"
},
{
"content": "build_path",
"role": "assistant"
}
] |
def x(environ):
settings = {}
for name, value in environ.items():
if not name.startswith('DJANGO_'):
continue
name = name.replace('DJANGO_', '', 1)
if _ignore_setting(name):
continue
try:
settings[name] = ast.literal_eval(value)
except (SyntaxError, ValueError) as err:
LOGGER.warn('Unable to parse setting %s=%s (%s)', name, value, err)
return settings
| get_settings_from_environment | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(environ):\n settings = {}\n for name, value in environ.items():\n if not name.startswith('DJANGO_'):\n continue\n name = name.replace('DJANGO_', '', 1)\n if _ignore_setting(name):\n continue\n try:\n settings[name] = ast.literal_eval(value)\n except (SyntaxError, ValueError) as err:\n LOGGER.warn('Unable to parse setting %s=%s (%s)', name, value, err)\n return settings\n",
"role": "user"
},
{
"content": "get_settings_from_environment",
"role": "assistant"
}
] |
def x(data, digest_size=DEFAULT_DIGEST_SIZE):
if digest_size % 3 != 0:
raise ValueError('digest_size must be a multiple of 3')
if not 0 <= digest_size <= 63:
raise ValueError('digest_size must be between 0 and 63 (bytes)')
sha512 = Digest(hashlib.sha512(data.encode(ENC)).digest())
return sha512[:digest_size]
| vmc_digest | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(data, digest_size=DEFAULT_DIGEST_SIZE):\n if digest_size % 3 != 0:\n raise ValueError('digest_size must be a multiple of 3')\n if not 0 <= digest_size <= 63:\n raise ValueError('digest_size must be between 0 and 63 (bytes)')\n sha512 = Digest(hashlib.sha512(data.encode(ENC)).digest())\n return sha512[:digest_size]\n",
"role": "user"
},
{
"content": "vmc_digest",
"role": "assistant"
}
] |
def x(matrix, top):
assert isinstance(matrix, ExpMatrix)
assert isinstance(top, (int, np.integer))
if top >= matrix.p:
logger.warning(
'Variance filter has no effect ("top" parameter is >= number of genes).'
)
return matrix.copy()
var = np.var(matrix.X, axis=1, ddof=1)
total_var = np.sum(var)
a = np.argsort(var)
a = a[::-1]
sel = np.zeros(matrix.p, dtype=np.bool_)
sel[a[:top]] = True
lost_p = matrix.p - top
lost_var = total_var - np.sum(var[sel])
logger.info(
'Selected the %d most variable genes (excluded %.1f%% of genes, representing %.1f%% of total variance).'
, top, 100 * (lost_p / float(matrix.p)), 100 * (lost_var / total_var))
matrix = matrix.loc[sel]
return matrix
| filter_variance | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(matrix, top):\n assert isinstance(matrix, ExpMatrix)\n assert isinstance(top, (int, np.integer))\n if top >= matrix.p:\n logger.warning(\n 'Variance filter has no effect (\"top\" parameter is >= number of genes).'\n )\n return matrix.copy()\n var = np.var(matrix.X, axis=1, ddof=1)\n total_var = np.sum(var)\n a = np.argsort(var)\n a = a[::-1]\n sel = np.zeros(matrix.p, dtype=np.bool_)\n sel[a[:top]] = True\n lost_p = matrix.p - top\n lost_var = total_var - np.sum(var[sel])\n logger.info(\n 'Selected the %d most variable genes (excluded %.1f%% of genes, representing %.1f%% of total variance).'\n , top, 100 * (lost_p / float(matrix.p)), 100 * (lost_var / total_var))\n matrix = matrix.loc[sel]\n return matrix\n",
"role": "user"
},
{
"content": "filter_variance",
"role": "assistant"
}
] |
def x(matrix, top):
assert isinstance(matrix, ExpMatrix)
assert isinstance(top, int)
if top >= matrix.p:
logger.warning(
'Gene expression filter with `top` parameter that is >= the number of genes!'
)
top = matrix.p
a = np.argsort(np.mean(matrix.X, axis=1))
a = a[::-1]
sel = np.zeros(matrix.p, dtype=np.bool_)
sel[a[:top]] = True
matrix = matrix.loc[sel]
return matrix
| filter_mean | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(matrix, top):\n assert isinstance(matrix, ExpMatrix)\n assert isinstance(top, int)\n if top >= matrix.p:\n logger.warning(\n 'Gene expression filter with `top` parameter that is >= the number of genes!'\n )\n top = matrix.p\n a = np.argsort(np.mean(matrix.X, axis=1))\n a = a[::-1]\n sel = np.zeros(matrix.p, dtype=np.bool_)\n sel[a[:top]] = True\n matrix = matrix.loc[sel]\n return matrix\n",
"role": "user"
},
{
"content": "filter_mean",
"role": "assistant"
}
] |
def x(matrix, top, percentile=50):
assert isinstance(matrix, ExpMatrix)
assert isinstance(top, int)
assert isinstance(percentile, (int, float))
if top >= matrix.p:
logger.warning(
'Gene expression filter with `top` parameter that is >= the number of genes!'
)
top = matrix.p
a = np.argsort(np.percentile(matrix.X, percentile, axis=1))
a = a[::-1]
sel = np.zeros(matrix.p, dtype=np.bool_)
sel[a[:top]] = True
matrix = matrix.loc[sel]
return matrix
| filter_percentile | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(matrix, top, percentile=50):\n assert isinstance(matrix, ExpMatrix)\n assert isinstance(top, int)\n assert isinstance(percentile, (int, float))\n if top >= matrix.p:\n logger.warning(\n 'Gene expression filter with `top` parameter that is >= the number of genes!'\n )\n top = matrix.p\n a = np.argsort(np.percentile(matrix.X, percentile, axis=1))\n a = a[::-1]\n sel = np.zeros(matrix.p, dtype=np.bool_)\n sel[a[:top]] = True\n matrix = matrix.loc[sel]\n return matrix\n",
"role": "user"
},
{
"content": "filter_percentile",
"role": "assistant"
}
] |
def x(chrom):
try:
c = int(chrom)
except:
if chrom in ['X', 'Y']:
return chrom
elif chrom == 'MT':
return '_MT'
else:
return '__' + chrom
else:
return '%02d' % c
| transform_chrom | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(chrom):\n try:\n c = int(chrom)\n except:\n if chrom in ['X', 'Y']:\n return chrom\n elif chrom == 'MT':\n return '_MT'\n else:\n return '__' + chrom\n else:\n return '%02d' % c\n",
"role": "user"
},
{
"content": "transform_chrom",
"role": "assistant"
}
] |
def x(fasta_file, fancy_sort=True):
chromlen = []
with gzip.open(fasta_file, 'rt', encoding='ascii') as fh:
fasta = SeqIO.parse(fh, 'fasta')
for i, f in enumerate(fasta):
chromlen.append((f.id, len(f.seq)))
_LOGGER.info('Processed chromosome "%s"...', f.id)
chromlen = pd.Series(OrderedDict(chromlen))
chromlen.index.name = 'Chromosome'
chromlen.name = 'Length'
if fancy_sort:
chrom_for_sorting = chromlen.index.to_series().apply(_transform_chrom)
a = chrom_for_sorting.argsort(kind='mergesort')
chromlen = chromlen.iloc[a]
return chromlen
| get_chromosome_lengths | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fasta_file, fancy_sort=True):\n chromlen = []\n with gzip.open(fasta_file, 'rt', encoding='ascii') as fh:\n fasta = SeqIO.parse(fh, 'fasta')\n for i, f in enumerate(fasta):\n chromlen.append((f.id, len(f.seq)))\n _LOGGER.info('Processed chromosome \"%s\"...', f.id)\n chromlen = pd.Series(OrderedDict(chromlen))\n chromlen.index.name = 'Chromosome'\n chromlen.name = 'Length'\n if fancy_sort:\n chrom_for_sorting = chromlen.index.to_series().apply(_transform_chrom)\n a = chrom_for_sorting.argsort(kind='mergesort')\n chromlen = chromlen.iloc[a]\n return chromlen\n",
"role": "user"
},
{
"content": "get_chromosome_lengths",
"role": "assistant"
}
] |
def x(schema):
def x(schema):
if isinstance(schema, dict):
if 'allOf' in schema:
for x in schema['allOf']:
sub_schema = x
sub_schema.pop('title', None)
schema = _merge_dicts(schema, sub_schema)
schema.pop('allOf')
schema = traverse(schema)
elif 'properties' in schema:
for x in schema.get('properties', []):
schema['properties'][x] = traverse(schema['properties'][x])
elif 'items' in schema:
schema['items'] = traverse(schema['items'])
return schema
return traverse(schema)
| resolve_schema | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(schema):\n\n def x(schema):\n if isinstance(schema, dict):\n if 'allOf' in schema:\n for x in schema['allOf']:\n sub_schema = x\n sub_schema.pop('title', None)\n schema = _merge_dicts(schema, sub_schema)\n schema.pop('allOf')\n schema = traverse(schema)\n elif 'properties' in schema:\n for x in schema.get('properties', []):\n schema['properties'][x] = traverse(schema['properties'][x])\n elif 'items' in schema:\n schema['items'] = traverse(schema['items'])\n return schema\n return traverse(schema)\n",
"role": "user"
},
{
"content": "resolve_schema",
"role": "assistant"
}
] |
def x(first, second):
new = deepcopy(first)
for k, v in second.items():
if isinstance(v, dict) and v:
ret = _merge_dicts(new.get(k, dict()), v)
new[k] = ret
else:
new[k] = second[k]
return new
| merge_dicts | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(first, second):\n new = deepcopy(first)\n for k, v in second.items():\n if isinstance(v, dict) and v:\n ret = _merge_dicts(new.get(k, dict()), v)\n new[k] = ret\n else:\n new[k] = second[k]\n return new\n",
"role": "user"
},
{
"content": "merge_dicts",
"role": "assistant"
}
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.