source_code
stringlengths 45
7.24k
| function_name
stringlengths 1
111
| messages
list |
|---|---|---|
def x(list_idl):
seqs = set()
for idl in list_idl.loci2seq:
[seqs.add(s) for s in list_idl.loci2seq[idl]]
return seqs
|
get_seqs
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(list_idl):\n seqs = set()\n for idl in list_idl.loci2seq:\n [seqs.add(s) for s in list_idl.loci2seq[idl]]\n return seqs\n",
"role": "user"
},
{
"content": "get_seqs",
"role": "assistant"
}
] |
def x(s1, s2, i1, i2):
c = len(set(s1).intersection(s2))
t = min(len(s1), len(s2))
pct = 1.0 * c / t * t
is_gt = up_threshold(pct, t * 1.0, parameters.similar)
logger.debug('_common: pct %s of clusters:%s %s = %s' % (1.0 * c / t,
i1, i2, is_gt))
if pct < parameters.similar and is_gt and pct > 0:
pct = parameters.similar
return pct / t
|
common
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(s1, s2, i1, i2):\n c = len(set(s1).intersection(s2))\n t = min(len(s1), len(s2))\n pct = 1.0 * c / t * t\n is_gt = up_threshold(pct, t * 1.0, parameters.similar)\n logger.debug('_common: pct %s of clusters:%s %s = %s' % (1.0 * c / t,\n i1, i2, is_gt))\n if pct < parameters.similar and is_gt and pct > 0:\n pct = parameters.similar\n return pct / t\n",
"role": "user"
},
{
"content": "common",
"role": "assistant"
}
] |
def x(pairs, common, clus_seen, loci_similarity):
all_true1 = all([all([(common and loci_similarity[p, c] > parameters.
similar) for p in pairs if (p, c) in loci_similarity]) for c in
clus_seen])
all_true2 = all([all([(common and loci_similarity[c, p] > parameters.
similar) for p in pairs if (c, p) in loci_similarity]) for c in
clus_seen])
return all_true1 * all_true2
|
is_consistent
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(pairs, common, clus_seen, loci_similarity):\n all_true1 = all([all([(common and loci_similarity[p, c] > parameters.\n similar) for p in pairs if (p, c) in loci_similarity]) for c in\n clus_seen])\n all_true2 = all([all([(common and loci_similarity[c, p] > parameters.\n similar) for p in pairs if (c, p) in loci_similarity]) for c in\n clus_seen])\n return all_true1 * all_true2\n",
"role": "user"
},
{
"content": "is_consistent",
"role": "assistant"
}
] |
def x(loci, loci_similarity):
n_cluster = 0
internal_cluster = {}
clus_seen = {}
loci_sorted = sorted(loci_similarity.iteritems(), key=operator.
itemgetter(1), reverse=True)
for pairs, sim in loci_sorted:
common = sim > parameters.similar
n_cluster += 1
logger.debug('_merge_similar:try new cluster %s' % n_cluster)
new_c = cluster(n_cluster)
p_seen, p_unseen = [], []
size = min(len(_get_seqs(loci[pairs[0]])), len(_get_seqs(loci[pairs
[1]])))
if common:
consistent = _is_consistent(pairs, common, clus_seen,
loci_similarity)
logger.debug('_merge_similar: clusters seen: %s' % clus_seen)
logger.debug(
'_merge_similar: id %s common %s|%s total %s consistent %s' %
(pairs, sim, common, size, consistent))
if not consistent:
continue
if pairs[0] in clus_seen:
p_seen.append(pairs[0])
p_unseen.append(pairs[1])
if pairs[1] in clus_seen:
p_seen.append(pairs[1])
p_unseen.append(pairs[0])
if len(p_seen) == 0:
new_c = _merge_cluster(loci[pairs[0]], new_c)
new_c = _merge_cluster(loci[pairs[1]], new_c)
[clus_seen.update({p: n_cluster}) for p in pairs]
internal_cluster[n_cluster] = new_c
if len(p_seen) == 1:
idc_seen = clus_seen[p_seen[0]]
internal_cluster[idc_seen] = _merge_cluster(loci[p_unseen[0
]], internal_cluster[idc_seen])
clus_seen[p_unseen[0]] = idc_seen
else:
logger.debug('_merge_similar: id %s %s are different' % pairs)
continue
internal_cluster.update(_add_unseen(loci, clus_seen, n_cluster))
logger.debug('_merge_similar: total clus %s' % len(internal_cluster.keys())
)
return internal_cluster
|
merge_similar
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(loci, loci_similarity):\n n_cluster = 0\n internal_cluster = {}\n clus_seen = {}\n loci_sorted = sorted(loci_similarity.iteritems(), key=operator.\n itemgetter(1), reverse=True)\n for pairs, sim in loci_sorted:\n common = sim > parameters.similar\n n_cluster += 1\n logger.debug('_merge_similar:try new cluster %s' % n_cluster)\n new_c = cluster(n_cluster)\n p_seen, p_unseen = [], []\n size = min(len(_get_seqs(loci[pairs[0]])), len(_get_seqs(loci[pairs\n [1]])))\n if common:\n consistent = _is_consistent(pairs, common, clus_seen,\n loci_similarity)\n logger.debug('_merge_similar: clusters seen: %s' % clus_seen)\n logger.debug(\n '_merge_similar: id %s common %s|%s total %s consistent %s' %\n (pairs, sim, common, size, consistent))\n if not consistent:\n continue\n if pairs[0] in clus_seen:\n p_seen.append(pairs[0])\n p_unseen.append(pairs[1])\n if pairs[1] in clus_seen:\n p_seen.append(pairs[1])\n p_unseen.append(pairs[0])\n if len(p_seen) == 0:\n new_c = _merge_cluster(loci[pairs[0]], new_c)\n new_c = _merge_cluster(loci[pairs[1]], new_c)\n [clus_seen.update({p: n_cluster}) for p in pairs]\n internal_cluster[n_cluster] = new_c\n if len(p_seen) == 1:\n idc_seen = clus_seen[p_seen[0]]\n internal_cluster[idc_seen] = _merge_cluster(loci[p_unseen[0\n ]], internal_cluster[idc_seen])\n clus_seen[p_unseen[0]] = idc_seen\n else:\n logger.debug('_merge_similar: id %s %s are different' % pairs)\n continue\n internal_cluster.update(_add_unseen(loci, clus_seen, n_cluster))\n logger.debug('_merge_similar: total clus %s' % len(internal_cluster.keys())\n )\n return internal_cluster\n",
"role": "user"
},
{
"content": "merge_similar",
"role": "assistant"
}
] |
def x(old, new):
logger.debug('_merge_cluster: %s to %s' % (old.id, new.id))
logger.debug('_merge_cluster: add idls %s' % old.loci2seq.keys())
for idl in old.loci2seq:
new.add_id_member(old.loci2seq[idl], idl)
return new
|
merge_cluster
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(old, new):\n logger.debug('_merge_cluster: %s to %s' % (old.id, new.id))\n logger.debug('_merge_cluster: add idls %s' % old.loci2seq.keys())\n for idl in old.loci2seq:\n new.add_id_member(old.loci2seq[idl], idl)\n return new\n",
"role": "user"
},
{
"content": "merge_cluster",
"role": "assistant"
}
] |
def x(list_c, s2p, n_cluster):
logger.debug('_solve_conflict: count once')
if parameters.decision_cluster == 'bayes':
return decide_by_bayes(list_c, s2p)
loci_similarity = _calculate_similarity(list_c)
loci_similarity = sorted(loci_similarity.iteritems(), key=operator.
itemgetter(1), reverse=True)
common = sum([score for p, score in loci_similarity])
while common > 0:
n_cluster += 1
logger.debug('_solve_conflict: ma %s' % loci_similarity)
pairs = loci_similarity[0][0]
score = loci_similarity[0][1]
logger.debug('_solve_conflict: common %s, new %s' % (score, n_cluster))
if parameters.decision_cluster.startswith('most-voted'):
list_c = _split_cluster_by_most_vote(list_c, pairs)
else:
list_c = _split_cluster(list_c, pairs, n_cluster)
list_c = {k: v for k, v in list_c.iteritems() if len(v.loci2seq) > 0}
loci_similarity = _calculate_similarity(list_c)
loci_similarity = sorted(loci_similarity.iteritems(), key=operator.
itemgetter(1), reverse=True)
common = sum([score for p, score in loci_similarity])
logger.debug('_solve_conflict: solved clusters %s' % len(list_c.keys())
)
return list_c
|
solve_conflict
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(list_c, s2p, n_cluster):\n logger.debug('_solve_conflict: count once')\n if parameters.decision_cluster == 'bayes':\n return decide_by_bayes(list_c, s2p)\n loci_similarity = _calculate_similarity(list_c)\n loci_similarity = sorted(loci_similarity.iteritems(), key=operator.\n itemgetter(1), reverse=True)\n common = sum([score for p, score in loci_similarity])\n while common > 0:\n n_cluster += 1\n logger.debug('_solve_conflict: ma %s' % loci_similarity)\n pairs = loci_similarity[0][0]\n score = loci_similarity[0][1]\n logger.debug('_solve_conflict: common %s, new %s' % (score, n_cluster))\n if parameters.decision_cluster.startswith('most-voted'):\n list_c = _split_cluster_by_most_vote(list_c, pairs)\n else:\n list_c = _split_cluster(list_c, pairs, n_cluster)\n list_c = {k: v for k, v in list_c.iteritems() if len(v.loci2seq) > 0}\n loci_similarity = _calculate_similarity(list_c)\n loci_similarity = sorted(loci_similarity.iteritems(), key=operator.\n itemgetter(1), reverse=True)\n common = sum([score for p, score in loci_similarity])\n logger.debug('_solve_conflict: solved clusters %s' % len(list_c.keys())\n )\n return list_c\n",
"role": "user"
},
{
"content": "solve_conflict",
"role": "assistant"
}
] |
def x(c, pairs, n):
old = c[p[0]]
new = c[p[1]]
new_c = cluster(n)
common = set(_get_seqs(old)).intersection(_get_seqs(new))
for idl in old.loci2seq:
in_common = list(set(common).intersection(old.loci2seq[idl]))
if len(in_common) > 0:
logger.debug('_split_cluster: in_common %s with pair 1' % len(
in_common))
new_c.add_id_member(in_common, idl)
old.loci2seq[idl] = list(set(old.loci2seq[idl]) - set(common))
logger.debug('_split_cluster: len old %s with pair 1' % len(old
.loci2seq))
for idl in new.loci2seq:
in_common = list(set(common).intersection(new.loci2seq[idl]))
if len(in_common) > 0:
logger.debug('_split_cluster: in_common %s with pair 2' % len(
in_common))
new_c.add_id_member(in_common, idl)
new.loci2seq[idl] = list(set(new.loci2seq[idl]) - set(common))
logger.debug('_split_cluster: len old %s with pair 2' % len(new
.loci2seq))
old.update()
new.update()
old.loci2seq = {k: v for k, v in old.loci2seq.iteritems() if len(v) > 0}
new.loci2seq = {k: v for k, v in new.loci2seq.iteritems() if len(v) > 0}
c[n] = new
c[p[0]] = old
c[p[1]] = new
return c
|
split_cluster
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(c, pairs, n):\n old = c[p[0]]\n new = c[p[1]]\n new_c = cluster(n)\n common = set(_get_seqs(old)).intersection(_get_seqs(new))\n for idl in old.loci2seq:\n in_common = list(set(common).intersection(old.loci2seq[idl]))\n if len(in_common) > 0:\n logger.debug('_split_cluster: in_common %s with pair 1' % len(\n in_common))\n new_c.add_id_member(in_common, idl)\n old.loci2seq[idl] = list(set(old.loci2seq[idl]) - set(common))\n logger.debug('_split_cluster: len old %s with pair 1' % len(old\n .loci2seq))\n for idl in new.loci2seq:\n in_common = list(set(common).intersection(new.loci2seq[idl]))\n if len(in_common) > 0:\n logger.debug('_split_cluster: in_common %s with pair 2' % len(\n in_common))\n new_c.add_id_member(in_common, idl)\n new.loci2seq[idl] = list(set(new.loci2seq[idl]) - set(common))\n logger.debug('_split_cluster: len old %s with pair 2' % len(new\n .loci2seq))\n old.update()\n new.update()\n old.loci2seq = {k: v for k, v in old.loci2seq.iteritems() if len(v) > 0}\n new.loci2seq = {k: v for k, v in new.loci2seq.iteritems() if len(v) > 0}\n c[n] = new\n c[p[0]] = old\n c[p[1]] = new\n return c\n",
"role": "user"
},
{
"content": "split_cluster",
"role": "assistant"
}
] |
def x(c, p):
old, new = c[p[0]], c[p[1]]
old_size = _get_seqs(old)
new_size = _get_seqs(new)
logger.debug('_most_vote: size of %s with %s - %s with %s' % (old.id,
len(old_size), new.id, len(new_size)))
if len(old_size) > len(new_size):
keep, remove = old, new
else:
keep, remove = new, old
common = list(set(old_size).intersection(new_size))
logger.debug('_most_vote: keep %s remove %s with common %s' % (keep.id,
remove.id, len(common)))
for idl in remove.loci2seq:
if len(common) > 0:
remove.loci2seq[idl] = list(set(remove.loci2seq[idl]) - set(common)
)
keep.loci2seq = {k: v for k, v in keep.loci2seq.iteritems() if len(v) > 0}
remove.loci2seq = {k: v for k, v in remove.loci2seq.iteritems() if len(
v) > 0}
keep.update()
remove.update()
c[keep.id] = keep
c[remove.id] = remove
return c
|
split_cluster_by_most_vote
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(c, p):\n old, new = c[p[0]], c[p[1]]\n old_size = _get_seqs(old)\n new_size = _get_seqs(new)\n logger.debug('_most_vote: size of %s with %s - %s with %s' % (old.id,\n len(old_size), new.id, len(new_size)))\n if len(old_size) > len(new_size):\n keep, remove = old, new\n else:\n keep, remove = new, old\n common = list(set(old_size).intersection(new_size))\n logger.debug('_most_vote: keep %s remove %s with common %s' % (keep.id,\n remove.id, len(common)))\n for idl in remove.loci2seq:\n if len(common) > 0:\n remove.loci2seq[idl] = list(set(remove.loci2seq[idl]) - set(common)\n )\n keep.loci2seq = {k: v for k, v in keep.loci2seq.iteritems() if len(v) > 0}\n remove.loci2seq = {k: v for k, v in remove.loci2seq.iteritems() if len(\n v) > 0}\n keep.update()\n remove.update()\n c[keep.id] = keep\n c[remove.id] = remove\n return c\n",
"role": "user"
},
{
"content": "split_cluster_by_most_vote",
"role": "assistant"
}
] |
def x(list_c):
global REMOVED
init = len(list_c)
list_c = {k: v for k, v in list_c.iteritems() if len(_get_seqs(v)) >
parameters.min_seqs}
logger.debug('_clean_cluster: number of clusters %s ' % len(list_c.keys()))
list_c = {k: _select_loci(v) for k, v in list_c.iteritems()}
end = len(list_c)
REMOVED += init - end
return list_c
|
clean_cluster
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(list_c):\n global REMOVED\n init = len(list_c)\n list_c = {k: v for k, v in list_c.iteritems() if len(_get_seqs(v)) >\n parameters.min_seqs}\n logger.debug('_clean_cluster: number of clusters %s ' % len(list_c.keys()))\n list_c = {k: _select_loci(v) for k, v in list_c.iteritems()}\n end = len(list_c)\n REMOVED += init - end\n return list_c\n",
"role": "user"
},
{
"content": "clean_cluster",
"role": "assistant"
}
] |
def x(c):
loci_len = {k: len(v) for k, v in c.loci2seq.iteritems()}
logger.debug('_select_loci: number of loci %s' % len(c.loci2seq.keys()))
loci_len_sort = sorted(loci_len.iteritems(), key=operator.itemgetter(1),
reverse=True)
max_size = loci_len_sort[0][1]
logger.debug('_select_loci: max size %s' % max_size)
loci_clean = {locus: c.loci2seq[locus] for locus, size in loci_len_sort if
size > 0.8 * max_size}
c.loci2seq = loci_clean
removed = list(set(c.idmembers.keys()) - set(_get_seqs(c)))
c.add_id_member(removed, loci_len_sort[0][0])
logger.debug('_select_loci: number of loci %s after cleaning' % len(c.
loci2seq.keys()))
return c
|
select_loci
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(c):\n loci_len = {k: len(v) for k, v in c.loci2seq.iteritems()}\n logger.debug('_select_loci: number of loci %s' % len(c.loci2seq.keys()))\n loci_len_sort = sorted(loci_len.iteritems(), key=operator.itemgetter(1),\n reverse=True)\n max_size = loci_len_sort[0][1]\n logger.debug('_select_loci: max size %s' % max_size)\n loci_clean = {locus: c.loci2seq[locus] for locus, size in loci_len_sort if\n size > 0.8 * max_size}\n c.loci2seq = loci_clean\n removed = list(set(c.idmembers.keys()) - set(_get_seqs(c)))\n c.add_id_member(removed, loci_len_sort[0][0])\n logger.debug('_select_loci: number of loci %s after cleaning' % len(c.\n loci2seq.keys()))\n return c\n",
"role": "user"
},
{
"content": "select_loci",
"role": "assistant"
}
] |
def x(c, locilen_sorted, seen_seqs, filtered, maxseq, n_cluster):
first_run = 0
seen_seqs = list()
n_cluster += 1
logger.debug('_solve_loci:new cluster %s' % n_cluster)
new_c = cluster(n_cluster)
for idl, lenl in locilen_sorted:
locus_seqs = c.loci2seq[idl]
if first_run == 0:
seen_seqs = locus_seqs
first_run = 1
first_idl = idl
intersect = list(set(seen_seqs).intersection(locus_seqs))
common = 0
if intersect:
common = len(intersect) * 1.0 / min(len(seen_seqs), len(locus_seqs)
)
logger.debug(
'_sole_loci:id %s idl %s len %s max %s seen %s inter %s common %s '
% (c.id, idl, lenl, maxseq, len(seen_seqs), len(intersect),
common))
if common * 1.0 >= 0.6:
if lenl * 1.0 >= 0.6 * maxseq:
c, new_c, seen_seqs = _merge_loci_in_cluster(c, new_c, idl,
seen_seqs)
else:
c, new_c, seen_seqs = _merge_with_first_loci(c, new_c,
first_idl, idl, seen_seqs)
else:
c = _remove_seqs_from_loci(c, idl, seen_seqs)
filtered[n_cluster] = new_c
return c, seen_seqs, filtered, n_cluster
|
solve_loci_deprecated
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(c, locilen_sorted, seen_seqs, filtered, maxseq, n_cluster):\n first_run = 0\n seen_seqs = list()\n n_cluster += 1\n logger.debug('_solve_loci:new cluster %s' % n_cluster)\n new_c = cluster(n_cluster)\n for idl, lenl in locilen_sorted:\n locus_seqs = c.loci2seq[idl]\n if first_run == 0:\n seen_seqs = locus_seqs\n first_run = 1\n first_idl = idl\n intersect = list(set(seen_seqs).intersection(locus_seqs))\n common = 0\n if intersect:\n common = len(intersect) * 1.0 / min(len(seen_seqs), len(locus_seqs)\n )\n logger.debug(\n '_sole_loci:id %s idl %s len %s max %s seen %s inter %s common %s '\n % (c.id, idl, lenl, maxseq, len(seen_seqs), len(intersect),\n common))\n if common * 1.0 >= 0.6:\n if lenl * 1.0 >= 0.6 * maxseq:\n c, new_c, seen_seqs = _merge_loci_in_cluster(c, new_c, idl,\n seen_seqs)\n else:\n c, new_c, seen_seqs = _merge_with_first_loci(c, new_c,\n first_idl, idl, seen_seqs)\n else:\n c = _remove_seqs_from_loci(c, idl, seen_seqs)\n filtered[n_cluster] = new_c\n return c, seen_seqs, filtered, n_cluster\n",
"role": "user"
},
{
"content": "solve_loci_deprecated",
"role": "assistant"
}
] |
def x(string):
ann = set()
if not string:
return 'This cluster is inter-genic.'
for item in string:
for db in item:
ann = ann.union(set(item[db]))
return 'annotated as: %s ...' % ','.join(list(ann)[:3])
|
get_description
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(string):\n ann = set()\n if not string:\n return 'This cluster is inter-genic.'\n for item in string:\n for db in item:\n ann = ann.union(set(item[db]))\n return 'annotated as: %s ...' % ','.join(list(ann)[:3])\n",
"role": "user"
},
{
"content": "get_description",
"role": "assistant"
}
] |
def x(profile):
x = set()
for sample in profile:
x = x.union(set(profile[sample].keys()))
if not x:
return ''
end, start = max(x), min(x)
x = range(start, end, 4)
scaled_profile = defaultdict(list)
for pos in x:
for sample in profile:
y = _get_closer(profile[sample], pos)
if y:
scaled_profile[sample].append(profile[sample][y])
else:
scaled_profile[sample].append(0)
return {'x': list(x), 'y': scaled_profile, 'names': scaled_profile.keys()}
|
set_format
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(profile):\n x = set()\n for sample in profile:\n x = x.union(set(profile[sample].keys()))\n if not x:\n return ''\n end, start = max(x), min(x)\n x = range(start, end, 4)\n scaled_profile = defaultdict(list)\n for pos in x:\n for sample in profile:\n y = _get_closer(profile[sample], pos)\n if y:\n scaled_profile[sample].append(profile[sample][y])\n else:\n scaled_profile[sample].append(0)\n return {'x': list(x), 'y': scaled_profile, 'names': scaled_profile.keys()}\n",
"role": "user"
},
{
"content": "set_format",
"role": "assistant"
}
] |
def x(con, data):
with con:
cur = con.cursor()
cur.execute('DROP TABLE IF EXISTS clusters;')
cur.execute(
'CREATE TABLE clusters(Id INT, Description TEXT, Locus TEXT, Annotation TEXT, Sequences TEXT, Profile TXT, Precursor TXT)'
)
for c in data[0]:
locus = json.dumps(data[0][c]['loci'])
annotation = json.dumps(data[0][c]['ann'])
description = _get_description(data[0][c]['ann'])
sequences = json.dumps(_get_sequences(data[0][c]))
keys = data[0][c]['freq'][0].values()[0].keys()
profile = 'Not available.'
if 'profile' in data[0][c]:
profile = json.dumps(_set_format(data[0][c]['profile']))
precursor = json.dumps(data[0][c].get('precursor'))
cur.execute(
"INSERT INTO clusters VALUES(%s, '%s', '%s', '%s', '%s', '%s', '%s')"
% (c, description, locus, annotation, sequences, profile,
precursor))
|
insert_data
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(con, data):\n with con:\n cur = con.cursor()\n cur.execute('DROP TABLE IF EXISTS clusters;')\n cur.execute(\n 'CREATE TABLE clusters(Id INT, Description TEXT, Locus TEXT, Annotation TEXT, Sequences TEXT, Profile TXT, Precursor TXT)'\n )\n for c in data[0]:\n locus = json.dumps(data[0][c]['loci'])\n annotation = json.dumps(data[0][c]['ann'])\n description = _get_description(data[0][c]['ann'])\n sequences = json.dumps(_get_sequences(data[0][c]))\n keys = data[0][c]['freq'][0].values()[0].keys()\n profile = 'Not available.'\n if 'profile' in data[0][c]:\n profile = json.dumps(_set_format(data[0][c]['profile']))\n precursor = json.dumps(data[0][c].get('precursor'))\n cur.execute(\n \"INSERT INTO clusters VALUES(%s, '%s', '%s', '%s', '%s', '%s', '%s')\"\n % (c, description, locus, annotation, sequences, profile,\n precursor))\n",
"role": "user"
},
{
"content": "insert_data",
"role": "assistant"
}
] |
def x(file_in):
loc_id = 1
bedfile_clusters = ''
bamfile = pybedtools.BedTool(file_in)
bed = pybedtools.BedTool.bam_to_bed(bamfile)
for c, start, end, name, q, strand in bed:
loc_id += 1
bedfile_clusters += '%s\t%s\t%s\t%s\t%s\t%s\n' % (c, start, end,
name, loc_id, strand)
return bedfile_clusters
|
parse_align_file
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(file_in):\n loc_id = 1\n bedfile_clusters = ''\n bamfile = pybedtools.BedTool(file_in)\n bed = pybedtools.BedTool.bam_to_bed(bamfile)\n for c, start, end, name, q, strand in bed:\n loc_id += 1\n bedfile_clusters += '%s\\t%s\\t%s\\t%s\\t%s\\t%s\\n' % (c, start, end,\n name, loc_id, strand)\n return bedfile_clusters\n",
"role": "user"
},
{
"content": "parse_align_file",
"role": "assistant"
}
] |
def x(seq_obj, in_file):
name = ''
index = 1
total = defaultdict(int)
with open(in_file) as handle_in:
line = handle_in.readline().strip()
cols = line.split('\t')
samples = cols[2:]
for line in handle_in:
line = line.strip()
cols = line.split('\t')
name = int(cols[0].replace('seq_', ''))
seq = cols[1]
exp = {}
for i in range(len(samples)):
exp[samples[i]] = int(cols[i + 2])
total[samples[i]] += int(cols[i + 2])
index = index + 1
if name in seq_obj:
seq_obj[name].set_freq(exp)
seq_obj[name].set_seq(seq)
seq_obj = _normalize_seqs(seq_obj, total)
return seq_obj, total, index
|
parse_ma_file
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(seq_obj, in_file):\n name = ''\n index = 1\n total = defaultdict(int)\n with open(in_file) as handle_in:\n line = handle_in.readline().strip()\n cols = line.split('\\t')\n samples = cols[2:]\n for line in handle_in:\n line = line.strip()\n cols = line.split('\\t')\n name = int(cols[0].replace('seq_', ''))\n seq = cols[1]\n exp = {}\n for i in range(len(samples)):\n exp[samples[i]] = int(cols[i + 2])\n total[samples[i]] += int(cols[i + 2])\n index = index + 1\n if name in seq_obj:\n seq_obj[name].set_freq(exp)\n seq_obj[name].set_seq(seq)\n seq_obj = _normalize_seqs(seq_obj, total)\n return seq_obj, total, index\n",
"role": "user"
},
{
"content": "parse_ma_file",
"role": "assistant"
}
] |
def x(cols, field='name'):
field = field.lower()
try:
group = cols[2]
attrs = cols[8].split(';')
name = [attr.strip().split(' ')[1] for attr in attrs if attr.strip(
).split(' ')[0].lower().endswith(field)]
if not name:
name = [attr.strip().split(' ')[1] for attr in attrs if attr.
strip().split(' ')[0].lower().endswith('gene_id')]
if not name:
name = ['None']
biotype = [attr.strip().split(' ')[1] for attr in attrs if attr.
strip().split(' ')[0].lower().endswith('biotype')]
if biotype:
group = biotype[0]
c = cols[0]
s = int(cols[3])
e = int(cols[4])
st = cols[6]
return [c, s, e, st, group, name[0]]
except (Exception, e):
logger.error(cols)
logger.error('File is not in correct format')
logger.error('Expect chr source feature start end . strand attributes')
logger.error("Attributes are 'gene_name SNCA; gene_id ENSG; '")
logger.error('The 3rd column is used as type of small RNA (like miRNA)'
)
logger.error("at least should contains '; *name NAME; '")
logger.error(e)
raise
|
read_gtf_line
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cols, field='name'):\n field = field.lower()\n try:\n group = cols[2]\n attrs = cols[8].split(';')\n name = [attr.strip().split(' ')[1] for attr in attrs if attr.strip(\n ).split(' ')[0].lower().endswith(field)]\n if not name:\n name = [attr.strip().split(' ')[1] for attr in attrs if attr.\n strip().split(' ')[0].lower().endswith('gene_id')]\n if not name:\n name = ['None']\n biotype = [attr.strip().split(' ')[1] for attr in attrs if attr.\n strip().split(' ')[0].lower().endswith('biotype')]\n if biotype:\n group = biotype[0]\n c = cols[0]\n s = int(cols[3])\n e = int(cols[4])\n st = cols[6]\n return [c, s, e, st, group, name[0]]\n except (Exception, e):\n logger.error(cols)\n logger.error('File is not in correct format')\n logger.error('Expect chr source feature start end . strand attributes')\n logger.error(\"Attributes are 'gene_name SNCA; gene_id ENSG; '\")\n logger.error('The 3rd column is used as type of small RNA (like miRNA)'\n )\n logger.error(\"at least should contains '; *name NAME; '\")\n logger.error(e)\n raise\n",
"role": "user"
},
{
"content": "read_gtf_line",
"role": "assistant"
}
] |
def x(pos_a, pos_b):
strd = '-'
if pos_a[2] in pos_b[2]:
strd = '+'
if pos_a[2] in '+' and pos_b[2] in '+':
lento5 = pos_a[0] - pos_b[1] + 1
lento3 = pos_a[1] - pos_b[1] + 1
if pos_a[2] in '+' and pos_b[2] in '-':
lento5 = pos_a[1] - pos_b[0] + 1
lento3 = pos_a[0] - pos_b[1] + 1
if pos_a[2] in '-' and pos_b[2] in '+':
lento5 = pos_a[0] - pos_b[1] + 1
lento3 = pos_a[1] - pos_b[0] + 1
if pos_a[2] in '-' and pos_b[2] in '-':
lento3 = pos_a[0] - pos_b[0] + 1
lento5 = pos_a[1] - pos_b[1] + 1
else:
lento5 = pos_a[0] - pos_b[0] + 1
lento3 = pos_a[1] - pos_b[1] + 1
return lento5, lento3, strd
|
position_in_feature
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(pos_a, pos_b):\n strd = '-'\n if pos_a[2] in pos_b[2]:\n strd = '+'\n if pos_a[2] in '+' and pos_b[2] in '+':\n lento5 = pos_a[0] - pos_b[1] + 1\n lento3 = pos_a[1] - pos_b[1] + 1\n if pos_a[2] in '+' and pos_b[2] in '-':\n lento5 = pos_a[1] - pos_b[0] + 1\n lento3 = pos_a[0] - pos_b[1] + 1\n if pos_a[2] in '-' and pos_b[2] in '+':\n lento5 = pos_a[0] - pos_b[1] + 1\n lento3 = pos_a[1] - pos_b[0] + 1\n if pos_a[2] in '-' and pos_b[2] in '-':\n lento3 = pos_a[0] - pos_b[0] + 1\n lento5 = pos_a[1] - pos_b[1] + 1\n else:\n lento5 = pos_a[0] - pos_b[0] + 1\n lento3 = pos_a[1] - pos_b[1] + 1\n return lento5, lento3, strd\n",
"role": "user"
},
{
"content": "position_in_feature",
"role": "assistant"
}
] |
def x(c, clus_obj, db, type_ann, feature_id='name'):
id_sa, id_ea, id_id, id_idl, id_sta = 1, 2, 3, 4, 5
if type_ann == 'bed':
id_sb = 7
id_eb = 8
id_stb = 11
id_tag = 9
ida = 0
clus_id = clus_obj.clus
loci_id = clus_obj.loci
db = os.path.splitext(db)[0]
logger.debug('Type:%s\n' % type_ann)
for cols in c.features():
if type_ann == 'gtf':
cb, sb, eb, stb, db, tag = read_gtf_line(cols[6:], feature_id)
else:
sb = int(cols[id_sb])
eb = int(cols[id_eb])
stb = cols[id_stb]
tag = cols[id_tag]
id = int(cols[id_id])
idl = int(cols[id_idl])
if id in clus_id:
clus = clus_id[id]
sa = int(cols[id_sa])
ea = int(cols[id_ea])
ida += 1
lento5, lento3, strd = _position_in_feature([sa, ea, cols[
id_sta]], [sb, eb, stb])
if db in loci_id[idl].db_ann:
ann = annotation(db, tag, strd, lento5, lento3)
tdb = loci_id[idl].db_ann[db]
tdb.add_db_ann(ida, ann)
loci_id[idl].add_db(db, tdb)
else:
ann = annotation(db, tag, strd, lento5, lento3)
tdb = dbannotation(1)
tdb.add_db_ann(ida, ann)
loci_id[idl].add_db(db, tdb)
clus_id[id] = clus
clus_obj.clus = clus_id
clus_obj.loci = loci_id
return clus_obj
|
anncluster
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(c, clus_obj, db, type_ann, feature_id='name'):\n id_sa, id_ea, id_id, id_idl, id_sta = 1, 2, 3, 4, 5\n if type_ann == 'bed':\n id_sb = 7\n id_eb = 8\n id_stb = 11\n id_tag = 9\n ida = 0\n clus_id = clus_obj.clus\n loci_id = clus_obj.loci\n db = os.path.splitext(db)[0]\n logger.debug('Type:%s\\n' % type_ann)\n for cols in c.features():\n if type_ann == 'gtf':\n cb, sb, eb, stb, db, tag = read_gtf_line(cols[6:], feature_id)\n else:\n sb = int(cols[id_sb])\n eb = int(cols[id_eb])\n stb = cols[id_stb]\n tag = cols[id_tag]\n id = int(cols[id_id])\n idl = int(cols[id_idl])\n if id in clus_id:\n clus = clus_id[id]\n sa = int(cols[id_sa])\n ea = int(cols[id_ea])\n ida += 1\n lento5, lento3, strd = _position_in_feature([sa, ea, cols[\n id_sta]], [sb, eb, stb])\n if db in loci_id[idl].db_ann:\n ann = annotation(db, tag, strd, lento5, lento3)\n tdb = loci_id[idl].db_ann[db]\n tdb.add_db_ann(ida, ann)\n loci_id[idl].add_db(db, tdb)\n else:\n ann = annotation(db, tag, strd, lento5, lento3)\n tdb = dbannotation(1)\n tdb.add_db_ann(ida, ann)\n loci_id[idl].add_db(db, tdb)\n clus_id[id] = clus\n clus_obj.clus = clus_id\n clus_obj.loci = loci_id\n return clus_obj\n",
"role": "user"
},
{
"content": "anncluster",
"role": "assistant"
}
] |
def x(bam_in, genome, out):
if not genome:
logger.info('No genome given. skipping.')
return None
out_file = op.join(out, op.basename(bam_in) + '_cov.tsv')
if file_exists(out_file):
return None
fai = genome + '.fai'
cov = pybedtools.BedTool(bam_in).genome_coverage(g=fai, max=1)
cov.saveas(out_file)
total = 0
for region in cov:
if region[0] == 'genome' and int(region[1]) != 0:
total += float(region[4])
logger.info('Total genome with sequences: %s ' % total)
|
detect_complexity
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(bam_in, genome, out):\n if not genome:\n logger.info('No genome given. skipping.')\n return None\n out_file = op.join(out, op.basename(bam_in) + '_cov.tsv')\n if file_exists(out_file):\n return None\n fai = genome + '.fai'\n cov = pybedtools.BedTool(bam_in).genome_coverage(g=fai, max=1)\n cov.saveas(out_file)\n total = 0\n for region in cov:\n if region[0] == 'genome' and int(region[1]) != 0:\n total += float(region[4])\n logger.info('Total genome with sequences: %s ' % total)\n",
"role": "user"
},
{
"content": "detect_complexity",
"role": "assistant"
}
] |
def x(bam_in, mask=None):
seq_obj = defaultdict(int)
if mask:
mask_file = op.splitext(bam_in)[0] + '_mask.bam'
if not file_exists(mask_file):
pybedtools.BedTool(bam_file).intersect(b=mask, v=True).saveas(
mask_file)
bam_in = mask_file
out_file = op.splitext(bam_in)[0] + '_rmlw.bam'
run('samtools index %s' % bam_in)
if not file_exists(bam_in + '.bai'):
raise IOError(
'Failed to created bam index of %s. Try to do it manually' % bam_in
)
bam_handle = pysam.AlignmentFile(bam_in, 'rb')
with pysam.AlignmentFile(out_file, 'wb', template=bam_handle
) as out_handle:
for read in bam_handle.fetch():
seq_name = int(read.query_name.replace('seq_', ''))
match_size = [nts for oper, nts in read.cigartuples if oper == 0]
subs_size = [nts for oper, nts in read.cigartuples if oper == 4]
if match_size[0] < 17:
continue
if subs_size:
if subs_size[0] > 3:
continue
try:
nh = read.get_tag('NH')
except KeyError:
nh = 1
seq_obj[seq_name] = sequence(seq_name)
seq_obj[seq_name].align = nh
out_handle.write(read)
return out_file, seq_obj
|
clean_bam_file
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(bam_in, mask=None):\n seq_obj = defaultdict(int)\n if mask:\n mask_file = op.splitext(bam_in)[0] + '_mask.bam'\n if not file_exists(mask_file):\n pybedtools.BedTool(bam_file).intersect(b=mask, v=True).saveas(\n mask_file)\n bam_in = mask_file\n out_file = op.splitext(bam_in)[0] + '_rmlw.bam'\n run('samtools index %s' % bam_in)\n if not file_exists(bam_in + '.bai'):\n raise IOError(\n 'Failed to created bam index of %s. Try to do it manually' % bam_in\n )\n bam_handle = pysam.AlignmentFile(bam_in, 'rb')\n with pysam.AlignmentFile(out_file, 'wb', template=bam_handle\n ) as out_handle:\n for read in bam_handle.fetch():\n seq_name = int(read.query_name.replace('seq_', ''))\n match_size = [nts for oper, nts in read.cigartuples if oper == 0]\n subs_size = [nts for oper, nts in read.cigartuples if oper == 4]\n if match_size[0] < 17:\n continue\n if subs_size:\n if subs_size[0] > 3:\n continue\n try:\n nh = read.get_tag('NH')\n except KeyError:\n nh = 1\n seq_obj[seq_name] = sequence(seq_name)\n seq_obj[seq_name].align = nh\n out_handle.write(read)\n return out_file, seq_obj\n",
"role": "user"
},
{
"content": "clean_bam_file",
"role": "assistant"
}
] |
def x(c, current_seq, MIN_SEQ, non_un_gl=False):
current_loci = {}
current_clus = {}
sequence2clusters = defaultdict(set)
lindex = 0
eindex = 0
previous_id = 0
for line in c.features():
c, start, end, name, score, strand, c_id = line
name = int(name.replace('seq_', ''))
pos = int(start) if strand == '+' else int(end)
if name not in current_seq:
continue
if c.find('Un_gl') > -1 and non_un_gl:
continue
if c_id != previous_id:
if previous_id > 0:
if len(current_clus[eindex].idmembers) < MIN_SEQ:
for s in current_clus[eindex].idmembers:
sequence2clusters[s] = sequence2clusters[s] - set([
eindex])
del current_clus[eindex]
logger.debug('detect_cluster: %s %s %s' % (c_id, previous_id, name)
)
lindex += 1
eindex += 1
current_clus[eindex] = cluster(eindex)
newpos = position(lindex, c, start, end, strand)
current_loci[lindex] = newpos
current_loci[lindex].end = int(end)
current_loci[lindex].coverage[pos] += 1
size = range(pos, pos + current_seq[name].len)
current_loci[lindex].counts.update(dict(zip(size, [current_seq[name
].total()] * current_seq[name].len)))
current_clus[eindex].idmembers[name] = 1
current_clus[eindex].add_id_member([name], lindex)
current_seq[name].add_pos(lindex, pos)
previous_id = c_id
sequence2clusters[name].add(eindex)
logger.info('%s Clusters read' % eindex)
metacluster_obj, cluster_id = _find_metaclusters(current_clus,
sequence2clusters, current_seq, MIN_SEQ)
return cluster_info_obj(current_clus, metacluster_obj, current_loci,
current_seq)
|
detect_clusters
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(c, current_seq, MIN_SEQ, non_un_gl=False):\n current_loci = {}\n current_clus = {}\n sequence2clusters = defaultdict(set)\n lindex = 0\n eindex = 0\n previous_id = 0\n for line in c.features():\n c, start, end, name, score, strand, c_id = line\n name = int(name.replace('seq_', ''))\n pos = int(start) if strand == '+' else int(end)\n if name not in current_seq:\n continue\n if c.find('Un_gl') > -1 and non_un_gl:\n continue\n if c_id != previous_id:\n if previous_id > 0:\n if len(current_clus[eindex].idmembers) < MIN_SEQ:\n for s in current_clus[eindex].idmembers:\n sequence2clusters[s] = sequence2clusters[s] - set([\n eindex])\n del current_clus[eindex]\n logger.debug('detect_cluster: %s %s %s' % (c_id, previous_id, name)\n )\n lindex += 1\n eindex += 1\n current_clus[eindex] = cluster(eindex)\n newpos = position(lindex, c, start, end, strand)\n current_loci[lindex] = newpos\n current_loci[lindex].end = int(end)\n current_loci[lindex].coverage[pos] += 1\n size = range(pos, pos + current_seq[name].len)\n current_loci[lindex].counts.update(dict(zip(size, [current_seq[name\n ].total()] * current_seq[name].len)))\n current_clus[eindex].idmembers[name] = 1\n current_clus[eindex].add_id_member([name], lindex)\n current_seq[name].add_pos(lindex, pos)\n previous_id = c_id\n sequence2clusters[name].add(eindex)\n logger.info('%s Clusters read' % eindex)\n metacluster_obj, cluster_id = _find_metaclusters(current_clus,\n sequence2clusters, current_seq, MIN_SEQ)\n return cluster_info_obj(current_clus, metacluster_obj, current_loci,\n current_seq)\n",
"role": "user"
},
{
"content": "detect_clusters",
"role": "assistant"
}
] |
def x(clus_obj, sequence2clusters, current_seq, min_seqs):
seen = defaultdict(int)
metacluster = defaultdict(set)
c_index = len(sequence2clusters)
logger.info('Creating meta-clusters based on shared sequences: %s' %
c_index)
meta_idx = 1
bar = ProgressBar(maxval=c_index)
bar.start()
bar.update()
for itern, name in enumerate(sequence2clusters):
clusters = sequence2clusters[name]
if len(clusters) == 0:
c_index -= 1
continue
current_seq[name].align = 1
meta_idx += 1
bar.update(itern)
already_in = _common(clusters, seen)
_update(clusters, meta_idx, seen)
metacluster[meta_idx] = metacluster[meta_idx].union(clusters)
if already_in:
for seen_metacluster in already_in:
clusters2merge = metacluster[seen_metacluster]
metacluster[meta_idx] = metacluster[meta_idx].union(
clusters2merge)
_update(clusters2merge, meta_idx, seen)
del metacluster[seen_metacluster]
logger.info('%s metaclusters from %s sequences' % (len(metacluster),
c_index))
return metacluster, seen
|
find_metaclusters
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(clus_obj, sequence2clusters, current_seq, min_seqs):\n seen = defaultdict(int)\n metacluster = defaultdict(set)\n c_index = len(sequence2clusters)\n logger.info('Creating meta-clusters based on shared sequences: %s' %\n c_index)\n meta_idx = 1\n bar = ProgressBar(maxval=c_index)\n bar.start()\n bar.update()\n for itern, name in enumerate(sequence2clusters):\n clusters = sequence2clusters[name]\n if len(clusters) == 0:\n c_index -= 1\n continue\n current_seq[name].align = 1\n meta_idx += 1\n bar.update(itern)\n already_in = _common(clusters, seen)\n _update(clusters, meta_idx, seen)\n metacluster[meta_idx] = metacluster[meta_idx].union(clusters)\n if already_in:\n for seen_metacluster in already_in:\n clusters2merge = metacluster[seen_metacluster]\n metacluster[meta_idx] = metacluster[meta_idx].union(\n clusters2merge)\n _update(clusters2merge, meta_idx, seen)\n del metacluster[seen_metacluster]\n logger.info('%s metaclusters from %s sequences' % (len(metacluster),\n c_index))\n return metacluster, seen\n",
"role": "user"
},
{
"content": "find_metaclusters",
"role": "assistant"
}
] |
def x(clus_obj, min_seqs):
logger.info('Creating meta-clusters based on shared sequences.')
seen = defaultdict()
metacluster = defaultdict(list)
c_index = clus_obj.keys()
meta_idx = 0
with ProgressBar(maxval=len(c_index), redirect_stdout=True) as p:
for itern, c in enumerate(c_index):
p.update(itern)
clus = clus_obj[c]
if len(clus.idmembers.keys()) < min_seqs:
del clus_obj[c]
continue
logger.debug('reading cluster %s' % c)
logger.debug('loci2seq %s' % clus.loci2seq)
already_in, not_in = _get_seqs_from_cluster(clus.idmembers.keys
(), seen)
logger.debug('seen %s news %s' % (already_in, not_in))
meta_idx += 1
metacluster[meta_idx].append(c)
seen.update(dict(zip(not_in, [meta_idx] * len(not_in))))
if len(already_in) > 0:
logger.debug('seen in %s' % already_in)
for eindex in already_in:
for cluster in metacluster[eindex]:
metacluster[meta_idx].append(cluster)
prev_clus = clus_obj[cluster]
logger.debug('_find_families: prev %s current %s' %
(eindex, clus.id))
seqs_in = prev_clus.idmembers.keys()
seen.update(dict(zip(seqs_in, [meta_idx] * len(
seqs_in))))
del metacluster[eindex]
logger.info('%s clusters merged' % len(metacluster))
return metacluster, seen
|
find_families_deprecated
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(clus_obj, min_seqs):\n logger.info('Creating meta-clusters based on shared sequences.')\n seen = defaultdict()\n metacluster = defaultdict(list)\n c_index = clus_obj.keys()\n meta_idx = 0\n with ProgressBar(maxval=len(c_index), redirect_stdout=True) as p:\n for itern, c in enumerate(c_index):\n p.update(itern)\n clus = clus_obj[c]\n if len(clus.idmembers.keys()) < min_seqs:\n del clus_obj[c]\n continue\n logger.debug('reading cluster %s' % c)\n logger.debug('loci2seq %s' % clus.loci2seq)\n already_in, not_in = _get_seqs_from_cluster(clus.idmembers.keys\n (), seen)\n logger.debug('seen %s news %s' % (already_in, not_in))\n meta_idx += 1\n metacluster[meta_idx].append(c)\n seen.update(dict(zip(not_in, [meta_idx] * len(not_in))))\n if len(already_in) > 0:\n logger.debug('seen in %s' % already_in)\n for eindex in already_in:\n for cluster in metacluster[eindex]:\n metacluster[meta_idx].append(cluster)\n prev_clus = clus_obj[cluster]\n logger.debug('_find_families: prev %s current %s' %\n (eindex, clus.id))\n seqs_in = prev_clus.idmembers.keys()\n seen.update(dict(zip(seqs_in, [meta_idx] * len(\n seqs_in))))\n del metacluster[eindex]\n logger.info('%s clusters merged' % len(metacluster))\n return metacluster, seen\n",
"role": "user"
},
{
"content": "find_families_deprecated",
"role": "assistant"
}
] |
def x(clus_obj):
new_cluster = {}
for cid in clus_obj.clus:
cluster = clus_obj.clus[cid]
cluster.update()
logger.debug('peak calling for %s' % cid)
bigger = cluster.locimaxid
if bigger in clus_obj.loci:
s, e = min(clus_obj.loci[bigger].counts.keys()), max(clus_obj.
loci[bigger].counts.keys())
scale = s
if clus_obj.loci[bigger].strand == '-':
scale = e
logger.debug('bigger %s at %s-%s' % (bigger, s, e))
dt = np.array([0] * (abs(e - s) + 12))
for pos in clus_obj.loci[bigger].counts:
ss = abs(int(pos) - scale) + 5
dt[ss] += clus_obj.loci[bigger].counts[pos]
x = np.array(range(0, len(dt)))
logger.debug('x %s and y %s' % (x, dt))
if len(x) > 35 + 12:
peaks = list(np.array(pysen.pysenMMean(x, dt)) - 5)
logger.debug(peaks)
else:
peaks = ['short']
cluster.peaks = peaks
new_cluster[cid] = cluster
clus_obj.clus = new_cluster
return clus_obj
|
peak_calling
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(clus_obj):\n new_cluster = {}\n for cid in clus_obj.clus:\n cluster = clus_obj.clus[cid]\n cluster.update()\n logger.debug('peak calling for %s' % cid)\n bigger = cluster.locimaxid\n if bigger in clus_obj.loci:\n s, e = min(clus_obj.loci[bigger].counts.keys()), max(clus_obj.\n loci[bigger].counts.keys())\n scale = s\n if clus_obj.loci[bigger].strand == '-':\n scale = e\n logger.debug('bigger %s at %s-%s' % (bigger, s, e))\n dt = np.array([0] * (abs(e - s) + 12))\n for pos in clus_obj.loci[bigger].counts:\n ss = abs(int(pos) - scale) + 5\n dt[ss] += clus_obj.loci[bigger].counts[pos]\n x = np.array(range(0, len(dt)))\n logger.debug('x %s and y %s' % (x, dt))\n if len(x) > 35 + 12:\n peaks = list(np.array(pysen.pysenMMean(x, dt)) - 5)\n logger.debug(peaks)\n else:\n peaks = ['short']\n cluster.peaks = peaks\n new_cluster[cid] = cluster\n clus_obj.clus = new_cluster\n return clus_obj\n",
"role": "user"
},
{
"content": "peak_calling",
"role": "assistant"
}
] |
def x(args):
if args.fasta:
name = None
seq = ''
reads = dict()
with open(args.fasta) as in_handle:
for line in in_handle:
if line.startswith('>'):
if name:
reads.update(_generate_reads(seq, name))
seq = ''
name = line[1:-1]
else:
seq += line.strip()
reads.update(_generate_reads(seq, name))
_write_reads(reads, args.out)
|
simulate
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(args):\n if args.fasta:\n name = None\n seq = ''\n reads = dict()\n with open(args.fasta) as in_handle:\n for line in in_handle:\n if line.startswith('>'):\n if name:\n reads.update(_generate_reads(seq, name))\n seq = ''\n name = line[1:-1]\n else:\n seq += line.strip()\n reads.update(_generate_reads(seq, name))\n _write_reads(reads, args.out)\n",
"role": "user"
},
{
"content": "simulate",
"role": "assistant"
}
] |
def x(seq, name):
reads = dict()
if len(seq) < 130 and len(seq) > 70:
reads.update(_mature(seq[:40], 0, name))
reads.update(_mature(seq[-40:], len(seq) - 40, name))
reads.update(_noise(seq, name))
reads.update(_noise(seq, name, 25))
return reads
|
generate_reads
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(seq, name):\n reads = dict()\n if len(seq) < 130 and len(seq) > 70:\n reads.update(_mature(seq[:40], 0, name))\n reads.update(_mature(seq[-40:], len(seq) - 40, name))\n reads.update(_noise(seq, name))\n reads.update(_noise(seq, name, 25))\n return reads\n",
"role": "user"
},
{
"content": "generate_reads",
"role": "assistant"
}
] |
def x(subseq, absolute, c, size=33, total=5000):
reads = dict()
probs = [0.1, 0.2, 0.4, 0.2, 0.1]
end = 5 + size
error = [-2, -1, 0, 1, 2]
for error5 in error:
for error3 in error:
s = 5 - error5
e = end - error3
seen = subseq[s:e]
counts = int(probs[error5 + 2] * probs[error3 + 2] * total) + 1
name = 'seq_%s_%s_%s_x%s' % (c, s + absolute, e + absolute, counts)
reads[name] = seen, counts
return reads
|
mature
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(subseq, absolute, c, size=33, total=5000):\n reads = dict()\n probs = [0.1, 0.2, 0.4, 0.2, 0.1]\n end = 5 + size\n error = [-2, -1, 0, 1, 2]\n for error5 in error:\n for error3 in error:\n s = 5 - error5\n e = end - error3\n seen = subseq[s:e]\n counts = int(probs[error5 + 2] * probs[error3 + 2] * total) + 1\n name = 'seq_%s_%s_%s_x%s' % (c, s + absolute, e + absolute, counts)\n reads[name] = seen, counts\n return reads\n",
"role": "user"
},
{
"content": "mature",
"role": "assistant"
}
] |
def x(seq, c, size=33, total=1000):
reads = dict()
seen = 0
while seen < total:
s = random.randint(0, len(seq) - size)
e = s + size + random.randint(-5, 5)
p = random.uniform(0, 0.1)
counts = int(p * total) + 1
seen += counts
name = 'seq_%s_%s_%s_x%s' % (c, s, e, counts)
reads[name] = seq[s:e], counts
return reads
|
noise
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(seq, c, size=33, total=1000):\n reads = dict()\n seen = 0\n while seen < total:\n s = random.randint(0, len(seq) - size)\n e = s + size + random.randint(-5, 5)\n p = random.uniform(0, 0.1)\n counts = int(p * total) + 1\n seen += counts\n name = 'seq_%s_%s_%s_x%s' % (c, s, e, counts)\n reads[name] = seq[s:e], counts\n return reads\n",
"role": "user"
},
{
"content": "noise",
"role": "assistant"
}
] |
def x(reads, prefix):
out_ma = prefix + '.ma'
out_fasta = prefix + '.fasta'
out_real = prefix + '.txt'
with open(out_ma, 'w') as ma_handle:
print('id\tseq\tsample', file=ma_handle, end='')
with open(out_fasta, 'w') as fa_handle:
with open(out_real, 'w') as read_handle:
for idx, r in enumerate(reads):
info = r.split('_')
print('seq_%s\t%s\t%s' % (idx, reads[r][0], reads[r][1]
), file=ma_handle, end='')
print('>seq_%s\n%s' % (idx, reads[r][0]), file=
fa_handle, end='')
print('%s\t%s\t%s\t%s\t%s\t%s\t%s' % (idx, r, reads[r][
0], reads[r][1], info[1], info[2], info[3]), file=
read_handle, end='')
|
write_reads
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(reads, prefix):\n out_ma = prefix + '.ma'\n out_fasta = prefix + '.fasta'\n out_real = prefix + '.txt'\n with open(out_ma, 'w') as ma_handle:\n print('id\\tseq\\tsample', file=ma_handle, end='')\n with open(out_fasta, 'w') as fa_handle:\n with open(out_real, 'w') as read_handle:\n for idx, r in enumerate(reads):\n info = r.split('_')\n print('seq_%s\\t%s\\t%s' % (idx, reads[r][0], reads[r][1]\n ), file=ma_handle, end='')\n print('>seq_%s\\n%s' % (idx, reads[r][0]), file=\n fa_handle, end='')\n print('%s\\t%s\\t%s\\t%s\\t%s\\t%s\\t%s' % (idx, r, reads[r][\n 0], reads[r][1], info[1], info[2], info[3]), file=\n read_handle, end='')\n",
"role": "user"
},
{
"content": "write_reads",
"role": "assistant"
}
] |
def x(args):
logger.info('Reading sequeces')
data = parse_ma_file(args.ma)
logger.info('Get sequences from sam')
is_align = _read_sam(args.sam)
is_json, is_db = _read_json(args.json)
res = _summarise_sam(data, is_align, is_json, is_db)
_write_suma(res, os.path.join(args.out, 'stats_align.dat'))
logger.info('Done')
|
stats
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(args):\n logger.info('Reading sequeces')\n data = parse_ma_file(args.ma)\n logger.info('Get sequences from sam')\n is_align = _read_sam(args.sam)\n is_json, is_db = _read_json(args.json)\n res = _summarise_sam(data, is_align, is_json, is_db)\n _write_suma(res, os.path.join(args.out, 'stats_align.dat'))\n logger.info('Done')\n",
"role": "user"
},
{
"content": "stats",
"role": "assistant"
}
] |
def x(fn_json):
is_json = set()
is_db = {}
with open(fn_json) as handle:
data = json.load(handle)
for item in list(data[0].values()):
seqs_name = [list(x.keys()) for x in item['seqs']]
db_name = item['valid'] if 'valid' in item else None
[is_json.add(name[0]) for name in seqs_name]
if db_name:
[is_db.update({name[0]: ','.join(db_name)}) for name in
seqs_name]
return is_json, is_db
|
read_json
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(fn_json):\n is_json = set()\n is_db = {}\n with open(fn_json) as handle:\n data = json.load(handle)\n for item in list(data[0].values()):\n seqs_name = [list(x.keys()) for x in item['seqs']]\n db_name = item['valid'] if 'valid' in item else None\n [is_json.add(name[0]) for name in seqs_name]\n if db_name:\n [is_db.update({name[0]: ','.join(db_name)}) for name in\n seqs_name]\n return is_json, is_db\n",
"role": "user"
},
{
"content": "read_json",
"role": "assistant"
}
] |
def x(cmd, data=None, checks=None, region=None, log_error=True, log_stdout=
False):
try:
logger.debug(' '.join(str(x) for x in cmd) if not isinstance(cmd,
basestring) else cmd)
_do_run(cmd, checks, log_stdout)
except:
if log_error:
logger.info('error at command')
raise
|
run
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cmd, data=None, checks=None, region=None, log_error=True, log_stdout=\n False):\n try:\n logger.debug(' '.join(str(x) for x in cmd) if not isinstance(cmd,\n basestring) else cmd)\n _do_run(cmd, checks, log_stdout)\n except:\n if log_error:\n logger.info('error at command')\n raise\n",
"role": "user"
},
{
"content": "run",
"role": "assistant"
}
] |
def x(cmd):
if isinstance(cmd, basestring):
if cmd.find(' | ') > 0 or cmd.find('>(') or cmd.find('<('):
return 'set -o pipefail; ' + cmd, True, find_bash()
else:
return cmd, True, None
else:
return [str(x) for x in cmd], False, None
|
normalize_cmd_args
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cmd):\n if isinstance(cmd, basestring):\n if cmd.find(' | ') > 0 or cmd.find('>(') or cmd.find('<('):\n return 'set -o pipefail; ' + cmd, True, find_bash()\n else:\n return cmd, True, None\n else:\n return [str(x) for x in cmd], False, None\n",
"role": "user"
},
{
"content": "normalize_cmd_args",
"role": "assistant"
}
] |
def x(cmd, checks, log_stdout=False):
cmd, shell_arg, executable_arg = _normalize_cmd_args(cmd)
s = subprocess.Popen(cmd, shell=shell_arg, executable=executable_arg,
stdout=subprocess.PIPE, stderr=subprocess.STDOUT, close_fds=True)
debug_stdout = collections.deque(maxlen=100)
while 1:
line = s.stdout.readline()
if line:
debug_stdout.append(line)
if log_stdout:
logger.debug(line.rstrip())
else:
logger.debug(line.rstrip())
exitcode = s.poll()
if exitcode is not None:
for line in s.stdout:
debug_stdout.append(line)
if exitcode is not None and exitcode != 0:
error_msg = ' '.join(cmd) if not isinstance(cmd, basestring
) else cmd
error_msg += '\n'
error_msg += ''.join(debug_stdout)
s.communicate()
s.stdout.close()
raise subprocess.CalledProcessError(exitcode, error_msg)
else:
break
s.communicate()
s.stdout.close()
if checks:
for check in checks:
if not check():
raise IOError('External command failed')
|
do_run
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cmd, checks, log_stdout=False):\n cmd, shell_arg, executable_arg = _normalize_cmd_args(cmd)\n s = subprocess.Popen(cmd, shell=shell_arg, executable=executable_arg,\n stdout=subprocess.PIPE, stderr=subprocess.STDOUT, close_fds=True)\n debug_stdout = collections.deque(maxlen=100)\n while 1:\n line = s.stdout.readline()\n if line:\n debug_stdout.append(line)\n if log_stdout:\n logger.debug(line.rstrip())\n else:\n logger.debug(line.rstrip())\n exitcode = s.poll()\n if exitcode is not None:\n for line in s.stdout:\n debug_stdout.append(line)\n if exitcode is not None and exitcode != 0:\n error_msg = ' '.join(cmd) if not isinstance(cmd, basestring\n ) else cmd\n error_msg += '\\n'\n error_msg += ''.join(debug_stdout)\n s.communicate()\n s.stdout.close()\n raise subprocess.CalledProcessError(exitcode, error_msg)\n else:\n break\n s.communicate()\n s.stdout.close()\n if checks:\n for check in checks:\n if not check():\n raise IOError('External command failed')\n",
"role": "user"
},
{
"content": "do_run",
"role": "assistant"
}
] |
def x(list_c, loci_obj, seq_obj):
seqs = defaultdict(set)
for c in list_c.values():
for l in c.loci2seq:
[seqs[s].add(c.id) for s in c.loci2seq[l]]
common = [s for s in seqs if len(seqs[s]) > 1]
seqs_in_c = defaultdict(float)
for c in list_c.values():
for l in c.loci2seq:
for s in c.loci2seq[l]:
if s in common:
pos = seq_obj[s].pos[l]
cov = 1.0 * loci_obj[l].coverage[pos]
if seqs_in_c[s, c.id] < cov:
seqs_in_c[s, c.id] = cov
seqs_in_c = _transform(seqs_in_c)
return seqs_in_c
|
dict_seq_locus
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(list_c, loci_obj, seq_obj):\n seqs = defaultdict(set)\n for c in list_c.values():\n for l in c.loci2seq:\n [seqs[s].add(c.id) for s in c.loci2seq[l]]\n common = [s for s in seqs if len(seqs[s]) > 1]\n seqs_in_c = defaultdict(float)\n for c in list_c.values():\n for l in c.loci2seq:\n for s in c.loci2seq[l]:\n if s in common:\n pos = seq_obj[s].pos[l]\n cov = 1.0 * loci_obj[l].coverage[pos]\n if seqs_in_c[s, c.id] < cov:\n seqs_in_c[s, c.id] = cov\n seqs_in_c = _transform(seqs_in_c)\n return seqs_in_c\n",
"role": "user"
},
{
"content": "dict_seq_locus",
"role": "assistant"
}
] |
def x(self, data):
for hypo in self.Values():
like = self.Likelihood(data, hypo)
self.Mult(hypo, like)
self.Normalize()
|
Update
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, data):\n for hypo in self.Values():\n like = self.Likelihood(data, hypo)\n self.Mult(hypo, like)\n self.Normalize()\n",
"role": "user"
},
{
"content": "Update",
"role": "assistant"
}
] |
def x(self, data, hypo):
mix = self.loci[hypo]
like = mix[data]
return like
|
Likelihood
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, data, hypo):\n mix = self.loci[hypo]\n like = mix[data]\n return like\n",
"role": "user"
},
{
"content": "Likelihood",
"role": "assistant"
}
] |
def x(clus_obj, index):
current = clus_obj.clus
clus_seqt = clus_obj.seq
clus_locit = clus_obj.loci
itern = 0
for idc in current.keys():
itern += 1
timestamp = str(idc)
seqListTemp = ()
f = open('/tmp/' + timestamp + '.fa', 'w')
for idl in current[idc].loci2seq.keys():
seqListTemp = list(set(seqListTemp).union(current[idc].loci2seq
[idl]))
maxscore = 0
for s in seqListTemp:
score = calculate_size(clus_seqt[s].freq)
maxscore = max(maxscore, score)
clus_seqt[s].score = score
seq = clus_seqt[s]
f.write('>' + s + '\n' + seq.seq + '\n')
f.close()
locilen_sorted = sorted(current[idc].locilen.iteritems(), key=
operator.itemgetter(1), reverse=True)
lmax = clus_locit[locilen_sorted[0][0]]
f = open('/tmp/' + timestamp + '.bed', 'w')
f.write('%s\t%s\t%s\t.\t.\t%s\n' % (lmax.chr, lmax.start, lmax.end,
lmax.strand))
f.close()
os.system('bedtools getfasta -s -fi ' + index + ' -bed /tmp/' +
timestamp + '.bed -fo /tmp/' + timestamp + '.pre.fa')
os.system('bowtie2-build /tmp/' + timestamp + '.pre.fa /tmp/' +
timestamp + '.pre.ind >/dev/null 2>&1')
os.system(
'bowtie2 --rdg 7,3 --mp 4 --end-to-end --no-head --no-sq -D 20 -R 3 -N 0 -i S,1,0.8 -L 3 -f /tmp/'
+ timestamp + '.pre.ind /tmp/' + timestamp + '.fa -S /tmp/' +
timestamp + '.map >>bowtie.log 2>&1')
f = open('/tmp/' + timestamp + '.map', 'r')
seqpos = {}
minv = 10000000
for line in f:
line = line.strip()
cols = line.split('\t')
seqpos[cols[0]] = int(cols[3])
if minv > int(cols[3]):
minv = int(cols[3])
f.close()
seqpos_sorted = sorted(seqpos.iteritems(), key=operator.itemgetter(
1), reverse=False)
showseq = ''
showseq_plain = ''
for s, pos in seqpos_sorted:
ratio = clus_seqt[s].score * 1.0 / maxscore * 100.0
realScore = math.log(ratio, 2) * 2
if realScore < 0:
realScore = 0
showseq_plain += (
'<br>%s<a style = "font-size:%.0fpx;"href = javascript:loadSeq("%s")>%s</a>'
% (''.join('.' for i in range(pos - 1)), realScore + 10, s,
clus_seqt[s].seq))
current[idc].showseq_plain = showseq_plain
os.system('rm /tmp/' + timestamp + '*')
clus_obj.clus = current
clus_obj.seq = clus_seqt
return clus_obj
|
show_seq
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(clus_obj, index):\n current = clus_obj.clus\n clus_seqt = clus_obj.seq\n clus_locit = clus_obj.loci\n itern = 0\n for idc in current.keys():\n itern += 1\n timestamp = str(idc)\n seqListTemp = ()\n f = open('/tmp/' + timestamp + '.fa', 'w')\n for idl in current[idc].loci2seq.keys():\n seqListTemp = list(set(seqListTemp).union(current[idc].loci2seq\n [idl]))\n maxscore = 0\n for s in seqListTemp:\n score = calculate_size(clus_seqt[s].freq)\n maxscore = max(maxscore, score)\n clus_seqt[s].score = score\n seq = clus_seqt[s]\n f.write('>' + s + '\\n' + seq.seq + '\\n')\n f.close()\n locilen_sorted = sorted(current[idc].locilen.iteritems(), key=\n operator.itemgetter(1), reverse=True)\n lmax = clus_locit[locilen_sorted[0][0]]\n f = open('/tmp/' + timestamp + '.bed', 'w')\n f.write('%s\\t%s\\t%s\\t.\\t.\\t%s\\n' % (lmax.chr, lmax.start, lmax.end,\n lmax.strand))\n f.close()\n os.system('bedtools getfasta -s -fi ' + index + ' -bed /tmp/' +\n timestamp + '.bed -fo /tmp/' + timestamp + '.pre.fa')\n os.system('bowtie2-build /tmp/' + timestamp + '.pre.fa /tmp/' +\n timestamp + '.pre.ind >/dev/null 2>&1')\n os.system(\n 'bowtie2 --rdg 7,3 --mp 4 --end-to-end --no-head --no-sq -D 20 -R 3 -N 0 -i S,1,0.8 -L 3 -f /tmp/'\n + timestamp + '.pre.ind /tmp/' + timestamp + '.fa -S /tmp/' +\n timestamp + '.map >>bowtie.log 2>&1')\n f = open('/tmp/' + timestamp + '.map', 'r')\n seqpos = {}\n minv = 10000000\n for line in f:\n line = line.strip()\n cols = line.split('\\t')\n seqpos[cols[0]] = int(cols[3])\n if minv > int(cols[3]):\n minv = int(cols[3])\n f.close()\n seqpos_sorted = sorted(seqpos.iteritems(), key=operator.itemgetter(\n 1), reverse=False)\n showseq = ''\n showseq_plain = ''\n for s, pos in seqpos_sorted:\n ratio = clus_seqt[s].score * 1.0 / maxscore * 100.0\n realScore = math.log(ratio, 2) * 2\n if realScore < 0:\n realScore = 0\n showseq_plain += (\n '<br>%s<a style = \"font-size:%.0fpx;\"href = javascript:loadSeq(\"%s\")>%s</a>'\n % (''.join('.' for i in range(pos - 1)), realScore + 10, s,\n clus_seqt[s].seq))\n current[idc].showseq_plain = showseq_plain\n os.system('rm /tmp/' + timestamp + '*')\n clus_obj.clus = current\n clus_obj.seq = clus_seqt\n return clus_obj\n",
"role": "user"
},
{
"content": "show_seq",
"role": "assistant"
}
] |
def x(s, t):
for ids in s:
obj = s[ids]
[obj.norm_freq.update({sample: 1.0 * obj.freq[sample] / (t[sample] +
1) * 1000000}) for sample in obj.norm_freq]
s[ids] = obj
return s
|
normalize_seqs
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(s, t):\n for ids in s:\n obj = s[ids]\n [obj.norm_freq.update({sample: 1.0 * obj.freq[sample] / (t[sample] +\n 1) * 1000000}) for sample in obj.norm_freq]\n s[ids] = obj\n return s\n",
"role": "user"
},
{
"content": "normalize_seqs",
"role": "assistant"
}
] |
def x(bam_in, precursors):
a = pybedtools.BedTool(bam_in)
b = pybedtools.BedTool(precursors)
c = a.intersect(b, u=True)
out_file = utils.splitext_plus(op.basename(bam_in))[0] + '_clean.bam'
c.saveas(out_file)
return op.abspath(out_file)
|
prepare_bam
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(bam_in, precursors):\n a = pybedtools.BedTool(bam_in)\n b = pybedtools.BedTool(precursors)\n c = a.intersect(b, u=True)\n out_file = utils.splitext_plus(op.basename(bam_in))[0] + '_clean.bam'\n c.saveas(out_file)\n return op.abspath(out_file)\n",
"role": "user"
},
{
"content": "prepare_bam",
"role": "assistant"
}
] |
def x(bed_file):
new_bed = utils.splitext_plus(bed_file)[0] + '_order.bed'
with open(bed_file) as in_handle:
with open(new_bed, 'w') as out_handle:
for line in in_handle:
cols = line.strip().split('\t')
cols[3] = _select_anno(cols[3]) + '_' + cols[4]
cols[4] = '0'
print('\t'.join(cols), file=out_handle, end='')
return new_bed
|
reorder_columns
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(bed_file):\n new_bed = utils.splitext_plus(bed_file)[0] + '_order.bed'\n with open(bed_file) as in_handle:\n with open(new_bed, 'w') as out_handle:\n for line in in_handle:\n cols = line.strip().split('\\t')\n cols[3] = _select_anno(cols[3]) + '_' + cols[4]\n cols[4] = '0'\n print('\\t'.join(cols), file=out_handle, end='')\n return new_bed\n",
"role": "user"
},
{
"content": "reorder_columns",
"role": "assistant"
}
] |
def x(cov_file):
new_cov = utils.splitext_plus(cov_file)[0] + '_fix.cov'
with open(cov_file) as in_handle:
with open(new_cov, 'w') as out_handle:
for line in in_handle:
cols = line.strip().split('\t')
cols[4] = cols[6]
print('\t'.join(cols[0:6]), file=out_handle, end='')
return new_cov
|
fix_score_column
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cov_file):\n new_cov = utils.splitext_plus(cov_file)[0] + '_fix.cov'\n with open(cov_file) as in_handle:\n with open(new_cov, 'w') as out_handle:\n for line in in_handle:\n cols = line.strip().split('\\t')\n cols[4] = cols[6]\n print('\\t'.join(cols[0:6]), file=out_handle, end='')\n return new_cov\n",
"role": "user"
},
{
"content": "fix_score_column",
"role": "assistant"
}
] |
def x(bam_in, bed_file, out_dir, prefix):
bed_file = _reorder_columns(bed_file)
counts_reads_cmd = (
'coverageBed -s -counts -b {bam_in} -a {bed_file} | sort -k4,4 > {out_dir}/loci.cov'
)
with utils.chdir(out_dir):
run(counts_reads_cmd.format(min_trimmed_read_len=
min_trimmed_read_len, max_trimmed_read_len=max_trimmed_read_len,
**locals()), 'Run counts_reads')
loci_file = _fix_score_column(op.join(out_dir, 'loci.cov'))
return loci_file
|
detect_regions
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(bam_in, bed_file, out_dir, prefix):\n bed_file = _reorder_columns(bed_file)\n counts_reads_cmd = (\n 'coverageBed -s -counts -b {bam_in} -a {bed_file} | sort -k4,4 > {out_dir}/loci.cov'\n )\n with utils.chdir(out_dir):\n run(counts_reads_cmd.format(min_trimmed_read_len=\n min_trimmed_read_len, max_trimmed_read_len=max_trimmed_read_len,\n **locals()), 'Run counts_reads')\n loci_file = _fix_score_column(op.join(out_dir, 'loci.cov'))\n return loci_file\n",
"role": "user"
},
{
"content": "detect_regions",
"role": "assistant"
}
] |
def x(cov_file, min_reads):
new_cov = op.join(op.dirname(cov_file), 'feat_antisense.txt')
with open(cov_file) as in_handle:
with open(new_cov, 'w') as out_handle:
print('name\tantisense', file=out_handle, end='')
for line in in_handle:
cols = line.strip().split('\t')
cols[6] = 0 if cols[6] < min_reads else cols[6]
print('%s\t%s' % (cols[3], cols[6]), file=out_handle, end='')
return new_cov
|
order_antisense_column
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cov_file, min_reads):\n new_cov = op.join(op.dirname(cov_file), 'feat_antisense.txt')\n with open(cov_file) as in_handle:\n with open(new_cov, 'w') as out_handle:\n print('name\\tantisense', file=out_handle, end='')\n for line in in_handle:\n cols = line.strip().split('\\t')\n cols[6] = 0 if cols[6] < min_reads else cols[6]\n print('%s\\t%s' % (cols[3], cols[6]), file=out_handle, end='')\n return new_cov\n",
"role": "user"
},
{
"content": "order_antisense_column",
"role": "assistant"
}
] |
def x(bam_in, loci_file, out_dir):
data = Counter()
a = pybedtools.BedTool(bam_in)
b = pybedtools.BedTool(loci_file)
c = a.intersect(b, s=True, bed=True, wo=True)
for line in c:
end = int(line[1]) + 1 + int(line[2]) if line[5] == '+' else int(line
[1]) + 1
start = int(line[1]) + 1 if line[5] == '+' else int(line[1]) + 1 + int(
line[2])
side5 = '%s\t5p\t%s' % (line[15], start)
side3 = '%s\t3p\t%s' % (line[15], end)
data[side5] += 1
data[side3] += 1
counts_reads = op.join(out_dir, 'locus_readpos.counts')
with open(counts_reads, 'w') as out_handle:
for k in data:
print(k, file=out_handle, end='')
return counts_reads
|
reads_per_position
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(bam_in, loci_file, out_dir):\n data = Counter()\n a = pybedtools.BedTool(bam_in)\n b = pybedtools.BedTool(loci_file)\n c = a.intersect(b, s=True, bed=True, wo=True)\n for line in c:\n end = int(line[1]) + 1 + int(line[2]) if line[5] == '+' else int(line\n [1]) + 1\n start = int(line[1]) + 1 if line[5] == '+' else int(line[1]) + 1 + int(\n line[2])\n side5 = '%s\\t5p\\t%s' % (line[15], start)\n side3 = '%s\\t3p\\t%s' % (line[15], end)\n data[side5] += 1\n data[side3] += 1\n counts_reads = op.join(out_dir, 'locus_readpos.counts')\n with open(counts_reads, 'w') as out_handle:\n for k in data:\n print(k, file=out_handle, end='')\n return counts_reads\n",
"role": "user"
},
{
"content": "reads_per_position",
"role": "assistant"
}
] |
def x(bam_in, loci_file, reference, out_dir):
lenvec_plus = op.join(out_dir, 'genomic_lenvec.plus')
lenvec_minus = op.join(out_dir, 'genomic_lenvec.minus')
compute_genomic_cmd = (
'compute_genomic_lenvectors {bam_in} {lenvec_plus} {lenvec_minus} {min_len} {max_len} '
)
index_genomic_cmd = 'index_genomic_lenvectors {lenvec} '
genomic_lenvec = op.join(out_dir, 'genomic_lenvec')
feat_len_file = op.join(out_dir, 'feat_lengths.txt')
compute_locus_cmd = (
'compute_locus_lenvectors {loci_file} {genomic_lenvec} {min_len} {max_len} > {feat_len_file}'
)
cov_S_file = op.join(out_dir, 'loci.cov_anti')
coverage_anti_cmd = (
'coverageBed -S -counts -b {bam_in} -a {loci_file} > {cov_S_file}')
feat_posentropy = op.join(out_dir, 'feat_posentropy.txt')
entropy_cmd = 'compute_locus_entropy.rb {counts_reads} > {feat_posentropy}'
with utils.chdir(out_dir):
run(compute_genomic_cmd.format(min_len=min_trimmed_read_len,
max_len=max_trimmed_read_len, **locals()), 'Run compute_genomic')
run(index_genomic_cmd.format(lenvec=lenvec_plus), 'Run index in plus')
run(index_genomic_cmd.format(lenvec=lenvec_minus), 'Run index in minus'
)
run(compute_locus_cmd.format(min_len=min_trimmed_read_len, max_len=
max_trimmed_read_len, **locals()), 'Run compute locus')
run(coverage_anti_cmd.format(**locals()), 'Run coverage antisense')
feat_antisense = _order_antisense_column(cov_S_file,
min_trimmed_read_len)
counts_reads = _reads_per_position(bam_in, loci_file, out_dir)
run(entropy_cmd.format(**locals()), 'Run entropy')
rnafold = calculate_structure(loci_file, reference)
|
create_features
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(bam_in, loci_file, reference, out_dir):\n lenvec_plus = op.join(out_dir, 'genomic_lenvec.plus')\n lenvec_minus = op.join(out_dir, 'genomic_lenvec.minus')\n compute_genomic_cmd = (\n 'compute_genomic_lenvectors {bam_in} {lenvec_plus} {lenvec_minus} {min_len} {max_len} '\n )\n index_genomic_cmd = 'index_genomic_lenvectors {lenvec} '\n genomic_lenvec = op.join(out_dir, 'genomic_lenvec')\n feat_len_file = op.join(out_dir, 'feat_lengths.txt')\n compute_locus_cmd = (\n 'compute_locus_lenvectors {loci_file} {genomic_lenvec} {min_len} {max_len} > {feat_len_file}'\n )\n cov_S_file = op.join(out_dir, 'loci.cov_anti')\n coverage_anti_cmd = (\n 'coverageBed -S -counts -b {bam_in} -a {loci_file} > {cov_S_file}')\n feat_posentropy = op.join(out_dir, 'feat_posentropy.txt')\n entropy_cmd = 'compute_locus_entropy.rb {counts_reads} > {feat_posentropy}'\n with utils.chdir(out_dir):\n run(compute_genomic_cmd.format(min_len=min_trimmed_read_len,\n max_len=max_trimmed_read_len, **locals()), 'Run compute_genomic')\n run(index_genomic_cmd.format(lenvec=lenvec_plus), 'Run index in plus')\n run(index_genomic_cmd.format(lenvec=lenvec_minus), 'Run index in minus'\n )\n run(compute_locus_cmd.format(min_len=min_trimmed_read_len, max_len=\n max_trimmed_read_len, **locals()), 'Run compute locus')\n run(coverage_anti_cmd.format(**locals()), 'Run coverage antisense')\n feat_antisense = _order_antisense_column(cov_S_file,\n min_trimmed_read_len)\n counts_reads = _reads_per_position(bam_in, loci_file, out_dir)\n run(entropy_cmd.format(**locals()), 'Run entropy')\n rnafold = calculate_structure(loci_file, reference)\n",
"role": "user"
},
{
"content": "create_features",
"role": "assistant"
}
] |
def x(args):
logger.info('reading sequeces')
data = load_data(args.json)
logger.info('create profile')
data = make_profile(data, os.path.join(args.out, 'profiles'), args)
logger.info('create database')
make_database(data, 'seqcluster.db', args.out)
logger.info(
'Done. Download https://github.com/lpantano/seqclusterViz/archive/master.zip to browse the output.'
)
|
report
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(args):\n logger.info('reading sequeces')\n data = load_data(args.json)\n logger.info('create profile')\n data = make_profile(data, os.path.join(args.out, 'profiles'), args)\n logger.info('create database')\n make_database(data, 'seqcluster.db', args.out)\n logger.info(\n 'Done. Download https://github.com/lpantano/seqclusterViz/archive/master.zip to browse the output.'\n )\n",
"role": "user"
},
{
"content": "report",
"role": "assistant"
}
] |
def x(peaks):
previous = peaks[0]
new_peaks = [previous]
for pos in peaks:
if pos > previous + 10:
new_peaks.add(pos)
previous = pos
return new_peaks
|
summarize_peaks
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(peaks):\n previous = peaks[0]\n new_peaks = [previous]\n for pos in peaks:\n if pos > previous + 10:\n new_peaks.add(pos)\n previous = pos\n return new_peaks\n",
"role": "user"
},
{
"content": "summarize_peaks",
"role": "assistant"
}
] |
def x(x, y, win=10):
previous = min(y)
peaks = []
intervals = range(x, y, win)
for pos in intervals:
if y[pos] > previous * 10:
previous = y[pos]
peaks.add(pos)
peaks = _summarize_peaks(peaks)
|
find_mature
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(x, y, win=10):\n previous = min(y)\n peaks = []\n intervals = range(x, y, win)\n for pos in intervals:\n if y[pos] > previous * 10:\n previous = y[pos]\n peaks.add(pos)\n peaks = _summarize_peaks(peaks)\n",
"role": "user"
},
{
"content": "find_mature",
"role": "assistant"
}
] |
def x(in_file):
keep = Counter()
with open_fastq(in_file) as handle:
for line in handle:
if line.startswith('@'):
if line.find('UMI') > -1:
logger.info(
'Find UMI tags in read names, collapsing by UMI.')
return collapse_umi(in_file)
seq = handle.next().strip()
handle.next()
qual = handle.next().strip()
if seq in keep:
keep[seq].update(qual)
else:
keep[seq] = quality(qual)
logger.info('Sequences loaded: %s' % len(keep))
return keep
|
collapse
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(in_file):\n keep = Counter()\n with open_fastq(in_file) as handle:\n for line in handle:\n if line.startswith('@'):\n if line.find('UMI') > -1:\n logger.info(\n 'Find UMI tags in read names, collapsing by UMI.')\n return collapse_umi(in_file)\n seq = handle.next().strip()\n handle.next()\n qual = handle.next().strip()\n if seq in keep:\n keep[seq].update(qual)\n else:\n keep[seq] = quality(qual)\n logger.info('Sequences loaded: %s' % len(keep))\n return keep\n",
"role": "user"
},
{
"content": "collapse",
"role": "assistant"
}
] |
def x(in_file):
keep = defaultdict(dict)
with open_fastq(in_file) as handle:
for line in handle:
if line.startswith('@'):
m = re.search('UMI_([ATGC]*)', line.strip())
umis = m.group(0)
seq = handle.next().strip()
handle.next()
qual = handle.next().strip()
if (umis, seq) in keep:
keep[umis, seq][1].update(qual)
keep[umis, seq][0].update(seq)
else:
keep[umis, seq] = [umi(seq), quality(qual)]
logger.info('Sequences loaded: %s' % len(keep))
return keep
|
collapse_umi
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(in_file):\n keep = defaultdict(dict)\n with open_fastq(in_file) as handle:\n for line in handle:\n if line.startswith('@'):\n m = re.search('UMI_([ATGC]*)', line.strip())\n umis = m.group(0)\n seq = handle.next().strip()\n handle.next()\n qual = handle.next().strip()\n if (umis, seq) in keep:\n keep[umis, seq][1].update(qual)\n keep[umis, seq][0].update(seq)\n else:\n keep[umis, seq] = [umi(seq), quality(qual)]\n logger.info('Sequences loaded: %s' % len(keep))\n return keep\n",
"role": "user"
},
{
"content": "collapse_umi",
"role": "assistant"
}
] |
def x(in_file):
_, ext = os.path.splitext(in_file)
if ext == '.gz':
return gzip.open(in_file, 'rb')
if ext in ['.fastq', '.fq', '.fasta', '.fa']:
return open(in_file, 'r')
return ValueError('File needs to be fastq|fasta|fq|fa [.gz]')
|
open_fastq
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(in_file):\n _, ext = os.path.splitext(in_file)\n if ext == '.gz':\n return gzip.open(in_file, 'rb')\n if ext in ['.fastq', '.fq', '.fasta', '.fa']:\n return open(in_file, 'r')\n return ValueError('File needs to be fastq|fasta|fq|fa [.gz]')\n",
"role": "user"
},
{
"content": "open_fastq",
"role": "assistant"
}
] |
def x(args):
try:
umi_fn = args.fastq
if _is_umi(args.fastq):
umis = collapse(args.fastq)
umi_fn = os.path.join(args.out, splitext_plus(os.path.basename(
args.fastq))[0] + '_umi_trimmed.fastq')
write_output(umi_fn, umis, args.minimum)
seqs = collapse(umi_fn)
out_file = splitext_plus(os.path.basename(args.fastq))[0
] + '_trimmed.fastq'
except IOError as e:
logger.error('I/O error({0}): {1}'.format(e.errno, e.strerror))
raise 'Can not read file'
out_file = os.path.join(args.out, out_file)
write_output(out_file, seqs, args.minimum)
return out_file
|
collapse_fastq
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(args):\n try:\n umi_fn = args.fastq\n if _is_umi(args.fastq):\n umis = collapse(args.fastq)\n umi_fn = os.path.join(args.out, splitext_plus(os.path.basename(\n args.fastq))[0] + '_umi_trimmed.fastq')\n write_output(umi_fn, umis, args.minimum)\n seqs = collapse(umi_fn)\n out_file = splitext_plus(os.path.basename(args.fastq))[0\n ] + '_trimmed.fastq'\n except IOError as e:\n logger.error('I/O error({0}): {1}'.format(e.errno, e.strerror))\n raise 'Can not read file'\n out_file = os.path.join(args.out, out_file)\n write_output(out_file, seqs, args.minimum)\n return out_file\n",
"role": "user"
},
{
"content": "collapse_fastq",
"role": "assistant"
}
] |
def x(doctree):
current = 0
num_children = len(doctree.children)
while current < num_children:
child = doctree.children[current]
child.replace_self(child.traverse(no_autoslides_filter))
if len(doctree.children) == num_children:
current += 1
else:
num_children = len(doctree.children)
|
filter_doctree_for_slides
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(doctree):\n current = 0\n num_children = len(doctree.children)\n while current < num_children:\n child = doctree.children[current]\n child.replace_self(child.traverse(no_autoslides_filter))\n if len(doctree.children) == num_children:\n current += 1\n else:\n num_children = len(doctree.children)\n",
"role": "user"
},
{
"content": "filter_doctree_for_slides",
"role": "assistant"
}
] |
def x(self, node, increment=True):
parent_title_node = node.parent.next_node(nodes.title)
nextslide_info = getattr(parent_title_node, 'nextslide_info', (
parent_title_node.deepcopy().children, 1))
nextslide_info = nextslide_info[0], nextslide_info[1] + 1
if node.args:
textnodes, messages = node.state.inline_text(node.args[0], 1)
new_title = nodes.title(node.args[0], '', *textnodes)
else:
title_nodes = nextslide_info[0][:]
if 'increment' in node.attributes:
title_nodes.append(nodes.Text(' (%s)' % nextslide_info[1]))
new_title = nodes.title('', '', *title_nodes)
new_title.nextslide_info = nextslide_info
return new_title
|
make_title_node
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, node, increment=True):\n parent_title_node = node.parent.next_node(nodes.title)\n nextslide_info = getattr(parent_title_node, 'nextslide_info', (\n parent_title_node.deepcopy().children, 1))\n nextslide_info = nextslide_info[0], nextslide_info[1] + 1\n if node.args:\n textnodes, messages = node.state.inline_text(node.args[0], 1)\n new_title = nodes.title(node.args[0], '', *textnodes)\n else:\n title_nodes = nextslide_info[0][:]\n if 'increment' in node.attributes:\n title_nodes.append(nodes.Text(' (%s)' % nextslide_info[1]))\n new_title = nodes.title('', '', *title_nodes)\n new_title.nextslide_info = nextslide_info\n return new_title\n",
"role": "user"
},
{
"content": "make_title_node",
"role": "assistant"
}
] |
def x(self, builder):
if 'theme' in self.attributes:
builder.apply_theme(self.attributes['theme'], builder.theme_options)
|
apply
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, builder):\n if 'theme' in self.attributes:\n builder.apply_theme(self.attributes['theme'], builder.theme_options)\n",
"role": "user"
},
{
"content": "apply",
"role": "assistant"
}
] |
def x(cls, builder, doctree=None):
result = {'theme': builder.config.slide_theme, 'autoslides': builder.
config.autoslides, 'slide_classes': []}
if doctree:
conf_node = cls.get(doctree)
if conf_node:
result.update(conf_node.attributes)
return result
|
get_conf
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cls, builder, doctree=None):\n result = {'theme': builder.config.slide_theme, 'autoslides': builder.\n config.autoslides, 'slide_classes': []}\n if doctree:\n conf_node = cls.get(doctree)\n if conf_node:\n result.update(conf_node.attributes)\n return result\n",
"role": "user"
},
{
"content": "get_conf",
"role": "assistant"
}
] |
def x(context):
COPY_LISTS = 'script_files', 'css_files'
for attr in COPY_LISTS:
if attr in context:
context[attr] = context[attr][:]
return context
|
fix_context
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(context):\n COPY_LISTS = 'script_files', 'css_files'\n for attr in COPY_LISTS:\n if attr in context:\n context[attr] = context[attr][:]\n return context\n",
"role": "user"
},
{
"content": "fix_context",
"role": "assistant"
}
] |
def x(d):
msg = bold('Welcome to the Hieroglyph %s quickstart utility.') % (version
(),)
print(msg)
msg = """
This will ask questions for creating a Hieroglyph project, and then ask
some basic Sphinx questions.
"""
print(msg)
d.update({'version': datetime.date.today().strftime('%Y.%m.%d'),
'release': datetime.date.today().strftime('%Y.%m.%d'), 'make_mode':
True})
if 'project' not in d:
print('\nThe presentation title will be included on the title slide.')
sphinx.quickstart.do_prompt(d, 'project', 'Presentation title')
if 'author' not in d:
sphinx.quickstart.do_prompt(d, 'author', 'Author name(s)')
theme_entrypoints = pkg_resources.iter_entry_points('hieroglyph.theme')
themes = [t.load() for t in theme_entrypoints]
msg = '\nAvailable themes:\n\n'
for theme in themes:
msg += '\n'.join([bold(theme['name']), theme['desc'], '', ''])
msg += 'Which theme would you like to use?'
print(msg)
sphinx.quickstart.do_prompt(d, 'slide_theme', 'Slide Theme', themes[0][
'name'], sphinx.quickstart.choice(*[t['name'] for t in themes]))
print('')
sphinx.quickstart.ask_user(d)
|
ask_user
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(d):\n msg = bold('Welcome to the Hieroglyph %s quickstart utility.') % (version\n (),)\n print(msg)\n msg = \"\"\"\nThis will ask questions for creating a Hieroglyph project, and then ask\nsome basic Sphinx questions.\n\"\"\"\n print(msg)\n d.update({'version': datetime.date.today().strftime('%Y.%m.%d'),\n 'release': datetime.date.today().strftime('%Y.%m.%d'), 'make_mode':\n True})\n if 'project' not in d:\n print('\\nThe presentation title will be included on the title slide.')\n sphinx.quickstart.do_prompt(d, 'project', 'Presentation title')\n if 'author' not in d:\n sphinx.quickstart.do_prompt(d, 'author', 'Author name(s)')\n theme_entrypoints = pkg_resources.iter_entry_points('hieroglyph.theme')\n themes = [t.load() for t in theme_entrypoints]\n msg = '\\nAvailable themes:\\n\\n'\n for theme in themes:\n msg += '\\n'.join([bold(theme['name']), theme['desc'], '', ''])\n msg += 'Which theme would you like to use?'\n print(msg)\n sphinx.quickstart.do_prompt(d, 'slide_theme', 'Slide Theme', themes[0][\n 'name'], sphinx.quickstart.choice(*[t['name'] for t in themes]))\n print('')\n sphinx.quickstart.ask_user(d)\n",
"role": "user"
},
{
"content": "ask_user",
"role": "assistant"
}
] |
def x(self):
return {'title': self.title, 'level': self.level, 'content': self.
content, 'classes': self.classes, 'slide_classes': self.
_filter_classes(exclude='content-'), 'content_classes': self.
_filter_classes(include='content-'), 'slide_number': self.
slide_number, 'config': self._translator.builder.config, 'id': self.id}
|
get_slide_context
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n return {'title': self.title, 'level': self.level, 'content': self.\n content, 'classes': self.classes, 'slide_classes': self.\n _filter_classes(exclude='content-'), 'content_classes': self.\n _filter_classes(include='content-'), 'slide_number': self.\n slide_number, 'config': self._translator.builder.config, 'id': self.id}\n",
"role": "user"
},
{
"content": "get_slide_context",
"role": "assistant"
}
] |
def x(self, slide_no):
if self.builder.config.slide_numbers:
self.body.append('\n<div class="slide-no">%s</div>\n' % (slide_no,))
|
add_slide_number
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, slide_no):\n if self.builder.config.slide_numbers:\n self.body.append('\\n<div class=\"slide-no\">%s</div>\\n' % (slide_no,))\n",
"role": "user"
},
{
"content": "add_slide_number",
"role": "assistant"
}
] |
def x(self, slide_no):
if self.builder.config.slide_footer:
self.body.append('\n<div class="slide-footer">%s</div>\n' % (self.
builder.config.slide_footer,))
|
add_slide_footer
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, slide_no):\n if self.builder.config.slide_footer:\n self.body.append('\\n<div class=\"slide-footer\">%s</div>\\n' % (self.\n builder.config.slide_footer,))\n",
"role": "user"
},
{
"content": "add_slide_footer",
"role": "assistant"
}
] |
def x(app):
from hieroglyph import writer
if app.builder.name not in HTML_BUILDERS:
return
if app.config.slide_link_html_to_slides:
app.builder.templates.loaders.append(SphinxFileSystemLoader(os.path
.join(os.path.dirname(__file__), 'themes', 'slides')))
if not app.config.html_sidebars:
app.config.html_sidebars = {'**': ['localtoc.html',
'relations.html', 'sourcelink.html', SLIDELINK_TEMPLATE,
'searchbox.html']}
else:
included = False
for glob, templates in app.config.html_sidebars:
if SLIDELINK_TEMPLATE in templates:
included = True
break
if not included:
app.config.html_sidebars.setdefault('**', []).append(
SLIDELINK_TEMPLATE)
if app.config.slide_link_html_sections_to_slides:
if sphinx.version_info >= (1, 6, 0):
override_translator = type('SlideLinkTranslator', (app.builder.
get_translator_class(), object), {'depart_title': writer.
depart_title})
app.set_translator(app.builder, override_translator)
else:
app.builder.translator_class = type('SlideLinkTranslator', (app
.builder.translator_class, object), {'depart_title': writer
.depart_title})
|
inspect_config
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(app):\n from hieroglyph import writer\n if app.builder.name not in HTML_BUILDERS:\n return\n if app.config.slide_link_html_to_slides:\n app.builder.templates.loaders.append(SphinxFileSystemLoader(os.path\n .join(os.path.dirname(__file__), 'themes', 'slides')))\n if not app.config.html_sidebars:\n app.config.html_sidebars = {'**': ['localtoc.html',\n 'relations.html', 'sourcelink.html', SLIDELINK_TEMPLATE,\n 'searchbox.html']}\n else:\n included = False\n for glob, templates in app.config.html_sidebars:\n if SLIDELINK_TEMPLATE in templates:\n included = True\n break\n if not included:\n app.config.html_sidebars.setdefault('**', []).append(\n SLIDELINK_TEMPLATE)\n if app.config.slide_link_html_sections_to_slides:\n if sphinx.version_info >= (1, 6, 0):\n override_translator = type('SlideLinkTranslator', (app.builder.\n get_translator_class(), object), {'depart_title': writer.\n depart_title})\n app.set_translator(app.builder, override_translator)\n else:\n app.builder.translator_class = type('SlideLinkTranslator', (app\n .builder.translator_class, object), {'depart_title': writer\n .depart_title})\n",
"role": "user"
},
{
"content": "inspect_config",
"role": "assistant"
}
] |
def x(builder, pagename=None):
return builder.get_relative_uri(pagename or builder.current_docname, os
.path.join(builder.app.config.slide_relative_path, pagename or
builder.current_docname))
|
slide_path
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(builder, pagename=None):\n return builder.get_relative_uri(pagename or builder.current_docname, os\n .path.join(builder.app.config.slide_relative_path, pagename or\n builder.current_docname))\n",
"role": "user"
},
{
"content": "slide_path",
"role": "assistant"
}
] |
def x(builder, pagename=None):
return builder.get_relative_uri(pagename or builder.current_docname, os
.path.join(builder.app.config.slide_html_relative_path, pagename or
builder.current_docname))
|
html_path
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(builder, pagename=None):\n return builder.get_relative_uri(pagename or builder.current_docname, os\n .path.join(builder.app.config.slide_html_relative_path, pagename or\n builder.current_docname))\n",
"role": "user"
},
{
"content": "html_path",
"role": "assistant"
}
] |
def x(app, pagename, templatename, context, doctree):
context['show_slidelink'
] = app.config.slide_link_html_to_slides and hasattr(app.builder,
'get_outfilename')
if context['show_slidelink']:
context['slide_path'] = slide_path(app.builder, pagename)
|
add_link
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(app, pagename, templatename, context, doctree):\n context['show_slidelink'\n ] = app.config.slide_link_html_to_slides and hasattr(app.builder,\n 'get_outfilename')\n if context['show_slidelink']:\n context['slide_path'] = slide_path(app.builder, pagename)\n",
"role": "user"
},
{
"content": "add_link",
"role": "assistant"
}
] |
def x(self, themename, themeoptions):
self._theme_stack.append((self.theme, self.theme_options))
theme_factory = HTMLThemeFactory(self.app)
theme_factory.load_additional_themes(self.get_builtin_theme_dirs() +
self.config.slide_theme_path)
self.theme = theme_factory.create(themename)
self.theme_options = themeoptions.copy()
self.templates.init(self, self.theme)
self.templates.environment.filters['json'] = json.dumps
if self.theme not in self._additional_themes:
self._additional_themes.append(self.theme)
|
apply_theme
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, themename, themeoptions):\n self._theme_stack.append((self.theme, self.theme_options))\n theme_factory = HTMLThemeFactory(self.app)\n theme_factory.load_additional_themes(self.get_builtin_theme_dirs() +\n self.config.slide_theme_path)\n self.theme = theme_factory.create(themename)\n self.theme_options = themeoptions.copy()\n self.templates.init(self, self.theme)\n self.templates.environment.filters['json'] = json.dumps\n if self.theme not in self._additional_themes:\n self._additional_themes.append(self.theme)\n",
"role": "user"
},
{
"content": "apply_theme",
"role": "assistant"
}
] |
def x(self, doctree):
super(AbstractSlideBuilder, self).post_process_images(doctree)
relative_base = ['..'] * doctree.attributes.get('source')[len(self.
srcdir) + 1:].count('/')
for node in doctree.traverse(nodes.image):
if node.get('candidates') is None:
node['candidates'] = '*',
if node['uri'].startswith(self.outdir):
node['uri'] = '/'.join(relative_base + [node['uri'][len(self.
outdir) + 1:]])
|
post_process_images
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, doctree):\n super(AbstractSlideBuilder, self).post_process_images(doctree)\n relative_base = ['..'] * doctree.attributes.get('source')[len(self.\n srcdir) + 1:].count('/')\n for node in doctree.traverse(nodes.image):\n if node.get('candidates') is None:\n node['candidates'] = '*',\n if node['uri'].startswith(self.outdir):\n node['uri'] = '/'.join(relative_base + [node['uri'][len(self.\n outdir) + 1:]])\n",
"role": "user"
},
{
"content": "post_process_images",
"role": "assistant"
}
] |
def x(section):
metadata = {}
metadata_lines = section.split('\n')
for line in metadata_lines:
colon_index = line.find(':')
if colon_index != -1:
key = line[:colon_index].strip()
val = line[colon_index + 1:].strip()
metadata[key] = val
return metadata
|
parse_metadata
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(section):\n metadata = {}\n metadata_lines = section.split('\\n')\n for line in metadata_lines:\n colon_index = line.find(':')\n if colon_index != -1:\n key = line[:colon_index].strip()\n val = line[colon_index + 1:].strip()\n metadata[key] = val\n return metadata\n",
"role": "user"
},
{
"content": "parse_metadata",
"role": "assistant"
}
] |
def x(html, metadata):
if metadata.get('build_lists') and metadata['build_lists'] == 'true':
html = html.replace('<ul>', '<ul class="build">')
html = html.replace('<ol>', '<ol class="build">')
return html
|
postprocess_html
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(html, metadata):\n if metadata.get('build_lists') and metadata['build_lists'] == 'true':\n html = html.replace('<ul>', '<ul class=\"build\">')\n html = html.replace('<ol>', '<ol class=\"build\">')\n return html\n",
"role": "user"
},
{
"content": "postprocess_html",
"role": "assistant"
}
] |
def x(linter, checker_method, augmentation):
if sys.version_info[0] <= 2:
checker = get_checker(linter, checker_method.im_class)
else:
try:
checker = get_checker(linter, checker_method.__self__.__class__)
except AttributeError:
checker = get_checker(linter, get_class(checker_method.
__module__, checker_method.__qualname__))
old_method = getattr(checker, checker_method.__name__)
def x(node):
def x():
old_method(node)
augmentation(chain, node)
setattr(checker, checker_method.__name__, augment_func)
|
augment_visit
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(linter, checker_method, augmentation):\n if sys.version_info[0] <= 2:\n checker = get_checker(linter, checker_method.im_class)\n else:\n try:\n checker = get_checker(linter, checker_method.__self__.__class__)\n except AttributeError:\n checker = get_checker(linter, get_class(checker_method.\n __module__, checker_method.__qualname__))\n old_method = getattr(checker, checker_method.__name__)\n\n def x(node):\n\n def x():\n old_method(node)\n augmentation(chain, node)\n setattr(checker, checker_method.__name__, augment_func)\n",
"role": "user"
},
{
"content": "augment_visit",
"role": "assistant"
}
] |
def x(linter, checker_method, message_id_or_symbol, test_func):
msgs_store = getattr(linter, 'msgs_store', linter)
def x(message_id_or_symbol):
if hasattr(msgs_store, 'check_message_id'):
return [msgs_store.check_message_id(message_id_or_symbol)]
elif hasattr(msgs_store, 'get_message_definition'):
return [msgs_store.get_message_definition(message_id_or_symbol)]
elif hasattr(msgs_store, 'get_message_definitions'):
return msgs_store.get_message_definitions(message_id_or_symbol)
else:
raise ValueError(
'pylint.utils.MessagesStore does not have a get_message_definition(s) method'
)
try:
pylint_messages = get_message_definitions(message_id_or_symbol)
symbols = [symbol for pylint_message in pylint_messages for symbol in
(pylint_message.msgid, pylint_message.symbol) if symbol is not None
]
except UnknownMessage:
symbols = [message_id_or_symbol]
def x(chain, node):
with Suppress(linter) as s:
if test_func(node):
s.suppress(*symbols)
chain()
augment_visit(linter, checker_method, do_suppress)
|
suppress_message
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(linter, checker_method, message_id_or_symbol, test_func):\n msgs_store = getattr(linter, 'msgs_store', linter)\n\n def x(message_id_or_symbol):\n if hasattr(msgs_store, 'check_message_id'):\n return [msgs_store.check_message_id(message_id_or_symbol)]\n elif hasattr(msgs_store, 'get_message_definition'):\n return [msgs_store.get_message_definition(message_id_or_symbol)]\n elif hasattr(msgs_store, 'get_message_definitions'):\n return msgs_store.get_message_definitions(message_id_or_symbol)\n else:\n raise ValueError(\n 'pylint.utils.MessagesStore does not have a get_message_definition(s) method'\n )\n try:\n pylint_messages = get_message_definitions(message_id_or_symbol)\n symbols = [symbol for pylint_message in pylint_messages for symbol in\n (pylint_message.msgid, pylint_message.symbol) if symbol is not None\n ]\n except UnknownMessage:\n symbols = [message_id_or_symbol]\n\n def x(chain, node):\n with Suppress(linter) as s:\n if test_func(node):\n s.suppress(*symbols)\n chain()\n augment_visit(linter, checker_method, do_suppress)\n",
"role": "user"
},
{
"content": "suppress_message",
"role": "assistant"
}
] |
def x(self, hash_values):
q = """SELECT threat_type,platform_type,threat_entry_type, expires_at < current_timestamp AS has_expired
FROM full_hash WHERE value IN ({})
"""
output = []
with self.get_cursor() as dbc:
placeholders = ','.join(['?'] * len(hash_values))
dbc.execute(q.format(placeholders), [sqlite3.Binary(hv) for hv in
hash_values])
for h in dbc.fetchall():
threat_type, platform_type, threat_entry_type, has_expired = h
threat_list = ThreatList(threat_type, platform_type,
threat_entry_type)
output.append((threat_list, has_expired))
return output
|
lookup_full_hashes
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, hash_values):\n q = \"\"\"SELECT threat_type,platform_type,threat_entry_type, expires_at < current_timestamp AS has_expired\n FROM full_hash WHERE value IN ({})\n \"\"\"\n output = []\n with self.get_cursor() as dbc:\n placeholders = ','.join(['?'] * len(hash_values))\n dbc.execute(q.format(placeholders), [sqlite3.Binary(hv) for hv in\n hash_values])\n for h in dbc.fetchall():\n threat_type, platform_type, threat_entry_type, has_expired = h\n threat_list = ThreatList(threat_type, platform_type,\n threat_entry_type)\n output.append((threat_list, has_expired))\n return output\n",
"role": "user"
},
{
"content": "lookup_full_hashes",
"role": "assistant"
}
] |
def x(self, cues):
q = """SELECT value, MAX(negative_expires_at < current_timestamp) AS negative_cache_expired
FROM hash_prefix WHERE cue IN ({}) GROUP BY 1
"""
output = []
with self.get_cursor() as dbc:
dbc.execute(q.format(','.join(['?'] * len(cues))), [sqlite3.Binary(
cue) for cue in cues])
for h in dbc.fetchall():
value, negative_cache_expired = h
output.append((bytes(value), negative_cache_expired))
return output
|
lookup_hash_prefix
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, cues):\n q = \"\"\"SELECT value, MAX(negative_expires_at < current_timestamp) AS negative_cache_expired\n FROM hash_prefix WHERE cue IN ({}) GROUP BY 1\n \"\"\"\n output = []\n with self.get_cursor() as dbc:\n dbc.execute(q.format(','.join(['?'] * len(cues))), [sqlite3.Binary(\n cue) for cue in cues])\n for h in dbc.fetchall():\n value, negative_cache_expired = h\n output.append((bytes(value), negative_cache_expired))\n return output\n",
"role": "user"
},
{
"content": "lookup_hash_prefix",
"role": "assistant"
}
] |
def x(self, threat_list, hash_value, cache_duration, malware_threat_type):
log.info('Storing full hash %s to list %s with cache duration %s',
to_hex(hash_value), str(threat_list), cache_duration)
qi = """INSERT OR IGNORE INTO full_hash
(value, threat_type, platform_type, threat_entry_type, malware_threat_type, downloaded_at)
VALUES
(?, ?, ?, ?, ?, current_timestamp)
"""
qu = (
"UPDATE full_hash SET expires_at=datetime(current_timestamp, '+{} SECONDS') WHERE value=? AND threat_type=? AND platform_type=? AND threat_entry_type=?"
)
i_parameters = [sqlite3.Binary(hash_value), threat_list.threat_type,
threat_list.platform_type, threat_list.threat_entry_type,
malware_threat_type]
u_parameters = [sqlite3.Binary(hash_value), threat_list.threat_type,
threat_list.platform_type, threat_list.threat_entry_type]
with self.get_cursor() as dbc:
dbc.execute(qi, i_parameters)
dbc.execute(qu.format(int(cache_duration)), u_parameters)
|
store_full_hash
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, threat_list, hash_value, cache_duration, malware_threat_type):\n log.info('Storing full hash %s to list %s with cache duration %s',\n to_hex(hash_value), str(threat_list), cache_duration)\n qi = \"\"\"INSERT OR IGNORE INTO full_hash\n (value, threat_type, platform_type, threat_entry_type, malware_threat_type, downloaded_at)\n VALUES\n (?, ?, ?, ?, ?, current_timestamp)\n \"\"\"\n qu = (\n \"UPDATE full_hash SET expires_at=datetime(current_timestamp, '+{} SECONDS') WHERE value=? AND threat_type=? AND platform_type=? AND threat_entry_type=?\"\n )\n i_parameters = [sqlite3.Binary(hash_value), threat_list.threat_type,\n threat_list.platform_type, threat_list.threat_entry_type,\n malware_threat_type]\n u_parameters = [sqlite3.Binary(hash_value), threat_list.threat_type,\n threat_list.platform_type, threat_list.threat_entry_type]\n with self.get_cursor() as dbc:\n dbc.execute(qi, i_parameters)\n dbc.execute(qu.format(int(cache_duration)), u_parameters)\n",
"role": "user"
},
{
"content": "store_full_hash",
"role": "assistant"
}
] |
def x(self, keep_expired_for=60 * 60 * 12):
q = """DELETE FROM full_hash WHERE expires_at < datetime(current_timestamp, '-{} SECONDS')
"""
log.info('Cleaning up full_hash entries expired more than {} seconds ago.'
.format(keep_expired_for))
with self.get_cursor() as dbc:
dbc.execute(q.format(int(keep_expired_for)))
|
cleanup_full_hashes
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, keep_expired_for=60 * 60 * 12):\n q = \"\"\"DELETE FROM full_hash WHERE expires_at < datetime(current_timestamp, '-{} SECONDS')\n \"\"\"\n log.info('Cleaning up full_hash entries expired more than {} seconds ago.'\n .format(keep_expired_for))\n with self.get_cursor() as dbc:\n dbc.execute(q.format(int(keep_expired_for)))\n",
"role": "user"
},
{
"content": "cleanup_full_hashes",
"role": "assistant"
}
] |
def x(self):
q = 'SELECT threat_type,platform_type,threat_entry_type FROM threat_list'
output = []
with self.get_cursor() as dbc:
dbc.execute(q)
for h in dbc.fetchall():
threat_type, platform_type, threat_entry_type = h
threat_list = ThreatList(threat_type, platform_type,
threat_entry_type)
output.append(threat_list)
return output
|
get_threat_lists
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n q = 'SELECT threat_type,platform_type,threat_entry_type FROM threat_list'\n output = []\n with self.get_cursor() as dbc:\n dbc.execute(q)\n for h in dbc.fetchall():\n threat_type, platform_type, threat_entry_type = h\n threat_list = ThreatList(threat_type, platform_type,\n threat_entry_type)\n output.append(threat_list)\n return output\n",
"role": "user"
},
{
"content": "get_threat_lists",
"role": "assistant"
}
] |
def x(self):
q = (
'SELECT threat_type,platform_type,threat_entry_type,client_state FROM threat_list'
)
output = {}
with self.get_cursor() as dbc:
dbc.execute(q)
for h in dbc.fetchall():
threat_type, platform_type, threat_entry_type, client_state = h
threat_list_tuple = threat_type, platform_type, threat_entry_type
output[threat_list_tuple] = client_state
return output
|
get_client_state
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n q = (\n 'SELECT threat_type,platform_type,threat_entry_type,client_state FROM threat_list'\n )\n output = {}\n with self.get_cursor() as dbc:\n dbc.execute(q)\n for h in dbc.fetchall():\n threat_type, platform_type, threat_entry_type, client_state = h\n threat_list_tuple = threat_type, platform_type, threat_entry_type\n output[threat_list_tuple] = client_state\n return output\n",
"role": "user"
},
{
"content": "get_client_state",
"role": "assistant"
}
] |
def x(self, threat_list):
q = """INSERT OR IGNORE INTO threat_list
(threat_type, platform_type, threat_entry_type, timestamp)
VALUES
(?, ?, ?, current_timestamp)
"""
params = [threat_list.threat_type, threat_list.platform_type,
threat_list.threat_entry_type]
with self.get_cursor() as dbc:
dbc.execute(q, params)
|
add_threat_list
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, threat_list):\n q = \"\"\"INSERT OR IGNORE INTO threat_list\n (threat_type, platform_type, threat_entry_type, timestamp)\n VALUES\n (?, ?, ?, current_timestamp)\n \"\"\"\n params = [threat_list.threat_type, threat_list.platform_type,\n threat_list.threat_entry_type]\n with self.get_cursor() as dbc:\n dbc.execute(q, params)\n",
"role": "user"
},
{
"content": "add_threat_list",
"role": "assistant"
}
] |
def x(self, threat_list):
log.info('Deleting cached threat list "{}"'.format(repr(threat_list)))
q = """DELETE FROM threat_list
WHERE threat_type=? AND platform_type=? AND threat_entry_type=?
"""
params = [threat_list.threat_type, threat_list.platform_type,
threat_list.threat_entry_type]
with self.get_cursor() as dbc:
dbc.execute(q, params)
|
delete_threat_list
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, threat_list):\n log.info('Deleting cached threat list \"{}\"'.format(repr(threat_list)))\n q = \"\"\"DELETE FROM threat_list\n WHERE threat_type=? AND platform_type=? AND threat_entry_type=?\n \"\"\"\n params = [threat_list.threat_type, threat_list.platform_type,\n threat_list.threat_entry_type]\n with self.get_cursor() as dbc:\n dbc.execute(q, params)\n",
"role": "user"
},
{
"content": "delete_threat_list",
"role": "assistant"
}
] |
def x(self, threat_list):
q = """SELECT value FROM hash_prefix
WHERE threat_type=? AND platform_type=? AND threat_entry_type=?
ORDER BY value
"""
params = [threat_list.threat_type, threat_list.platform_type,
threat_list.threat_entry_type]
with self.get_cursor() as dbc:
dbc.execute(q, params)
all_hashes = b''.join(bytes(h[0]) for h in dbc.fetchall())
checksum = hashlib.sha256(all_hashes).digest()
return checksum
|
hash_prefix_list_checksum
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, threat_list):\n q = \"\"\"SELECT value FROM hash_prefix\n WHERE threat_type=? AND platform_type=? AND threat_entry_type=?\n ORDER BY value\n \"\"\"\n params = [threat_list.threat_type, threat_list.platform_type,\n threat_list.threat_entry_type]\n with self.get_cursor() as dbc:\n dbc.execute(q, params)\n all_hashes = b''.join(bytes(h[0]) for h in dbc.fetchall())\n checksum = hashlib.sha256(all_hashes).digest()\n return checksum\n",
"role": "user"
},
{
"content": "hash_prefix_list_checksum",
"role": "assistant"
}
] |
def x(self, threat_list, indices):
batch_size = 40
q = """DELETE FROM hash_prefix
WHERE threat_type=? AND platform_type=? AND threat_entry_type=? AND value IN ({})
"""
prefixes_to_remove = self.get_hash_prefix_values_to_remove(threat_list,
indices)
with self.get_cursor() as dbc:
for i in range(0, len(prefixes_to_remove), batch_size):
remove_batch = prefixes_to_remove[i:i + batch_size]
params = [threat_list.threat_type, threat_list.platform_type,
threat_list.threat_entry_type] + [sqlite3.Binary(b) for b in
remove_batch]
dbc.execute(q.format(','.join(['?'] * len(remove_batch))), params)
|
remove_hash_prefix_indices
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, threat_list, indices):\n batch_size = 40\n q = \"\"\"DELETE FROM hash_prefix\n WHERE threat_type=? AND platform_type=? AND threat_entry_type=? AND value IN ({})\n \"\"\"\n prefixes_to_remove = self.get_hash_prefix_values_to_remove(threat_list,\n indices)\n with self.get_cursor() as dbc:\n for i in range(0, len(prefixes_to_remove), batch_size):\n remove_batch = prefixes_to_remove[i:i + batch_size]\n params = [threat_list.threat_type, threat_list.platform_type,\n threat_list.threat_entry_type] + [sqlite3.Binary(b) for b in\n remove_batch]\n dbc.execute(q.format(','.join(['?'] * len(remove_batch))), params)\n",
"role": "user"
},
{
"content": "remove_hash_prefix_indices",
"role": "assistant"
}
] |
def x(self):
q = 'SELECT distinct value from hash_prefix'
output = []
with self.get_cursor() as dbc:
dbc.execute(q)
output = [bytes(r[0]) for r in dbc.fetchall()]
return output
|
dump_hash_prefix_values
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n q = 'SELECT distinct value from hash_prefix'\n output = []\n with self.get_cursor() as dbc:\n dbc.execute(q)\n output = [bytes(r[0]) for r in dbc.fetchall()]\n return output\n",
"role": "user"
},
{
"content": "dump_hash_prefix_values",
"role": "assistant"
}
] |
def x(tk):
used = False
try:
while True:
try:
method, args, kwargs, response_queue = (tk.tk._event_queue.
get_nowait())
except queue.Empty:
break
else:
used = True
if tk.tk._debug >= 2:
print('Calling event from main thread:', method.
__name__, args, kwargs)
try:
response_queue.put((False, method(*args, **kwargs)))
except SystemExit:
raise
except Exception:
from sys import exc_info
ex_type, ex_value, ex_tb = exc_info()
response_queue.put((True, (ex_type, ex_value, ex_tb)))
finally:
if used:
tk.after_idle(_check_events, tk)
else:
tk.after(tk.tk._check_period, _check_events, tk)
|
check_events
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(tk):\n used = False\n try:\n while True:\n try:\n method, args, kwargs, response_queue = (tk.tk._event_queue.\n get_nowait())\n except queue.Empty:\n break\n else:\n used = True\n if tk.tk._debug >= 2:\n print('Calling event from main thread:', method.\n __name__, args, kwargs)\n try:\n response_queue.put((False, method(*args, **kwargs)))\n except SystemExit:\n raise\n except Exception:\n from sys import exc_info\n ex_type, ex_value, ex_tb = exc_info()\n response_queue.put((True, (ex_type, ex_value, ex_tb)))\n finally:\n if used:\n tk.after_idle(_check_events, tk)\n else:\n tk.after(tk.tk._check_period, _check_events, tk)\n",
"role": "user"
},
{
"content": "check_events",
"role": "assistant"
}
] |
def x(self):
try:
self.storage.cleanup_full_hashes()
self.storage.commit()
self._sync_threat_lists()
self.storage.commit()
self._sync_hash_prefix_cache()
except Exception:
self.storage.rollback()
raise
|
update_hash_prefix_cache
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n try:\n self.storage.cleanup_full_hashes()\n self.storage.commit()\n self._sync_threat_lists()\n self.storage.commit()\n self._sync_hash_prefix_cache()\n except Exception:\n self.storage.rollback()\n raise\n",
"role": "user"
},
{
"content": "update_hash_prefix_cache",
"role": "assistant"
}
] |
def x(self, hash_prefixes):
client_state = self.storage.get_client_state()
self.api_client.fair_use_delay()
fh_response = self.api_client.get_full_hashes(hash_prefixes, client_state)
for m in fh_response.get('matches', []):
threat_list = ThreatList(m['threatType'], m['platformType'], m[
'threatEntryType'])
hash_value = b64decode(m['threat']['hash'])
cache_duration = int(m['cacheDuration'].rstrip('s'))
malware_threat_type = None
for metadata in m['threatEntryMetadata'].get('entries', []):
k = b64decode(metadata['key'])
v = b64decode(metadata['value'])
if k == 'malware_threat_type':
malware_threat_type = v
self.storage.store_full_hash(threat_list, hash_value,
cache_duration, malware_threat_type)
negative_cache_duration = int(fh_response['negativeCacheDuration'].
rstrip('s'))
for prefix_value in hash_prefixes:
self.storage.update_hash_prefix_expiration(prefix_value,
negative_cache_duration)
|
sync_full_hashes
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, hash_prefixes):\n client_state = self.storage.get_client_state()\n self.api_client.fair_use_delay()\n fh_response = self.api_client.get_full_hashes(hash_prefixes, client_state)\n for m in fh_response.get('matches', []):\n threat_list = ThreatList(m['threatType'], m['platformType'], m[\n 'threatEntryType'])\n hash_value = b64decode(m['threat']['hash'])\n cache_duration = int(m['cacheDuration'].rstrip('s'))\n malware_threat_type = None\n for metadata in m['threatEntryMetadata'].get('entries', []):\n k = b64decode(metadata['key'])\n v = b64decode(metadata['value'])\n if k == 'malware_threat_type':\n malware_threat_type = v\n self.storage.store_full_hash(threat_list, hash_value,\n cache_duration, malware_threat_type)\n negative_cache_duration = int(fh_response['negativeCacheDuration'].\n rstrip('s'))\n for prefix_value in hash_prefixes:\n self.storage.update_hash_prefix_expiration(prefix_value,\n negative_cache_duration)\n",
"role": "user"
},
{
"content": "sync_full_hashes",
"role": "assistant"
}
] |
def x(self, url):
if type(url) is not str:
url = url.encode('utf8')
if not url.strip():
raise ValueError('Empty input string.')
url_hashes = URL(url).hashes
try:
list_names = self._lookup_hashes(url_hashes)
self.storage.commit()
except Exception:
self.storage.rollback()
raise
if list_names:
return list_names
return None
|
lookup_url
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, url):\n if type(url) is not str:\n url = url.encode('utf8')\n if not url.strip():\n raise ValueError('Empty input string.')\n url_hashes = URL(url).hashes\n try:\n list_names = self._lookup_hashes(url_hashes)\n self.storage.commit()\n except Exception:\n self.storage.rollback()\n raise\n if list_names:\n return list_names\n return None\n",
"role": "user"
},
{
"content": "lookup_url",
"role": "assistant"
}
] |
def x(self, full_hashes):
full_hashes = list(full_hashes)
cues = [fh[0:4] for fh in full_hashes]
result = []
matching_prefixes = {}
matching_full_hashes = set()
is_potential_threat = False
for hash_prefix, negative_cache_expired in self.storage.lookup_hash_prefix(
cues):
for full_hash in full_hashes:
if full_hash.startswith(hash_prefix):
is_potential_threat = True
matching_prefixes[hash_prefix] = matching_prefixes.get(
hash_prefix, False) or negative_cache_expired
matching_full_hashes.add(full_hash)
if not is_potential_threat:
return []
matching_expired_threat_lists = set()
for threat_list, has_expired in self.storage.lookup_full_hashes(
matching_full_hashes):
if has_expired:
matching_expired_threat_lists.add(threat_list)
else:
result.append(threat_list)
if result:
return result
if len(matching_expired_threat_lists) == 0 and sum(map(int,
matching_prefixes.values())) == 0:
log.info('Negative cache hit.')
return []
self._sync_full_hashes(matching_prefixes.keys())
for threat_list, has_expired in self.storage.lookup_full_hashes(
matching_full_hashes):
if not has_expired:
result.append(threat_list)
return result
|
lookup_hashes
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, full_hashes):\n full_hashes = list(full_hashes)\n cues = [fh[0:4] for fh in full_hashes]\n result = []\n matching_prefixes = {}\n matching_full_hashes = set()\n is_potential_threat = False\n for hash_prefix, negative_cache_expired in self.storage.lookup_hash_prefix(\n cues):\n for full_hash in full_hashes:\n if full_hash.startswith(hash_prefix):\n is_potential_threat = True\n matching_prefixes[hash_prefix] = matching_prefixes.get(\n hash_prefix, False) or negative_cache_expired\n matching_full_hashes.add(full_hash)\n if not is_potential_threat:\n return []\n matching_expired_threat_lists = set()\n for threat_list, has_expired in self.storage.lookup_full_hashes(\n matching_full_hashes):\n if has_expired:\n matching_expired_threat_lists.add(threat_list)\n else:\n result.append(threat_list)\n if result:\n return result\n if len(matching_expired_threat_lists) == 0 and sum(map(int,\n matching_prefixes.values())) == 0:\n log.info('Negative cache hit.')\n return []\n self._sync_full_hashes(matching_prefixes.keys())\n for threat_list, has_expired in self.storage.lookup_full_hashes(\n matching_full_hashes):\n if not has_expired:\n result.append(threat_list)\n return result\n",
"role": "user"
},
{
"content": "lookup_hashes",
"role": "assistant"
}
] |
def x(self):
response = self.service.threatLists().list().execute()
self.set_wait_duration(response.get('minimumWaitDuration'))
return response['threatLists']
|
get_threats_lists
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n response = self.service.threatLists().list().execute()\n self.set_wait_duration(response.get('minimumWaitDuration'))\n return response['threatLists']\n",
"role": "user"
},
{
"content": "get_threats_lists",
"role": "assistant"
}
] |
def x(self, client_state):
request_body = {'client': {'clientId': self.client_id, 'clientVersion':
self.client_version}, 'listUpdateRequests': []}
for (threat_type, platform_type, threat_entry_type
), current_state in client_state.items():
request_body['listUpdateRequests'].append({'threatType':
threat_type, 'platformType': platform_type, 'threatEntryType':
threat_entry_type, 'state': current_state, 'constraints': {
'supportedCompressions': ['RAW']}})
response = self.service.threatListUpdates().fetch(body=request_body
).execute()
self.set_wait_duration(response.get('minimumWaitDuration'))
return response['listUpdateResponses']
|
get_threats_update
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, client_state):\n request_body = {'client': {'clientId': self.client_id, 'clientVersion':\n self.client_version}, 'listUpdateRequests': []}\n for (threat_type, platform_type, threat_entry_type\n ), current_state in client_state.items():\n request_body['listUpdateRequests'].append({'threatType':\n threat_type, 'platformType': platform_type, 'threatEntryType':\n threat_entry_type, 'state': current_state, 'constraints': {\n 'supportedCompressions': ['RAW']}})\n response = self.service.threatListUpdates().fetch(body=request_body\n ).execute()\n self.set_wait_duration(response.get('minimumWaitDuration'))\n return response['listUpdateResponses']\n",
"role": "user"
},
{
"content": "get_threats_update",
"role": "assistant"
}
] |
def x(self, prefixes, client_state):
request_body = {'client': {'clientId': self.client_id, 'clientVersion':
self.client_version}, 'clientStates': [], 'threatInfo': {
'threatTypes': [], 'platformTypes': [], 'threatEntryTypes': [],
'threatEntries': []}}
for prefix in prefixes:
request_body['threatInfo']['threatEntries'].append({'hash':
b64encode(prefix).decode()})
for (threatType, platformType, threatEntryType
), clientState in client_state.items():
request_body['clientStates'].append(clientState)
if threatType not in request_body['threatInfo']['threatTypes']:
request_body['threatInfo']['threatTypes'].append(threatType)
if platformType not in request_body['threatInfo']['platformTypes']:
request_body['threatInfo']['platformTypes'].append(platformType)
if threatEntryType not in request_body['threatInfo']['threatEntryTypes'
]:
request_body['threatInfo']['threatEntryTypes'].append(
threatEntryType)
response = self.service.fullHashes().find(body=request_body).execute()
self.set_wait_duration(response.get('minimumWaitDuration'))
return response
|
get_full_hashes
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, prefixes, client_state):\n request_body = {'client': {'clientId': self.client_id, 'clientVersion':\n self.client_version}, 'clientStates': [], 'threatInfo': {\n 'threatTypes': [], 'platformTypes': [], 'threatEntryTypes': [],\n 'threatEntries': []}}\n for prefix in prefixes:\n request_body['threatInfo']['threatEntries'].append({'hash':\n b64encode(prefix).decode()})\n for (threatType, platformType, threatEntryType\n ), clientState in client_state.items():\n request_body['clientStates'].append(clientState)\n if threatType not in request_body['threatInfo']['threatTypes']:\n request_body['threatInfo']['threatTypes'].append(threatType)\n if platformType not in request_body['threatInfo']['platformTypes']:\n request_body['threatInfo']['platformTypes'].append(platformType)\n if threatEntryType not in request_body['threatInfo']['threatEntryTypes'\n ]:\n request_body['threatInfo']['threatEntryTypes'].append(\n threatEntryType)\n response = self.service.fullHashes().find(body=request_body).execute()\n self.set_wait_duration(response.get('minimumWaitDuration'))\n return response\n",
"role": "user"
},
{
"content": "get_full_hashes",
"role": "assistant"
}
] |
def x(self):
for url_variant in self.url_permutations(self.canonical):
url_hash = self.digest(url_variant)
yield url_hash
|
hashes
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n for url_variant in self.url_permutations(self.canonical):\n url_hash = self.digest(url_variant)\n yield url_hash\n",
"role": "user"
},
{
"content": "hashes",
"role": "assistant"
}
] |
def x(self):
def x(u):
uu = urllib.unquote(u)
if uu == u:
return uu
else:
return full_unescape(uu)
def x(u):
uu = urlparse.unquote_to_bytes(u)
if uu == u:
return uu
else:
return full_unescape_to_bytes(uu)
def x(s):
safe_chars = '!"$&\'()*+,-./:;<=>?@[\\]^_`{|}~'
return urllib.quote(s, safe=safe_chars)
url = self.url.strip()
url = url.replace(b'\n', b'').replace(b'\r', b'').replace(b'\t', b'')
url = url.split(b'#', 1)[0]
if url.startswith(b'//'):
url = b'http:' + url
if len(url.split(b'://')) <= 1:
url = b'http://' + url
if self.__py3 and type(url) is bytes:
url = quote(full_unescape_to_bytes(url))
else:
url = quote(full_unescape(url))
url_parts = urlparse.urlsplit(url)
if not url_parts[0]:
url = 'http://{}'.format(url)
url_parts = urlparse.urlsplit(url)
protocol = url_parts.scheme
if self.__py3:
host = full_unescape_to_bytes(url_parts.hostname)
path = full_unescape_to_bytes(url_parts.path)
else:
host = full_unescape(url_parts.hostname)
path = full_unescape(url_parts.path)
query = url_parts.query
if not query and '?' not in url:
query = None
if not path:
path = b'/'
has_trailing_slash = path[-1:] == b'/'
path = posixpath.normpath(path).replace(b'//', b'/')
if has_trailing_slash and path[-1:] != b'/':
path = path + b'/'
port = url_parts.port
host = host.strip(b'.')
host = re.sub(b'\\.+', b'.', host).lower()
if host.isdigit():
try:
host = socket.inet_ntoa(struct.pack('!I', int(host)))
except Exception:
pass
elif host.startswith(b'0x') and b'.' not in host:
try:
host = socket.inet_ntoa(struct.pack('!I', int(host, 16)))
except Exception:
pass
quoted_path = quote(path)
quoted_host = quote(host)
if port is not None:
quoted_host = '{}:{}'.format(quoted_host, port)
canonical_url = '{}://{}{}'.format(protocol, quoted_host, quoted_path)
if query is not None:
canonical_url = '{}?{}'.format(canonical_url, query)
return canonical_url
|
canonical
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n\n def x(u):\n uu = urllib.unquote(u)\n if uu == u:\n return uu\n else:\n return full_unescape(uu)\n\n def x(u):\n uu = urlparse.unquote_to_bytes(u)\n if uu == u:\n return uu\n else:\n return full_unescape_to_bytes(uu)\n\n def x(s):\n safe_chars = '!\"$&\\'()*+,-./:;<=>?@[\\\\]^_`{|}~'\n return urllib.quote(s, safe=safe_chars)\n url = self.url.strip()\n url = url.replace(b'\\n', b'').replace(b'\\r', b'').replace(b'\\t', b'')\n url = url.split(b'#', 1)[0]\n if url.startswith(b'//'):\n url = b'http:' + url\n if len(url.split(b'://')) <= 1:\n url = b'http://' + url\n if self.__py3 and type(url) is bytes:\n url = quote(full_unescape_to_bytes(url))\n else:\n url = quote(full_unescape(url))\n url_parts = urlparse.urlsplit(url)\n if not url_parts[0]:\n url = 'http://{}'.format(url)\n url_parts = urlparse.urlsplit(url)\n protocol = url_parts.scheme\n if self.__py3:\n host = full_unescape_to_bytes(url_parts.hostname)\n path = full_unescape_to_bytes(url_parts.path)\n else:\n host = full_unescape(url_parts.hostname)\n path = full_unescape(url_parts.path)\n query = url_parts.query\n if not query and '?' not in url:\n query = None\n if not path:\n path = b'/'\n has_trailing_slash = path[-1:] == b'/'\n path = posixpath.normpath(path).replace(b'//', b'/')\n if has_trailing_slash and path[-1:] != b'/':\n path = path + b'/'\n port = url_parts.port\n host = host.strip(b'.')\n host = re.sub(b'\\\\.+', b'.', host).lower()\n if host.isdigit():\n try:\n host = socket.inet_ntoa(struct.pack('!I', int(host)))\n except Exception:\n pass\n elif host.startswith(b'0x') and b'.' not in host:\n try:\n host = socket.inet_ntoa(struct.pack('!I', int(host, 16)))\n except Exception:\n pass\n quoted_path = quote(path)\n quoted_host = quote(host)\n if port is not None:\n quoted_host = '{}:{}'.format(quoted_host, port)\n canonical_url = '{}://{}{}'.format(protocol, quoted_host, quoted_path)\n if query is not None:\n canonical_url = '{}?{}'.format(canonical_url, query)\n return canonical_url\n",
"role": "user"
},
{
"content": "canonical",
"role": "assistant"
}
] |
def x(url):
def x(host):
if re.match('\\d+\\.\\d+\\.\\d+\\.\\d+', host):
yield host
return
parts = host.split('.')
l = min(len(parts), 5)
if l > 4:
yield host
for i in range(l - 1):
yield '.'.join(parts[i - l:])
def x(path):
yield path
query = None
if '?' in path:
path, query = path.split('?', 1)
if query is not None:
yield path
path_parts = path.split('/')[0:-1]
curr_path = ''
for i in range(min(4, len(path_parts))):
curr_path = curr_path + path_parts[i] + '/'
yield curr_path
protocol, address_str = urllib.splittype(url)
host, path = urllib.splithost(address_str)
user, host = urllib.splituser(host)
host, port = urllib.splitport(host)
host = host.strip('/')
seen_permutations = set()
for h in url_host_permutations(host):
for p in url_path_permutations(path):
u = '{}{}'.format(h, p)
if u not in seen_permutations:
yield u
seen_permutations.add(u)
|
url_permutations
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(url):\n\n def x(host):\n if re.match('\\\\d+\\\\.\\\\d+\\\\.\\\\d+\\\\.\\\\d+', host):\n yield host\n return\n parts = host.split('.')\n l = min(len(parts), 5)\n if l > 4:\n yield host\n for i in range(l - 1):\n yield '.'.join(parts[i - l:])\n\n def x(path):\n yield path\n query = None\n if '?' in path:\n path, query = path.split('?', 1)\n if query is not None:\n yield path\n path_parts = path.split('/')[0:-1]\n curr_path = ''\n for i in range(min(4, len(path_parts))):\n curr_path = curr_path + path_parts[i] + '/'\n yield curr_path\n protocol, address_str = urllib.splittype(url)\n host, path = urllib.splithost(address_str)\n user, host = urllib.splituser(host)\n host, port = urllib.splitport(host)\n host = host.strip('/')\n seen_permutations = set()\n for h in url_host_permutations(host):\n for p in url_path_permutations(path):\n u = '{}{}'.format(h, p)\n if u not in seen_permutations:\n yield u\n seen_permutations.add(u)\n",
"role": "user"
},
{
"content": "url_permutations",
"role": "assistant"
}
] |
def x(v1, v2):
def x(v):
v = re.sub('(\\.0+)*$', '', v)
result = []
for part in v.split('.'):
m = re.match('^(\\d+)$', part)
if m:
result.append(int(m.group(1)))
continue
m = re.match('^(\\d+)([a-zA-Z]+)$', part)
if m:
result.append(int(m.group(1)))
result.append(m.group(2))
continue
m = re.match('^(\\d+)([a-zA-Z]+)(\\d+)$', part)
if m:
result.append(int(m.group(1)))
result.append(m.group(2))
result.append(int(m.group(3)))
continue
return tuple(result)
n1 = normalize(v1)
n2 = normalize(v2)
return (n1 > n2) - (n1 < n2)
|
compare_versions
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(v1, v2):\n\n def x(v):\n v = re.sub('(\\\\.0+)*$', '', v)\n result = []\n for part in v.split('.'):\n m = re.match('^(\\\\d+)$', part)\n if m:\n result.append(int(m.group(1)))\n continue\n m = re.match('^(\\\\d+)([a-zA-Z]+)$', part)\n if m:\n result.append(int(m.group(1)))\n result.append(m.group(2))\n continue\n m = re.match('^(\\\\d+)([a-zA-Z]+)(\\\\d+)$', part)\n if m:\n result.append(int(m.group(1)))\n result.append(m.group(2))\n result.append(int(m.group(3)))\n continue\n return tuple(result)\n n1 = normalize(v1)\n n2 = normalize(v2)\n return (n1 > n2) - (n1 < n2)\n",
"role": "user"
},
{
"content": "compare_versions",
"role": "assistant"
}
] |
def x(spec):
m = re.search('([<>=]?=?)?\\s*([0-9.a-zA-Z]+)', spec)
return m.group(2), m.group(1)
|
split_version_specifier
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(spec):\n m = re.search('([<>=]?=?)?\\\\s*([0-9.a-zA-Z]+)', spec)\n return m.group(2), m.group(1)\n",
"role": "user"
},
{
"content": "split_version_specifier",
"role": "assistant"
}
] |
def x(package):
pkg_config_exe = os.environ.get('PKG_CONFIG', None) or 'pkg-config'
cmd = '{0} --exists {1}'.format(pkg_config_exe, package).split()
return call(cmd) == 0
|
exists
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(package):\n pkg_config_exe = os.environ.get('PKG_CONFIG', None) or 'pkg-config'\n cmd = '{0} --exists {1}'.format(pkg_config_exe, package).split()\n return call(cmd) == 0\n",
"role": "user"
},
{
"content": "exists",
"role": "assistant"
}
] |
def x(package, static=False):
_raise_if_not_exists(package)
return _query(package, *_build_options('--libs', static=static))
|
libs
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(package, static=False):\n _raise_if_not_exists(package)\n return _query(package, *_build_options('--libs', static=static))\n",
"role": "user"
},
{
"content": "libs",
"role": "assistant"
}
] |
def x(package):
_raise_if_not_exists(package)
result = _query(package, '--print-variables')
names = (x.strip() for x in result.split('\n') if x != '')
return dict((x, _query(package, '--variable={0}'.format(x)).strip()) for
x in names)
|
variables
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(package):\n _raise_if_not_exists(package)\n result = _query(package, '--print-variables')\n names = (x.strip() for x in result.split('\\n') if x != '')\n return dict((x, _query(package, '--variable={0}'.format(x)).strip()) for\n x in names)\n",
"role": "user"
},
{
"content": "variables",
"role": "assistant"
}
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.