source_code
stringlengths
45
7.24k
function_name
stringlengths
1
111
messages
list
def x(list_idl): seqs = set() for idl in list_idl.loci2seq: [seqs.add(s) for s in list_idl.loci2seq[idl]] return seqs
get_seqs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(list_idl):\n seqs = set()\n for idl in list_idl.loci2seq:\n [seqs.add(s) for s in list_idl.loci2seq[idl]]\n return seqs\n", "role": "user" }, { "content": "get_seqs", "role": "assistant" } ]
def x(s1, s2, i1, i2): c = len(set(s1).intersection(s2)) t = min(len(s1), len(s2)) pct = 1.0 * c / t * t is_gt = up_threshold(pct, t * 1.0, parameters.similar) logger.debug('_common: pct %s of clusters:%s %s = %s' % (1.0 * c / t, i1, i2, is_gt)) if pct < parameters.similar and is_gt and pct > 0: pct = parameters.similar return pct / t
common
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(s1, s2, i1, i2):\n c = len(set(s1).intersection(s2))\n t = min(len(s1), len(s2))\n pct = 1.0 * c / t * t\n is_gt = up_threshold(pct, t * 1.0, parameters.similar)\n logger.debug('_common: pct %s of clusters:%s %s = %s' % (1.0 * c / t,\n i1, i2, is_gt))\n if pct < parameters.similar and is_gt and pct > 0:\n pct = parameters.similar\n return pct / t\n", "role": "user" }, { "content": "common", "role": "assistant" } ]
def x(pairs, common, clus_seen, loci_similarity): all_true1 = all([all([(common and loci_similarity[p, c] > parameters. similar) for p in pairs if (p, c) in loci_similarity]) for c in clus_seen]) all_true2 = all([all([(common and loci_similarity[c, p] > parameters. similar) for p in pairs if (c, p) in loci_similarity]) for c in clus_seen]) return all_true1 * all_true2
is_consistent
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(pairs, common, clus_seen, loci_similarity):\n all_true1 = all([all([(common and loci_similarity[p, c] > parameters.\n similar) for p in pairs if (p, c) in loci_similarity]) for c in\n clus_seen])\n all_true2 = all([all([(common and loci_similarity[c, p] > parameters.\n similar) for p in pairs if (c, p) in loci_similarity]) for c in\n clus_seen])\n return all_true1 * all_true2\n", "role": "user" }, { "content": "is_consistent", "role": "assistant" } ]
def x(loci, loci_similarity): n_cluster = 0 internal_cluster = {} clus_seen = {} loci_sorted = sorted(loci_similarity.iteritems(), key=operator. itemgetter(1), reverse=True) for pairs, sim in loci_sorted: common = sim > parameters.similar n_cluster += 1 logger.debug('_merge_similar:try new cluster %s' % n_cluster) new_c = cluster(n_cluster) p_seen, p_unseen = [], [] size = min(len(_get_seqs(loci[pairs[0]])), len(_get_seqs(loci[pairs [1]]))) if common: consistent = _is_consistent(pairs, common, clus_seen, loci_similarity) logger.debug('_merge_similar: clusters seen: %s' % clus_seen) logger.debug( '_merge_similar: id %s common %s|%s total %s consistent %s' % (pairs, sim, common, size, consistent)) if not consistent: continue if pairs[0] in clus_seen: p_seen.append(pairs[0]) p_unseen.append(pairs[1]) if pairs[1] in clus_seen: p_seen.append(pairs[1]) p_unseen.append(pairs[0]) if len(p_seen) == 0: new_c = _merge_cluster(loci[pairs[0]], new_c) new_c = _merge_cluster(loci[pairs[1]], new_c) [clus_seen.update({p: n_cluster}) for p in pairs] internal_cluster[n_cluster] = new_c if len(p_seen) == 1: idc_seen = clus_seen[p_seen[0]] internal_cluster[idc_seen] = _merge_cluster(loci[p_unseen[0 ]], internal_cluster[idc_seen]) clus_seen[p_unseen[0]] = idc_seen else: logger.debug('_merge_similar: id %s %s are different' % pairs) continue internal_cluster.update(_add_unseen(loci, clus_seen, n_cluster)) logger.debug('_merge_similar: total clus %s' % len(internal_cluster.keys()) ) return internal_cluster
merge_similar
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(loci, loci_similarity):\n n_cluster = 0\n internal_cluster = {}\n clus_seen = {}\n loci_sorted = sorted(loci_similarity.iteritems(), key=operator.\n itemgetter(1), reverse=True)\n for pairs, sim in loci_sorted:\n common = sim > parameters.similar\n n_cluster += 1\n logger.debug('_merge_similar:try new cluster %s' % n_cluster)\n new_c = cluster(n_cluster)\n p_seen, p_unseen = [], []\n size = min(len(_get_seqs(loci[pairs[0]])), len(_get_seqs(loci[pairs\n [1]])))\n if common:\n consistent = _is_consistent(pairs, common, clus_seen,\n loci_similarity)\n logger.debug('_merge_similar: clusters seen: %s' % clus_seen)\n logger.debug(\n '_merge_similar: id %s common %s|%s total %s consistent %s' %\n (pairs, sim, common, size, consistent))\n if not consistent:\n continue\n if pairs[0] in clus_seen:\n p_seen.append(pairs[0])\n p_unseen.append(pairs[1])\n if pairs[1] in clus_seen:\n p_seen.append(pairs[1])\n p_unseen.append(pairs[0])\n if len(p_seen) == 0:\n new_c = _merge_cluster(loci[pairs[0]], new_c)\n new_c = _merge_cluster(loci[pairs[1]], new_c)\n [clus_seen.update({p: n_cluster}) for p in pairs]\n internal_cluster[n_cluster] = new_c\n if len(p_seen) == 1:\n idc_seen = clus_seen[p_seen[0]]\n internal_cluster[idc_seen] = _merge_cluster(loci[p_unseen[0\n ]], internal_cluster[idc_seen])\n clus_seen[p_unseen[0]] = idc_seen\n else:\n logger.debug('_merge_similar: id %s %s are different' % pairs)\n continue\n internal_cluster.update(_add_unseen(loci, clus_seen, n_cluster))\n logger.debug('_merge_similar: total clus %s' % len(internal_cluster.keys())\n )\n return internal_cluster\n", "role": "user" }, { "content": "merge_similar", "role": "assistant" } ]
def x(old, new): logger.debug('_merge_cluster: %s to %s' % (old.id, new.id)) logger.debug('_merge_cluster: add idls %s' % old.loci2seq.keys()) for idl in old.loci2seq: new.add_id_member(old.loci2seq[idl], idl) return new
merge_cluster
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(old, new):\n logger.debug('_merge_cluster: %s to %s' % (old.id, new.id))\n logger.debug('_merge_cluster: add idls %s' % old.loci2seq.keys())\n for idl in old.loci2seq:\n new.add_id_member(old.loci2seq[idl], idl)\n return new\n", "role": "user" }, { "content": "merge_cluster", "role": "assistant" } ]
def x(list_c, s2p, n_cluster): logger.debug('_solve_conflict: count once') if parameters.decision_cluster == 'bayes': return decide_by_bayes(list_c, s2p) loci_similarity = _calculate_similarity(list_c) loci_similarity = sorted(loci_similarity.iteritems(), key=operator. itemgetter(1), reverse=True) common = sum([score for p, score in loci_similarity]) while common > 0: n_cluster += 1 logger.debug('_solve_conflict: ma %s' % loci_similarity) pairs = loci_similarity[0][0] score = loci_similarity[0][1] logger.debug('_solve_conflict: common %s, new %s' % (score, n_cluster)) if parameters.decision_cluster.startswith('most-voted'): list_c = _split_cluster_by_most_vote(list_c, pairs) else: list_c = _split_cluster(list_c, pairs, n_cluster) list_c = {k: v for k, v in list_c.iteritems() if len(v.loci2seq) > 0} loci_similarity = _calculate_similarity(list_c) loci_similarity = sorted(loci_similarity.iteritems(), key=operator. itemgetter(1), reverse=True) common = sum([score for p, score in loci_similarity]) logger.debug('_solve_conflict: solved clusters %s' % len(list_c.keys()) ) return list_c
solve_conflict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(list_c, s2p, n_cluster):\n logger.debug('_solve_conflict: count once')\n if parameters.decision_cluster == 'bayes':\n return decide_by_bayes(list_c, s2p)\n loci_similarity = _calculate_similarity(list_c)\n loci_similarity = sorted(loci_similarity.iteritems(), key=operator.\n itemgetter(1), reverse=True)\n common = sum([score for p, score in loci_similarity])\n while common > 0:\n n_cluster += 1\n logger.debug('_solve_conflict: ma %s' % loci_similarity)\n pairs = loci_similarity[0][0]\n score = loci_similarity[0][1]\n logger.debug('_solve_conflict: common %s, new %s' % (score, n_cluster))\n if parameters.decision_cluster.startswith('most-voted'):\n list_c = _split_cluster_by_most_vote(list_c, pairs)\n else:\n list_c = _split_cluster(list_c, pairs, n_cluster)\n list_c = {k: v for k, v in list_c.iteritems() if len(v.loci2seq) > 0}\n loci_similarity = _calculate_similarity(list_c)\n loci_similarity = sorted(loci_similarity.iteritems(), key=operator.\n itemgetter(1), reverse=True)\n common = sum([score for p, score in loci_similarity])\n logger.debug('_solve_conflict: solved clusters %s' % len(list_c.keys())\n )\n return list_c\n", "role": "user" }, { "content": "solve_conflict", "role": "assistant" } ]
def x(c, pairs, n): old = c[p[0]] new = c[p[1]] new_c = cluster(n) common = set(_get_seqs(old)).intersection(_get_seqs(new)) for idl in old.loci2seq: in_common = list(set(common).intersection(old.loci2seq[idl])) if len(in_common) > 0: logger.debug('_split_cluster: in_common %s with pair 1' % len( in_common)) new_c.add_id_member(in_common, idl) old.loci2seq[idl] = list(set(old.loci2seq[idl]) - set(common)) logger.debug('_split_cluster: len old %s with pair 1' % len(old .loci2seq)) for idl in new.loci2seq: in_common = list(set(common).intersection(new.loci2seq[idl])) if len(in_common) > 0: logger.debug('_split_cluster: in_common %s with pair 2' % len( in_common)) new_c.add_id_member(in_common, idl) new.loci2seq[idl] = list(set(new.loci2seq[idl]) - set(common)) logger.debug('_split_cluster: len old %s with pair 2' % len(new .loci2seq)) old.update() new.update() old.loci2seq = {k: v for k, v in old.loci2seq.iteritems() if len(v) > 0} new.loci2seq = {k: v for k, v in new.loci2seq.iteritems() if len(v) > 0} c[n] = new c[p[0]] = old c[p[1]] = new return c
split_cluster
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(c, pairs, n):\n old = c[p[0]]\n new = c[p[1]]\n new_c = cluster(n)\n common = set(_get_seqs(old)).intersection(_get_seqs(new))\n for idl in old.loci2seq:\n in_common = list(set(common).intersection(old.loci2seq[idl]))\n if len(in_common) > 0:\n logger.debug('_split_cluster: in_common %s with pair 1' % len(\n in_common))\n new_c.add_id_member(in_common, idl)\n old.loci2seq[idl] = list(set(old.loci2seq[idl]) - set(common))\n logger.debug('_split_cluster: len old %s with pair 1' % len(old\n .loci2seq))\n for idl in new.loci2seq:\n in_common = list(set(common).intersection(new.loci2seq[idl]))\n if len(in_common) > 0:\n logger.debug('_split_cluster: in_common %s with pair 2' % len(\n in_common))\n new_c.add_id_member(in_common, idl)\n new.loci2seq[idl] = list(set(new.loci2seq[idl]) - set(common))\n logger.debug('_split_cluster: len old %s with pair 2' % len(new\n .loci2seq))\n old.update()\n new.update()\n old.loci2seq = {k: v for k, v in old.loci2seq.iteritems() if len(v) > 0}\n new.loci2seq = {k: v for k, v in new.loci2seq.iteritems() if len(v) > 0}\n c[n] = new\n c[p[0]] = old\n c[p[1]] = new\n return c\n", "role": "user" }, { "content": "split_cluster", "role": "assistant" } ]
def x(c, p): old, new = c[p[0]], c[p[1]] old_size = _get_seqs(old) new_size = _get_seqs(new) logger.debug('_most_vote: size of %s with %s - %s with %s' % (old.id, len(old_size), new.id, len(new_size))) if len(old_size) > len(new_size): keep, remove = old, new else: keep, remove = new, old common = list(set(old_size).intersection(new_size)) logger.debug('_most_vote: keep %s remove %s with common %s' % (keep.id, remove.id, len(common))) for idl in remove.loci2seq: if len(common) > 0: remove.loci2seq[idl] = list(set(remove.loci2seq[idl]) - set(common) ) keep.loci2seq = {k: v for k, v in keep.loci2seq.iteritems() if len(v) > 0} remove.loci2seq = {k: v for k, v in remove.loci2seq.iteritems() if len( v) > 0} keep.update() remove.update() c[keep.id] = keep c[remove.id] = remove return c
split_cluster_by_most_vote
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(c, p):\n old, new = c[p[0]], c[p[1]]\n old_size = _get_seqs(old)\n new_size = _get_seqs(new)\n logger.debug('_most_vote: size of %s with %s - %s with %s' % (old.id,\n len(old_size), new.id, len(new_size)))\n if len(old_size) > len(new_size):\n keep, remove = old, new\n else:\n keep, remove = new, old\n common = list(set(old_size).intersection(new_size))\n logger.debug('_most_vote: keep %s remove %s with common %s' % (keep.id,\n remove.id, len(common)))\n for idl in remove.loci2seq:\n if len(common) > 0:\n remove.loci2seq[idl] = list(set(remove.loci2seq[idl]) - set(common)\n )\n keep.loci2seq = {k: v for k, v in keep.loci2seq.iteritems() if len(v) > 0}\n remove.loci2seq = {k: v for k, v in remove.loci2seq.iteritems() if len(\n v) > 0}\n keep.update()\n remove.update()\n c[keep.id] = keep\n c[remove.id] = remove\n return c\n", "role": "user" }, { "content": "split_cluster_by_most_vote", "role": "assistant" } ]
def x(list_c): global REMOVED init = len(list_c) list_c = {k: v for k, v in list_c.iteritems() if len(_get_seqs(v)) > parameters.min_seqs} logger.debug('_clean_cluster: number of clusters %s ' % len(list_c.keys())) list_c = {k: _select_loci(v) for k, v in list_c.iteritems()} end = len(list_c) REMOVED += init - end return list_c
clean_cluster
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(list_c):\n global REMOVED\n init = len(list_c)\n list_c = {k: v for k, v in list_c.iteritems() if len(_get_seqs(v)) >\n parameters.min_seqs}\n logger.debug('_clean_cluster: number of clusters %s ' % len(list_c.keys()))\n list_c = {k: _select_loci(v) for k, v in list_c.iteritems()}\n end = len(list_c)\n REMOVED += init - end\n return list_c\n", "role": "user" }, { "content": "clean_cluster", "role": "assistant" } ]
def x(c): loci_len = {k: len(v) for k, v in c.loci2seq.iteritems()} logger.debug('_select_loci: number of loci %s' % len(c.loci2seq.keys())) loci_len_sort = sorted(loci_len.iteritems(), key=operator.itemgetter(1), reverse=True) max_size = loci_len_sort[0][1] logger.debug('_select_loci: max size %s' % max_size) loci_clean = {locus: c.loci2seq[locus] for locus, size in loci_len_sort if size > 0.8 * max_size} c.loci2seq = loci_clean removed = list(set(c.idmembers.keys()) - set(_get_seqs(c))) c.add_id_member(removed, loci_len_sort[0][0]) logger.debug('_select_loci: number of loci %s after cleaning' % len(c. loci2seq.keys())) return c
select_loci
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(c):\n loci_len = {k: len(v) for k, v in c.loci2seq.iteritems()}\n logger.debug('_select_loci: number of loci %s' % len(c.loci2seq.keys()))\n loci_len_sort = sorted(loci_len.iteritems(), key=operator.itemgetter(1),\n reverse=True)\n max_size = loci_len_sort[0][1]\n logger.debug('_select_loci: max size %s' % max_size)\n loci_clean = {locus: c.loci2seq[locus] for locus, size in loci_len_sort if\n size > 0.8 * max_size}\n c.loci2seq = loci_clean\n removed = list(set(c.idmembers.keys()) - set(_get_seqs(c)))\n c.add_id_member(removed, loci_len_sort[0][0])\n logger.debug('_select_loci: number of loci %s after cleaning' % len(c.\n loci2seq.keys()))\n return c\n", "role": "user" }, { "content": "select_loci", "role": "assistant" } ]
def x(c, locilen_sorted, seen_seqs, filtered, maxseq, n_cluster): first_run = 0 seen_seqs = list() n_cluster += 1 logger.debug('_solve_loci:new cluster %s' % n_cluster) new_c = cluster(n_cluster) for idl, lenl in locilen_sorted: locus_seqs = c.loci2seq[idl] if first_run == 0: seen_seqs = locus_seqs first_run = 1 first_idl = idl intersect = list(set(seen_seqs).intersection(locus_seqs)) common = 0 if intersect: common = len(intersect) * 1.0 / min(len(seen_seqs), len(locus_seqs) ) logger.debug( '_sole_loci:id %s idl %s len %s max %s seen %s inter %s common %s ' % (c.id, idl, lenl, maxseq, len(seen_seqs), len(intersect), common)) if common * 1.0 >= 0.6: if lenl * 1.0 >= 0.6 * maxseq: c, new_c, seen_seqs = _merge_loci_in_cluster(c, new_c, idl, seen_seqs) else: c, new_c, seen_seqs = _merge_with_first_loci(c, new_c, first_idl, idl, seen_seqs) else: c = _remove_seqs_from_loci(c, idl, seen_seqs) filtered[n_cluster] = new_c return c, seen_seqs, filtered, n_cluster
solve_loci_deprecated
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(c, locilen_sorted, seen_seqs, filtered, maxseq, n_cluster):\n first_run = 0\n seen_seqs = list()\n n_cluster += 1\n logger.debug('_solve_loci:new cluster %s' % n_cluster)\n new_c = cluster(n_cluster)\n for idl, lenl in locilen_sorted:\n locus_seqs = c.loci2seq[idl]\n if first_run == 0:\n seen_seqs = locus_seqs\n first_run = 1\n first_idl = idl\n intersect = list(set(seen_seqs).intersection(locus_seqs))\n common = 0\n if intersect:\n common = len(intersect) * 1.0 / min(len(seen_seqs), len(locus_seqs)\n )\n logger.debug(\n '_sole_loci:id %s idl %s len %s max %s seen %s inter %s common %s '\n % (c.id, idl, lenl, maxseq, len(seen_seqs), len(intersect),\n common))\n if common * 1.0 >= 0.6:\n if lenl * 1.0 >= 0.6 * maxseq:\n c, new_c, seen_seqs = _merge_loci_in_cluster(c, new_c, idl,\n seen_seqs)\n else:\n c, new_c, seen_seqs = _merge_with_first_loci(c, new_c,\n first_idl, idl, seen_seqs)\n else:\n c = _remove_seqs_from_loci(c, idl, seen_seqs)\n filtered[n_cluster] = new_c\n return c, seen_seqs, filtered, n_cluster\n", "role": "user" }, { "content": "solve_loci_deprecated", "role": "assistant" } ]
def x(string): ann = set() if not string: return 'This cluster is inter-genic.' for item in string: for db in item: ann = ann.union(set(item[db])) return 'annotated as: %s ...' % ','.join(list(ann)[:3])
get_description
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(string):\n ann = set()\n if not string:\n return 'This cluster is inter-genic.'\n for item in string:\n for db in item:\n ann = ann.union(set(item[db]))\n return 'annotated as: %s ...' % ','.join(list(ann)[:3])\n", "role": "user" }, { "content": "get_description", "role": "assistant" } ]
def x(profile): x = set() for sample in profile: x = x.union(set(profile[sample].keys())) if not x: return '' end, start = max(x), min(x) x = range(start, end, 4) scaled_profile = defaultdict(list) for pos in x: for sample in profile: y = _get_closer(profile[sample], pos) if y: scaled_profile[sample].append(profile[sample][y]) else: scaled_profile[sample].append(0) return {'x': list(x), 'y': scaled_profile, 'names': scaled_profile.keys()}
set_format
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(profile):\n x = set()\n for sample in profile:\n x = x.union(set(profile[sample].keys()))\n if not x:\n return ''\n end, start = max(x), min(x)\n x = range(start, end, 4)\n scaled_profile = defaultdict(list)\n for pos in x:\n for sample in profile:\n y = _get_closer(profile[sample], pos)\n if y:\n scaled_profile[sample].append(profile[sample][y])\n else:\n scaled_profile[sample].append(0)\n return {'x': list(x), 'y': scaled_profile, 'names': scaled_profile.keys()}\n", "role": "user" }, { "content": "set_format", "role": "assistant" } ]
def x(con, data): with con: cur = con.cursor() cur.execute('DROP TABLE IF EXISTS clusters;') cur.execute( 'CREATE TABLE clusters(Id INT, Description TEXT, Locus TEXT, Annotation TEXT, Sequences TEXT, Profile TXT, Precursor TXT)' ) for c in data[0]: locus = json.dumps(data[0][c]['loci']) annotation = json.dumps(data[0][c]['ann']) description = _get_description(data[0][c]['ann']) sequences = json.dumps(_get_sequences(data[0][c])) keys = data[0][c]['freq'][0].values()[0].keys() profile = 'Not available.' if 'profile' in data[0][c]: profile = json.dumps(_set_format(data[0][c]['profile'])) precursor = json.dumps(data[0][c].get('precursor')) cur.execute( "INSERT INTO clusters VALUES(%s, '%s', '%s', '%s', '%s', '%s', '%s')" % (c, description, locus, annotation, sequences, profile, precursor))
insert_data
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(con, data):\n with con:\n cur = con.cursor()\n cur.execute('DROP TABLE IF EXISTS clusters;')\n cur.execute(\n 'CREATE TABLE clusters(Id INT, Description TEXT, Locus TEXT, Annotation TEXT, Sequences TEXT, Profile TXT, Precursor TXT)'\n )\n for c in data[0]:\n locus = json.dumps(data[0][c]['loci'])\n annotation = json.dumps(data[0][c]['ann'])\n description = _get_description(data[0][c]['ann'])\n sequences = json.dumps(_get_sequences(data[0][c]))\n keys = data[0][c]['freq'][0].values()[0].keys()\n profile = 'Not available.'\n if 'profile' in data[0][c]:\n profile = json.dumps(_set_format(data[0][c]['profile']))\n precursor = json.dumps(data[0][c].get('precursor'))\n cur.execute(\n \"INSERT INTO clusters VALUES(%s, '%s', '%s', '%s', '%s', '%s', '%s')\"\n % (c, description, locus, annotation, sequences, profile,\n precursor))\n", "role": "user" }, { "content": "insert_data", "role": "assistant" } ]
def x(file_in): loc_id = 1 bedfile_clusters = '' bamfile = pybedtools.BedTool(file_in) bed = pybedtools.BedTool.bam_to_bed(bamfile) for c, start, end, name, q, strand in bed: loc_id += 1 bedfile_clusters += '%s\t%s\t%s\t%s\t%s\t%s\n' % (c, start, end, name, loc_id, strand) return bedfile_clusters
parse_align_file
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(file_in):\n loc_id = 1\n bedfile_clusters = ''\n bamfile = pybedtools.BedTool(file_in)\n bed = pybedtools.BedTool.bam_to_bed(bamfile)\n for c, start, end, name, q, strand in bed:\n loc_id += 1\n bedfile_clusters += '%s\\t%s\\t%s\\t%s\\t%s\\t%s\\n' % (c, start, end,\n name, loc_id, strand)\n return bedfile_clusters\n", "role": "user" }, { "content": "parse_align_file", "role": "assistant" } ]
def x(seq_obj, in_file): name = '' index = 1 total = defaultdict(int) with open(in_file) as handle_in: line = handle_in.readline().strip() cols = line.split('\t') samples = cols[2:] for line in handle_in: line = line.strip() cols = line.split('\t') name = int(cols[0].replace('seq_', '')) seq = cols[1] exp = {} for i in range(len(samples)): exp[samples[i]] = int(cols[i + 2]) total[samples[i]] += int(cols[i + 2]) index = index + 1 if name in seq_obj: seq_obj[name].set_freq(exp) seq_obj[name].set_seq(seq) seq_obj = _normalize_seqs(seq_obj, total) return seq_obj, total, index
parse_ma_file
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(seq_obj, in_file):\n name = ''\n index = 1\n total = defaultdict(int)\n with open(in_file) as handle_in:\n line = handle_in.readline().strip()\n cols = line.split('\\t')\n samples = cols[2:]\n for line in handle_in:\n line = line.strip()\n cols = line.split('\\t')\n name = int(cols[0].replace('seq_', ''))\n seq = cols[1]\n exp = {}\n for i in range(len(samples)):\n exp[samples[i]] = int(cols[i + 2])\n total[samples[i]] += int(cols[i + 2])\n index = index + 1\n if name in seq_obj:\n seq_obj[name].set_freq(exp)\n seq_obj[name].set_seq(seq)\n seq_obj = _normalize_seqs(seq_obj, total)\n return seq_obj, total, index\n", "role": "user" }, { "content": "parse_ma_file", "role": "assistant" } ]
def x(cols, field='name'): field = field.lower() try: group = cols[2] attrs = cols[8].split(';') name = [attr.strip().split(' ')[1] for attr in attrs if attr.strip( ).split(' ')[0].lower().endswith(field)] if not name: name = [attr.strip().split(' ')[1] for attr in attrs if attr. strip().split(' ')[0].lower().endswith('gene_id')] if not name: name = ['None'] biotype = [attr.strip().split(' ')[1] for attr in attrs if attr. strip().split(' ')[0].lower().endswith('biotype')] if biotype: group = biotype[0] c = cols[0] s = int(cols[3]) e = int(cols[4]) st = cols[6] return [c, s, e, st, group, name[0]] except (Exception, e): logger.error(cols) logger.error('File is not in correct format') logger.error('Expect chr source feature start end . strand attributes') logger.error("Attributes are 'gene_name SNCA; gene_id ENSG; '") logger.error('The 3rd column is used as type of small RNA (like miRNA)' ) logger.error("at least should contains '; *name NAME; '") logger.error(e) raise
read_gtf_line
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cols, field='name'):\n field = field.lower()\n try:\n group = cols[2]\n attrs = cols[8].split(';')\n name = [attr.strip().split(' ')[1] for attr in attrs if attr.strip(\n ).split(' ')[0].lower().endswith(field)]\n if not name:\n name = [attr.strip().split(' ')[1] for attr in attrs if attr.\n strip().split(' ')[0].lower().endswith('gene_id')]\n if not name:\n name = ['None']\n biotype = [attr.strip().split(' ')[1] for attr in attrs if attr.\n strip().split(' ')[0].lower().endswith('biotype')]\n if biotype:\n group = biotype[0]\n c = cols[0]\n s = int(cols[3])\n e = int(cols[4])\n st = cols[6]\n return [c, s, e, st, group, name[0]]\n except (Exception, e):\n logger.error(cols)\n logger.error('File is not in correct format')\n logger.error('Expect chr source feature start end . strand attributes')\n logger.error(\"Attributes are 'gene_name SNCA; gene_id ENSG; '\")\n logger.error('The 3rd column is used as type of small RNA (like miRNA)'\n )\n logger.error(\"at least should contains '; *name NAME; '\")\n logger.error(e)\n raise\n", "role": "user" }, { "content": "read_gtf_line", "role": "assistant" } ]
def x(pos_a, pos_b): strd = '-' if pos_a[2] in pos_b[2]: strd = '+' if pos_a[2] in '+' and pos_b[2] in '+': lento5 = pos_a[0] - pos_b[1] + 1 lento3 = pos_a[1] - pos_b[1] + 1 if pos_a[2] in '+' and pos_b[2] in '-': lento5 = pos_a[1] - pos_b[0] + 1 lento3 = pos_a[0] - pos_b[1] + 1 if pos_a[2] in '-' and pos_b[2] in '+': lento5 = pos_a[0] - pos_b[1] + 1 lento3 = pos_a[1] - pos_b[0] + 1 if pos_a[2] in '-' and pos_b[2] in '-': lento3 = pos_a[0] - pos_b[0] + 1 lento5 = pos_a[1] - pos_b[1] + 1 else: lento5 = pos_a[0] - pos_b[0] + 1 lento3 = pos_a[1] - pos_b[1] + 1 return lento5, lento3, strd
position_in_feature
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(pos_a, pos_b):\n strd = '-'\n if pos_a[2] in pos_b[2]:\n strd = '+'\n if pos_a[2] in '+' and pos_b[2] in '+':\n lento5 = pos_a[0] - pos_b[1] + 1\n lento3 = pos_a[1] - pos_b[1] + 1\n if pos_a[2] in '+' and pos_b[2] in '-':\n lento5 = pos_a[1] - pos_b[0] + 1\n lento3 = pos_a[0] - pos_b[1] + 1\n if pos_a[2] in '-' and pos_b[2] in '+':\n lento5 = pos_a[0] - pos_b[1] + 1\n lento3 = pos_a[1] - pos_b[0] + 1\n if pos_a[2] in '-' and pos_b[2] in '-':\n lento3 = pos_a[0] - pos_b[0] + 1\n lento5 = pos_a[1] - pos_b[1] + 1\n else:\n lento5 = pos_a[0] - pos_b[0] + 1\n lento3 = pos_a[1] - pos_b[1] + 1\n return lento5, lento3, strd\n", "role": "user" }, { "content": "position_in_feature", "role": "assistant" } ]
def x(c, clus_obj, db, type_ann, feature_id='name'): id_sa, id_ea, id_id, id_idl, id_sta = 1, 2, 3, 4, 5 if type_ann == 'bed': id_sb = 7 id_eb = 8 id_stb = 11 id_tag = 9 ida = 0 clus_id = clus_obj.clus loci_id = clus_obj.loci db = os.path.splitext(db)[0] logger.debug('Type:%s\n' % type_ann) for cols in c.features(): if type_ann == 'gtf': cb, sb, eb, stb, db, tag = read_gtf_line(cols[6:], feature_id) else: sb = int(cols[id_sb]) eb = int(cols[id_eb]) stb = cols[id_stb] tag = cols[id_tag] id = int(cols[id_id]) idl = int(cols[id_idl]) if id in clus_id: clus = clus_id[id] sa = int(cols[id_sa]) ea = int(cols[id_ea]) ida += 1 lento5, lento3, strd = _position_in_feature([sa, ea, cols[ id_sta]], [sb, eb, stb]) if db in loci_id[idl].db_ann: ann = annotation(db, tag, strd, lento5, lento3) tdb = loci_id[idl].db_ann[db] tdb.add_db_ann(ida, ann) loci_id[idl].add_db(db, tdb) else: ann = annotation(db, tag, strd, lento5, lento3) tdb = dbannotation(1) tdb.add_db_ann(ida, ann) loci_id[idl].add_db(db, tdb) clus_id[id] = clus clus_obj.clus = clus_id clus_obj.loci = loci_id return clus_obj
anncluster
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(c, clus_obj, db, type_ann, feature_id='name'):\n id_sa, id_ea, id_id, id_idl, id_sta = 1, 2, 3, 4, 5\n if type_ann == 'bed':\n id_sb = 7\n id_eb = 8\n id_stb = 11\n id_tag = 9\n ida = 0\n clus_id = clus_obj.clus\n loci_id = clus_obj.loci\n db = os.path.splitext(db)[0]\n logger.debug('Type:%s\\n' % type_ann)\n for cols in c.features():\n if type_ann == 'gtf':\n cb, sb, eb, stb, db, tag = read_gtf_line(cols[6:], feature_id)\n else:\n sb = int(cols[id_sb])\n eb = int(cols[id_eb])\n stb = cols[id_stb]\n tag = cols[id_tag]\n id = int(cols[id_id])\n idl = int(cols[id_idl])\n if id in clus_id:\n clus = clus_id[id]\n sa = int(cols[id_sa])\n ea = int(cols[id_ea])\n ida += 1\n lento5, lento3, strd = _position_in_feature([sa, ea, cols[\n id_sta]], [sb, eb, stb])\n if db in loci_id[idl].db_ann:\n ann = annotation(db, tag, strd, lento5, lento3)\n tdb = loci_id[idl].db_ann[db]\n tdb.add_db_ann(ida, ann)\n loci_id[idl].add_db(db, tdb)\n else:\n ann = annotation(db, tag, strd, lento5, lento3)\n tdb = dbannotation(1)\n tdb.add_db_ann(ida, ann)\n loci_id[idl].add_db(db, tdb)\n clus_id[id] = clus\n clus_obj.clus = clus_id\n clus_obj.loci = loci_id\n return clus_obj\n", "role": "user" }, { "content": "anncluster", "role": "assistant" } ]
def x(bam_in, genome, out): if not genome: logger.info('No genome given. skipping.') return None out_file = op.join(out, op.basename(bam_in) + '_cov.tsv') if file_exists(out_file): return None fai = genome + '.fai' cov = pybedtools.BedTool(bam_in).genome_coverage(g=fai, max=1) cov.saveas(out_file) total = 0 for region in cov: if region[0] == 'genome' and int(region[1]) != 0: total += float(region[4]) logger.info('Total genome with sequences: %s ' % total)
detect_complexity
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(bam_in, genome, out):\n if not genome:\n logger.info('No genome given. skipping.')\n return None\n out_file = op.join(out, op.basename(bam_in) + '_cov.tsv')\n if file_exists(out_file):\n return None\n fai = genome + '.fai'\n cov = pybedtools.BedTool(bam_in).genome_coverage(g=fai, max=1)\n cov.saveas(out_file)\n total = 0\n for region in cov:\n if region[0] == 'genome' and int(region[1]) != 0:\n total += float(region[4])\n logger.info('Total genome with sequences: %s ' % total)\n", "role": "user" }, { "content": "detect_complexity", "role": "assistant" } ]
def x(bam_in, mask=None): seq_obj = defaultdict(int) if mask: mask_file = op.splitext(bam_in)[0] + '_mask.bam' if not file_exists(mask_file): pybedtools.BedTool(bam_file).intersect(b=mask, v=True).saveas( mask_file) bam_in = mask_file out_file = op.splitext(bam_in)[0] + '_rmlw.bam' run('samtools index %s' % bam_in) if not file_exists(bam_in + '.bai'): raise IOError( 'Failed to created bam index of %s. Try to do it manually' % bam_in ) bam_handle = pysam.AlignmentFile(bam_in, 'rb') with pysam.AlignmentFile(out_file, 'wb', template=bam_handle ) as out_handle: for read in bam_handle.fetch(): seq_name = int(read.query_name.replace('seq_', '')) match_size = [nts for oper, nts in read.cigartuples if oper == 0] subs_size = [nts for oper, nts in read.cigartuples if oper == 4] if match_size[0] < 17: continue if subs_size: if subs_size[0] > 3: continue try: nh = read.get_tag('NH') except KeyError: nh = 1 seq_obj[seq_name] = sequence(seq_name) seq_obj[seq_name].align = nh out_handle.write(read) return out_file, seq_obj
clean_bam_file
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(bam_in, mask=None):\n seq_obj = defaultdict(int)\n if mask:\n mask_file = op.splitext(bam_in)[0] + '_mask.bam'\n if not file_exists(mask_file):\n pybedtools.BedTool(bam_file).intersect(b=mask, v=True).saveas(\n mask_file)\n bam_in = mask_file\n out_file = op.splitext(bam_in)[0] + '_rmlw.bam'\n run('samtools index %s' % bam_in)\n if not file_exists(bam_in + '.bai'):\n raise IOError(\n 'Failed to created bam index of %s. Try to do it manually' % bam_in\n )\n bam_handle = pysam.AlignmentFile(bam_in, 'rb')\n with pysam.AlignmentFile(out_file, 'wb', template=bam_handle\n ) as out_handle:\n for read in bam_handle.fetch():\n seq_name = int(read.query_name.replace('seq_', ''))\n match_size = [nts for oper, nts in read.cigartuples if oper == 0]\n subs_size = [nts for oper, nts in read.cigartuples if oper == 4]\n if match_size[0] < 17:\n continue\n if subs_size:\n if subs_size[0] > 3:\n continue\n try:\n nh = read.get_tag('NH')\n except KeyError:\n nh = 1\n seq_obj[seq_name] = sequence(seq_name)\n seq_obj[seq_name].align = nh\n out_handle.write(read)\n return out_file, seq_obj\n", "role": "user" }, { "content": "clean_bam_file", "role": "assistant" } ]
def x(c, current_seq, MIN_SEQ, non_un_gl=False): current_loci = {} current_clus = {} sequence2clusters = defaultdict(set) lindex = 0 eindex = 0 previous_id = 0 for line in c.features(): c, start, end, name, score, strand, c_id = line name = int(name.replace('seq_', '')) pos = int(start) if strand == '+' else int(end) if name not in current_seq: continue if c.find('Un_gl') > -1 and non_un_gl: continue if c_id != previous_id: if previous_id > 0: if len(current_clus[eindex].idmembers) < MIN_SEQ: for s in current_clus[eindex].idmembers: sequence2clusters[s] = sequence2clusters[s] - set([ eindex]) del current_clus[eindex] logger.debug('detect_cluster: %s %s %s' % (c_id, previous_id, name) ) lindex += 1 eindex += 1 current_clus[eindex] = cluster(eindex) newpos = position(lindex, c, start, end, strand) current_loci[lindex] = newpos current_loci[lindex].end = int(end) current_loci[lindex].coverage[pos] += 1 size = range(pos, pos + current_seq[name].len) current_loci[lindex].counts.update(dict(zip(size, [current_seq[name ].total()] * current_seq[name].len))) current_clus[eindex].idmembers[name] = 1 current_clus[eindex].add_id_member([name], lindex) current_seq[name].add_pos(lindex, pos) previous_id = c_id sequence2clusters[name].add(eindex) logger.info('%s Clusters read' % eindex) metacluster_obj, cluster_id = _find_metaclusters(current_clus, sequence2clusters, current_seq, MIN_SEQ) return cluster_info_obj(current_clus, metacluster_obj, current_loci, current_seq)
detect_clusters
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(c, current_seq, MIN_SEQ, non_un_gl=False):\n current_loci = {}\n current_clus = {}\n sequence2clusters = defaultdict(set)\n lindex = 0\n eindex = 0\n previous_id = 0\n for line in c.features():\n c, start, end, name, score, strand, c_id = line\n name = int(name.replace('seq_', ''))\n pos = int(start) if strand == '+' else int(end)\n if name not in current_seq:\n continue\n if c.find('Un_gl') > -1 and non_un_gl:\n continue\n if c_id != previous_id:\n if previous_id > 0:\n if len(current_clus[eindex].idmembers) < MIN_SEQ:\n for s in current_clus[eindex].idmembers:\n sequence2clusters[s] = sequence2clusters[s] - set([\n eindex])\n del current_clus[eindex]\n logger.debug('detect_cluster: %s %s %s' % (c_id, previous_id, name)\n )\n lindex += 1\n eindex += 1\n current_clus[eindex] = cluster(eindex)\n newpos = position(lindex, c, start, end, strand)\n current_loci[lindex] = newpos\n current_loci[lindex].end = int(end)\n current_loci[lindex].coverage[pos] += 1\n size = range(pos, pos + current_seq[name].len)\n current_loci[lindex].counts.update(dict(zip(size, [current_seq[name\n ].total()] * current_seq[name].len)))\n current_clus[eindex].idmembers[name] = 1\n current_clus[eindex].add_id_member([name], lindex)\n current_seq[name].add_pos(lindex, pos)\n previous_id = c_id\n sequence2clusters[name].add(eindex)\n logger.info('%s Clusters read' % eindex)\n metacluster_obj, cluster_id = _find_metaclusters(current_clus,\n sequence2clusters, current_seq, MIN_SEQ)\n return cluster_info_obj(current_clus, metacluster_obj, current_loci,\n current_seq)\n", "role": "user" }, { "content": "detect_clusters", "role": "assistant" } ]
def x(clus_obj, sequence2clusters, current_seq, min_seqs): seen = defaultdict(int) metacluster = defaultdict(set) c_index = len(sequence2clusters) logger.info('Creating meta-clusters based on shared sequences: %s' % c_index) meta_idx = 1 bar = ProgressBar(maxval=c_index) bar.start() bar.update() for itern, name in enumerate(sequence2clusters): clusters = sequence2clusters[name] if len(clusters) == 0: c_index -= 1 continue current_seq[name].align = 1 meta_idx += 1 bar.update(itern) already_in = _common(clusters, seen) _update(clusters, meta_idx, seen) metacluster[meta_idx] = metacluster[meta_idx].union(clusters) if already_in: for seen_metacluster in already_in: clusters2merge = metacluster[seen_metacluster] metacluster[meta_idx] = metacluster[meta_idx].union( clusters2merge) _update(clusters2merge, meta_idx, seen) del metacluster[seen_metacluster] logger.info('%s metaclusters from %s sequences' % (len(metacluster), c_index)) return metacluster, seen
find_metaclusters
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(clus_obj, sequence2clusters, current_seq, min_seqs):\n seen = defaultdict(int)\n metacluster = defaultdict(set)\n c_index = len(sequence2clusters)\n logger.info('Creating meta-clusters based on shared sequences: %s' %\n c_index)\n meta_idx = 1\n bar = ProgressBar(maxval=c_index)\n bar.start()\n bar.update()\n for itern, name in enumerate(sequence2clusters):\n clusters = sequence2clusters[name]\n if len(clusters) == 0:\n c_index -= 1\n continue\n current_seq[name].align = 1\n meta_idx += 1\n bar.update(itern)\n already_in = _common(clusters, seen)\n _update(clusters, meta_idx, seen)\n metacluster[meta_idx] = metacluster[meta_idx].union(clusters)\n if already_in:\n for seen_metacluster in already_in:\n clusters2merge = metacluster[seen_metacluster]\n metacluster[meta_idx] = metacluster[meta_idx].union(\n clusters2merge)\n _update(clusters2merge, meta_idx, seen)\n del metacluster[seen_metacluster]\n logger.info('%s metaclusters from %s sequences' % (len(metacluster),\n c_index))\n return metacluster, seen\n", "role": "user" }, { "content": "find_metaclusters", "role": "assistant" } ]
def x(clus_obj, min_seqs): logger.info('Creating meta-clusters based on shared sequences.') seen = defaultdict() metacluster = defaultdict(list) c_index = clus_obj.keys() meta_idx = 0 with ProgressBar(maxval=len(c_index), redirect_stdout=True) as p: for itern, c in enumerate(c_index): p.update(itern) clus = clus_obj[c] if len(clus.idmembers.keys()) < min_seqs: del clus_obj[c] continue logger.debug('reading cluster %s' % c) logger.debug('loci2seq %s' % clus.loci2seq) already_in, not_in = _get_seqs_from_cluster(clus.idmembers.keys (), seen) logger.debug('seen %s news %s' % (already_in, not_in)) meta_idx += 1 metacluster[meta_idx].append(c) seen.update(dict(zip(not_in, [meta_idx] * len(not_in)))) if len(already_in) > 0: logger.debug('seen in %s' % already_in) for eindex in already_in: for cluster in metacluster[eindex]: metacluster[meta_idx].append(cluster) prev_clus = clus_obj[cluster] logger.debug('_find_families: prev %s current %s' % (eindex, clus.id)) seqs_in = prev_clus.idmembers.keys() seen.update(dict(zip(seqs_in, [meta_idx] * len( seqs_in)))) del metacluster[eindex] logger.info('%s clusters merged' % len(metacluster)) return metacluster, seen
find_families_deprecated
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(clus_obj, min_seqs):\n logger.info('Creating meta-clusters based on shared sequences.')\n seen = defaultdict()\n metacluster = defaultdict(list)\n c_index = clus_obj.keys()\n meta_idx = 0\n with ProgressBar(maxval=len(c_index), redirect_stdout=True) as p:\n for itern, c in enumerate(c_index):\n p.update(itern)\n clus = clus_obj[c]\n if len(clus.idmembers.keys()) < min_seqs:\n del clus_obj[c]\n continue\n logger.debug('reading cluster %s' % c)\n logger.debug('loci2seq %s' % clus.loci2seq)\n already_in, not_in = _get_seqs_from_cluster(clus.idmembers.keys\n (), seen)\n logger.debug('seen %s news %s' % (already_in, not_in))\n meta_idx += 1\n metacluster[meta_idx].append(c)\n seen.update(dict(zip(not_in, [meta_idx] * len(not_in))))\n if len(already_in) > 0:\n logger.debug('seen in %s' % already_in)\n for eindex in already_in:\n for cluster in metacluster[eindex]:\n metacluster[meta_idx].append(cluster)\n prev_clus = clus_obj[cluster]\n logger.debug('_find_families: prev %s current %s' %\n (eindex, clus.id))\n seqs_in = prev_clus.idmembers.keys()\n seen.update(dict(zip(seqs_in, [meta_idx] * len(\n seqs_in))))\n del metacluster[eindex]\n logger.info('%s clusters merged' % len(metacluster))\n return metacluster, seen\n", "role": "user" }, { "content": "find_families_deprecated", "role": "assistant" } ]
def x(clus_obj): new_cluster = {} for cid in clus_obj.clus: cluster = clus_obj.clus[cid] cluster.update() logger.debug('peak calling for %s' % cid) bigger = cluster.locimaxid if bigger in clus_obj.loci: s, e = min(clus_obj.loci[bigger].counts.keys()), max(clus_obj. loci[bigger].counts.keys()) scale = s if clus_obj.loci[bigger].strand == '-': scale = e logger.debug('bigger %s at %s-%s' % (bigger, s, e)) dt = np.array([0] * (abs(e - s) + 12)) for pos in clus_obj.loci[bigger].counts: ss = abs(int(pos) - scale) + 5 dt[ss] += clus_obj.loci[bigger].counts[pos] x = np.array(range(0, len(dt))) logger.debug('x %s and y %s' % (x, dt)) if len(x) > 35 + 12: peaks = list(np.array(pysen.pysenMMean(x, dt)) - 5) logger.debug(peaks) else: peaks = ['short'] cluster.peaks = peaks new_cluster[cid] = cluster clus_obj.clus = new_cluster return clus_obj
peak_calling
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(clus_obj):\n new_cluster = {}\n for cid in clus_obj.clus:\n cluster = clus_obj.clus[cid]\n cluster.update()\n logger.debug('peak calling for %s' % cid)\n bigger = cluster.locimaxid\n if bigger in clus_obj.loci:\n s, e = min(clus_obj.loci[bigger].counts.keys()), max(clus_obj.\n loci[bigger].counts.keys())\n scale = s\n if clus_obj.loci[bigger].strand == '-':\n scale = e\n logger.debug('bigger %s at %s-%s' % (bigger, s, e))\n dt = np.array([0] * (abs(e - s) + 12))\n for pos in clus_obj.loci[bigger].counts:\n ss = abs(int(pos) - scale) + 5\n dt[ss] += clus_obj.loci[bigger].counts[pos]\n x = np.array(range(0, len(dt)))\n logger.debug('x %s and y %s' % (x, dt))\n if len(x) > 35 + 12:\n peaks = list(np.array(pysen.pysenMMean(x, dt)) - 5)\n logger.debug(peaks)\n else:\n peaks = ['short']\n cluster.peaks = peaks\n new_cluster[cid] = cluster\n clus_obj.clus = new_cluster\n return clus_obj\n", "role": "user" }, { "content": "peak_calling", "role": "assistant" } ]
def x(args): if args.fasta: name = None seq = '' reads = dict() with open(args.fasta) as in_handle: for line in in_handle: if line.startswith('>'): if name: reads.update(_generate_reads(seq, name)) seq = '' name = line[1:-1] else: seq += line.strip() reads.update(_generate_reads(seq, name)) _write_reads(reads, args.out)
simulate
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(args):\n if args.fasta:\n name = None\n seq = ''\n reads = dict()\n with open(args.fasta) as in_handle:\n for line in in_handle:\n if line.startswith('>'):\n if name:\n reads.update(_generate_reads(seq, name))\n seq = ''\n name = line[1:-1]\n else:\n seq += line.strip()\n reads.update(_generate_reads(seq, name))\n _write_reads(reads, args.out)\n", "role": "user" }, { "content": "simulate", "role": "assistant" } ]
def x(seq, name): reads = dict() if len(seq) < 130 and len(seq) > 70: reads.update(_mature(seq[:40], 0, name)) reads.update(_mature(seq[-40:], len(seq) - 40, name)) reads.update(_noise(seq, name)) reads.update(_noise(seq, name, 25)) return reads
generate_reads
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(seq, name):\n reads = dict()\n if len(seq) < 130 and len(seq) > 70:\n reads.update(_mature(seq[:40], 0, name))\n reads.update(_mature(seq[-40:], len(seq) - 40, name))\n reads.update(_noise(seq, name))\n reads.update(_noise(seq, name, 25))\n return reads\n", "role": "user" }, { "content": "generate_reads", "role": "assistant" } ]
def x(subseq, absolute, c, size=33, total=5000): reads = dict() probs = [0.1, 0.2, 0.4, 0.2, 0.1] end = 5 + size error = [-2, -1, 0, 1, 2] for error5 in error: for error3 in error: s = 5 - error5 e = end - error3 seen = subseq[s:e] counts = int(probs[error5 + 2] * probs[error3 + 2] * total) + 1 name = 'seq_%s_%s_%s_x%s' % (c, s + absolute, e + absolute, counts) reads[name] = seen, counts return reads
mature
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(subseq, absolute, c, size=33, total=5000):\n reads = dict()\n probs = [0.1, 0.2, 0.4, 0.2, 0.1]\n end = 5 + size\n error = [-2, -1, 0, 1, 2]\n for error5 in error:\n for error3 in error:\n s = 5 - error5\n e = end - error3\n seen = subseq[s:e]\n counts = int(probs[error5 + 2] * probs[error3 + 2] * total) + 1\n name = 'seq_%s_%s_%s_x%s' % (c, s + absolute, e + absolute, counts)\n reads[name] = seen, counts\n return reads\n", "role": "user" }, { "content": "mature", "role": "assistant" } ]
def x(seq, c, size=33, total=1000): reads = dict() seen = 0 while seen < total: s = random.randint(0, len(seq) - size) e = s + size + random.randint(-5, 5) p = random.uniform(0, 0.1) counts = int(p * total) + 1 seen += counts name = 'seq_%s_%s_%s_x%s' % (c, s, e, counts) reads[name] = seq[s:e], counts return reads
noise
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(seq, c, size=33, total=1000):\n reads = dict()\n seen = 0\n while seen < total:\n s = random.randint(0, len(seq) - size)\n e = s + size + random.randint(-5, 5)\n p = random.uniform(0, 0.1)\n counts = int(p * total) + 1\n seen += counts\n name = 'seq_%s_%s_%s_x%s' % (c, s, e, counts)\n reads[name] = seq[s:e], counts\n return reads\n", "role": "user" }, { "content": "noise", "role": "assistant" } ]
def x(reads, prefix): out_ma = prefix + '.ma' out_fasta = prefix + '.fasta' out_real = prefix + '.txt' with open(out_ma, 'w') as ma_handle: print('id\tseq\tsample', file=ma_handle, end='') with open(out_fasta, 'w') as fa_handle: with open(out_real, 'w') as read_handle: for idx, r in enumerate(reads): info = r.split('_') print('seq_%s\t%s\t%s' % (idx, reads[r][0], reads[r][1] ), file=ma_handle, end='') print('>seq_%s\n%s' % (idx, reads[r][0]), file= fa_handle, end='') print('%s\t%s\t%s\t%s\t%s\t%s\t%s' % (idx, r, reads[r][ 0], reads[r][1], info[1], info[2], info[3]), file= read_handle, end='')
write_reads
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(reads, prefix):\n out_ma = prefix + '.ma'\n out_fasta = prefix + '.fasta'\n out_real = prefix + '.txt'\n with open(out_ma, 'w') as ma_handle:\n print('id\\tseq\\tsample', file=ma_handle, end='')\n with open(out_fasta, 'w') as fa_handle:\n with open(out_real, 'w') as read_handle:\n for idx, r in enumerate(reads):\n info = r.split('_')\n print('seq_%s\\t%s\\t%s' % (idx, reads[r][0], reads[r][1]\n ), file=ma_handle, end='')\n print('>seq_%s\\n%s' % (idx, reads[r][0]), file=\n fa_handle, end='')\n print('%s\\t%s\\t%s\\t%s\\t%s\\t%s\\t%s' % (idx, r, reads[r][\n 0], reads[r][1], info[1], info[2], info[3]), file=\n read_handle, end='')\n", "role": "user" }, { "content": "write_reads", "role": "assistant" } ]
def x(args): logger.info('Reading sequeces') data = parse_ma_file(args.ma) logger.info('Get sequences from sam') is_align = _read_sam(args.sam) is_json, is_db = _read_json(args.json) res = _summarise_sam(data, is_align, is_json, is_db) _write_suma(res, os.path.join(args.out, 'stats_align.dat')) logger.info('Done')
stats
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(args):\n logger.info('Reading sequeces')\n data = parse_ma_file(args.ma)\n logger.info('Get sequences from sam')\n is_align = _read_sam(args.sam)\n is_json, is_db = _read_json(args.json)\n res = _summarise_sam(data, is_align, is_json, is_db)\n _write_suma(res, os.path.join(args.out, 'stats_align.dat'))\n logger.info('Done')\n", "role": "user" }, { "content": "stats", "role": "assistant" } ]
def x(fn_json): is_json = set() is_db = {} with open(fn_json) as handle: data = json.load(handle) for item in list(data[0].values()): seqs_name = [list(x.keys()) for x in item['seqs']] db_name = item['valid'] if 'valid' in item else None [is_json.add(name[0]) for name in seqs_name] if db_name: [is_db.update({name[0]: ','.join(db_name)}) for name in seqs_name] return is_json, is_db
read_json
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(fn_json):\n is_json = set()\n is_db = {}\n with open(fn_json) as handle:\n data = json.load(handle)\n for item in list(data[0].values()):\n seqs_name = [list(x.keys()) for x in item['seqs']]\n db_name = item['valid'] if 'valid' in item else None\n [is_json.add(name[0]) for name in seqs_name]\n if db_name:\n [is_db.update({name[0]: ','.join(db_name)}) for name in\n seqs_name]\n return is_json, is_db\n", "role": "user" }, { "content": "read_json", "role": "assistant" } ]
def x(cmd, data=None, checks=None, region=None, log_error=True, log_stdout= False): try: logger.debug(' '.join(str(x) for x in cmd) if not isinstance(cmd, basestring) else cmd) _do_run(cmd, checks, log_stdout) except: if log_error: logger.info('error at command') raise
run
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cmd, data=None, checks=None, region=None, log_error=True, log_stdout=\n False):\n try:\n logger.debug(' '.join(str(x) for x in cmd) if not isinstance(cmd,\n basestring) else cmd)\n _do_run(cmd, checks, log_stdout)\n except:\n if log_error:\n logger.info('error at command')\n raise\n", "role": "user" }, { "content": "run", "role": "assistant" } ]
def x(cmd): if isinstance(cmd, basestring): if cmd.find(' | ') > 0 or cmd.find('>(') or cmd.find('<('): return 'set -o pipefail; ' + cmd, True, find_bash() else: return cmd, True, None else: return [str(x) for x in cmd], False, None
normalize_cmd_args
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cmd):\n if isinstance(cmd, basestring):\n if cmd.find(' | ') > 0 or cmd.find('>(') or cmd.find('<('):\n return 'set -o pipefail; ' + cmd, True, find_bash()\n else:\n return cmd, True, None\n else:\n return [str(x) for x in cmd], False, None\n", "role": "user" }, { "content": "normalize_cmd_args", "role": "assistant" } ]
def x(cmd, checks, log_stdout=False): cmd, shell_arg, executable_arg = _normalize_cmd_args(cmd) s = subprocess.Popen(cmd, shell=shell_arg, executable=executable_arg, stdout=subprocess.PIPE, stderr=subprocess.STDOUT, close_fds=True) debug_stdout = collections.deque(maxlen=100) while 1: line = s.stdout.readline() if line: debug_stdout.append(line) if log_stdout: logger.debug(line.rstrip()) else: logger.debug(line.rstrip()) exitcode = s.poll() if exitcode is not None: for line in s.stdout: debug_stdout.append(line) if exitcode is not None and exitcode != 0: error_msg = ' '.join(cmd) if not isinstance(cmd, basestring ) else cmd error_msg += '\n' error_msg += ''.join(debug_stdout) s.communicate() s.stdout.close() raise subprocess.CalledProcessError(exitcode, error_msg) else: break s.communicate() s.stdout.close() if checks: for check in checks: if not check(): raise IOError('External command failed')
do_run
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cmd, checks, log_stdout=False):\n cmd, shell_arg, executable_arg = _normalize_cmd_args(cmd)\n s = subprocess.Popen(cmd, shell=shell_arg, executable=executable_arg,\n stdout=subprocess.PIPE, stderr=subprocess.STDOUT, close_fds=True)\n debug_stdout = collections.deque(maxlen=100)\n while 1:\n line = s.stdout.readline()\n if line:\n debug_stdout.append(line)\n if log_stdout:\n logger.debug(line.rstrip())\n else:\n logger.debug(line.rstrip())\n exitcode = s.poll()\n if exitcode is not None:\n for line in s.stdout:\n debug_stdout.append(line)\n if exitcode is not None and exitcode != 0:\n error_msg = ' '.join(cmd) if not isinstance(cmd, basestring\n ) else cmd\n error_msg += '\\n'\n error_msg += ''.join(debug_stdout)\n s.communicate()\n s.stdout.close()\n raise subprocess.CalledProcessError(exitcode, error_msg)\n else:\n break\n s.communicate()\n s.stdout.close()\n if checks:\n for check in checks:\n if not check():\n raise IOError('External command failed')\n", "role": "user" }, { "content": "do_run", "role": "assistant" } ]
def x(list_c, loci_obj, seq_obj): seqs = defaultdict(set) for c in list_c.values(): for l in c.loci2seq: [seqs[s].add(c.id) for s in c.loci2seq[l]] common = [s for s in seqs if len(seqs[s]) > 1] seqs_in_c = defaultdict(float) for c in list_c.values(): for l in c.loci2seq: for s in c.loci2seq[l]: if s in common: pos = seq_obj[s].pos[l] cov = 1.0 * loci_obj[l].coverage[pos] if seqs_in_c[s, c.id] < cov: seqs_in_c[s, c.id] = cov seqs_in_c = _transform(seqs_in_c) return seqs_in_c
dict_seq_locus
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(list_c, loci_obj, seq_obj):\n seqs = defaultdict(set)\n for c in list_c.values():\n for l in c.loci2seq:\n [seqs[s].add(c.id) for s in c.loci2seq[l]]\n common = [s for s in seqs if len(seqs[s]) > 1]\n seqs_in_c = defaultdict(float)\n for c in list_c.values():\n for l in c.loci2seq:\n for s in c.loci2seq[l]:\n if s in common:\n pos = seq_obj[s].pos[l]\n cov = 1.0 * loci_obj[l].coverage[pos]\n if seqs_in_c[s, c.id] < cov:\n seqs_in_c[s, c.id] = cov\n seqs_in_c = _transform(seqs_in_c)\n return seqs_in_c\n", "role": "user" }, { "content": "dict_seq_locus", "role": "assistant" } ]
def x(self, data): for hypo in self.Values(): like = self.Likelihood(data, hypo) self.Mult(hypo, like) self.Normalize()
Update
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, data):\n for hypo in self.Values():\n like = self.Likelihood(data, hypo)\n self.Mult(hypo, like)\n self.Normalize()\n", "role": "user" }, { "content": "Update", "role": "assistant" } ]
def x(self, data, hypo): mix = self.loci[hypo] like = mix[data] return like
Likelihood
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, data, hypo):\n mix = self.loci[hypo]\n like = mix[data]\n return like\n", "role": "user" }, { "content": "Likelihood", "role": "assistant" } ]
def x(clus_obj, index): current = clus_obj.clus clus_seqt = clus_obj.seq clus_locit = clus_obj.loci itern = 0 for idc in current.keys(): itern += 1 timestamp = str(idc) seqListTemp = () f = open('/tmp/' + timestamp + '.fa', 'w') for idl in current[idc].loci2seq.keys(): seqListTemp = list(set(seqListTemp).union(current[idc].loci2seq [idl])) maxscore = 0 for s in seqListTemp: score = calculate_size(clus_seqt[s].freq) maxscore = max(maxscore, score) clus_seqt[s].score = score seq = clus_seqt[s] f.write('>' + s + '\n' + seq.seq + '\n') f.close() locilen_sorted = sorted(current[idc].locilen.iteritems(), key= operator.itemgetter(1), reverse=True) lmax = clus_locit[locilen_sorted[0][0]] f = open('/tmp/' + timestamp + '.bed', 'w') f.write('%s\t%s\t%s\t.\t.\t%s\n' % (lmax.chr, lmax.start, lmax.end, lmax.strand)) f.close() os.system('bedtools getfasta -s -fi ' + index + ' -bed /tmp/' + timestamp + '.bed -fo /tmp/' + timestamp + '.pre.fa') os.system('bowtie2-build /tmp/' + timestamp + '.pre.fa /tmp/' + timestamp + '.pre.ind >/dev/null 2>&1') os.system( 'bowtie2 --rdg 7,3 --mp 4 --end-to-end --no-head --no-sq -D 20 -R 3 -N 0 -i S,1,0.8 -L 3 -f /tmp/' + timestamp + '.pre.ind /tmp/' + timestamp + '.fa -S /tmp/' + timestamp + '.map >>bowtie.log 2>&1') f = open('/tmp/' + timestamp + '.map', 'r') seqpos = {} minv = 10000000 for line in f: line = line.strip() cols = line.split('\t') seqpos[cols[0]] = int(cols[3]) if minv > int(cols[3]): minv = int(cols[3]) f.close() seqpos_sorted = sorted(seqpos.iteritems(), key=operator.itemgetter( 1), reverse=False) showseq = '' showseq_plain = '' for s, pos in seqpos_sorted: ratio = clus_seqt[s].score * 1.0 / maxscore * 100.0 realScore = math.log(ratio, 2) * 2 if realScore < 0: realScore = 0 showseq_plain += ( '<br>%s<a style = "font-size:%.0fpx;"href = javascript:loadSeq("%s")>%s</a>' % (''.join('.' for i in range(pos - 1)), realScore + 10, s, clus_seqt[s].seq)) current[idc].showseq_plain = showseq_plain os.system('rm /tmp/' + timestamp + '*') clus_obj.clus = current clus_obj.seq = clus_seqt return clus_obj
show_seq
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(clus_obj, index):\n current = clus_obj.clus\n clus_seqt = clus_obj.seq\n clus_locit = clus_obj.loci\n itern = 0\n for idc in current.keys():\n itern += 1\n timestamp = str(idc)\n seqListTemp = ()\n f = open('/tmp/' + timestamp + '.fa', 'w')\n for idl in current[idc].loci2seq.keys():\n seqListTemp = list(set(seqListTemp).union(current[idc].loci2seq\n [idl]))\n maxscore = 0\n for s in seqListTemp:\n score = calculate_size(clus_seqt[s].freq)\n maxscore = max(maxscore, score)\n clus_seqt[s].score = score\n seq = clus_seqt[s]\n f.write('>' + s + '\\n' + seq.seq + '\\n')\n f.close()\n locilen_sorted = sorted(current[idc].locilen.iteritems(), key=\n operator.itemgetter(1), reverse=True)\n lmax = clus_locit[locilen_sorted[0][0]]\n f = open('/tmp/' + timestamp + '.bed', 'w')\n f.write('%s\\t%s\\t%s\\t.\\t.\\t%s\\n' % (lmax.chr, lmax.start, lmax.end,\n lmax.strand))\n f.close()\n os.system('bedtools getfasta -s -fi ' + index + ' -bed /tmp/' +\n timestamp + '.bed -fo /tmp/' + timestamp + '.pre.fa')\n os.system('bowtie2-build /tmp/' + timestamp + '.pre.fa /tmp/' +\n timestamp + '.pre.ind >/dev/null 2>&1')\n os.system(\n 'bowtie2 --rdg 7,3 --mp 4 --end-to-end --no-head --no-sq -D 20 -R 3 -N 0 -i S,1,0.8 -L 3 -f /tmp/'\n + timestamp + '.pre.ind /tmp/' + timestamp + '.fa -S /tmp/' +\n timestamp + '.map >>bowtie.log 2>&1')\n f = open('/tmp/' + timestamp + '.map', 'r')\n seqpos = {}\n minv = 10000000\n for line in f:\n line = line.strip()\n cols = line.split('\\t')\n seqpos[cols[0]] = int(cols[3])\n if minv > int(cols[3]):\n minv = int(cols[3])\n f.close()\n seqpos_sorted = sorted(seqpos.iteritems(), key=operator.itemgetter(\n 1), reverse=False)\n showseq = ''\n showseq_plain = ''\n for s, pos in seqpos_sorted:\n ratio = clus_seqt[s].score * 1.0 / maxscore * 100.0\n realScore = math.log(ratio, 2) * 2\n if realScore < 0:\n realScore = 0\n showseq_plain += (\n '<br>%s<a style = \"font-size:%.0fpx;\"href = javascript:loadSeq(\"%s\")>%s</a>'\n % (''.join('.' for i in range(pos - 1)), realScore + 10, s,\n clus_seqt[s].seq))\n current[idc].showseq_plain = showseq_plain\n os.system('rm /tmp/' + timestamp + '*')\n clus_obj.clus = current\n clus_obj.seq = clus_seqt\n return clus_obj\n", "role": "user" }, { "content": "show_seq", "role": "assistant" } ]
def x(s, t): for ids in s: obj = s[ids] [obj.norm_freq.update({sample: 1.0 * obj.freq[sample] / (t[sample] + 1) * 1000000}) for sample in obj.norm_freq] s[ids] = obj return s
normalize_seqs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(s, t):\n for ids in s:\n obj = s[ids]\n [obj.norm_freq.update({sample: 1.0 * obj.freq[sample] / (t[sample] +\n 1) * 1000000}) for sample in obj.norm_freq]\n s[ids] = obj\n return s\n", "role": "user" }, { "content": "normalize_seqs", "role": "assistant" } ]
def x(bam_in, precursors): a = pybedtools.BedTool(bam_in) b = pybedtools.BedTool(precursors) c = a.intersect(b, u=True) out_file = utils.splitext_plus(op.basename(bam_in))[0] + '_clean.bam' c.saveas(out_file) return op.abspath(out_file)
prepare_bam
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(bam_in, precursors):\n a = pybedtools.BedTool(bam_in)\n b = pybedtools.BedTool(precursors)\n c = a.intersect(b, u=True)\n out_file = utils.splitext_plus(op.basename(bam_in))[0] + '_clean.bam'\n c.saveas(out_file)\n return op.abspath(out_file)\n", "role": "user" }, { "content": "prepare_bam", "role": "assistant" } ]
def x(bed_file): new_bed = utils.splitext_plus(bed_file)[0] + '_order.bed' with open(bed_file) as in_handle: with open(new_bed, 'w') as out_handle: for line in in_handle: cols = line.strip().split('\t') cols[3] = _select_anno(cols[3]) + '_' + cols[4] cols[4] = '0' print('\t'.join(cols), file=out_handle, end='') return new_bed
reorder_columns
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(bed_file):\n new_bed = utils.splitext_plus(bed_file)[0] + '_order.bed'\n with open(bed_file) as in_handle:\n with open(new_bed, 'w') as out_handle:\n for line in in_handle:\n cols = line.strip().split('\\t')\n cols[3] = _select_anno(cols[3]) + '_' + cols[4]\n cols[4] = '0'\n print('\\t'.join(cols), file=out_handle, end='')\n return new_bed\n", "role": "user" }, { "content": "reorder_columns", "role": "assistant" } ]
def x(cov_file): new_cov = utils.splitext_plus(cov_file)[0] + '_fix.cov' with open(cov_file) as in_handle: with open(new_cov, 'w') as out_handle: for line in in_handle: cols = line.strip().split('\t') cols[4] = cols[6] print('\t'.join(cols[0:6]), file=out_handle, end='') return new_cov
fix_score_column
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cov_file):\n new_cov = utils.splitext_plus(cov_file)[0] + '_fix.cov'\n with open(cov_file) as in_handle:\n with open(new_cov, 'w') as out_handle:\n for line in in_handle:\n cols = line.strip().split('\\t')\n cols[4] = cols[6]\n print('\\t'.join(cols[0:6]), file=out_handle, end='')\n return new_cov\n", "role": "user" }, { "content": "fix_score_column", "role": "assistant" } ]
def x(bam_in, bed_file, out_dir, prefix): bed_file = _reorder_columns(bed_file) counts_reads_cmd = ( 'coverageBed -s -counts -b {bam_in} -a {bed_file} | sort -k4,4 > {out_dir}/loci.cov' ) with utils.chdir(out_dir): run(counts_reads_cmd.format(min_trimmed_read_len= min_trimmed_read_len, max_trimmed_read_len=max_trimmed_read_len, **locals()), 'Run counts_reads') loci_file = _fix_score_column(op.join(out_dir, 'loci.cov')) return loci_file
detect_regions
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(bam_in, bed_file, out_dir, prefix):\n bed_file = _reorder_columns(bed_file)\n counts_reads_cmd = (\n 'coverageBed -s -counts -b {bam_in} -a {bed_file} | sort -k4,4 > {out_dir}/loci.cov'\n )\n with utils.chdir(out_dir):\n run(counts_reads_cmd.format(min_trimmed_read_len=\n min_trimmed_read_len, max_trimmed_read_len=max_trimmed_read_len,\n **locals()), 'Run counts_reads')\n loci_file = _fix_score_column(op.join(out_dir, 'loci.cov'))\n return loci_file\n", "role": "user" }, { "content": "detect_regions", "role": "assistant" } ]
def x(cov_file, min_reads): new_cov = op.join(op.dirname(cov_file), 'feat_antisense.txt') with open(cov_file) as in_handle: with open(new_cov, 'w') as out_handle: print('name\tantisense', file=out_handle, end='') for line in in_handle: cols = line.strip().split('\t') cols[6] = 0 if cols[6] < min_reads else cols[6] print('%s\t%s' % (cols[3], cols[6]), file=out_handle, end='') return new_cov
order_antisense_column
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cov_file, min_reads):\n new_cov = op.join(op.dirname(cov_file), 'feat_antisense.txt')\n with open(cov_file) as in_handle:\n with open(new_cov, 'w') as out_handle:\n print('name\\tantisense', file=out_handle, end='')\n for line in in_handle:\n cols = line.strip().split('\\t')\n cols[6] = 0 if cols[6] < min_reads else cols[6]\n print('%s\\t%s' % (cols[3], cols[6]), file=out_handle, end='')\n return new_cov\n", "role": "user" }, { "content": "order_antisense_column", "role": "assistant" } ]
def x(bam_in, loci_file, out_dir): data = Counter() a = pybedtools.BedTool(bam_in) b = pybedtools.BedTool(loci_file) c = a.intersect(b, s=True, bed=True, wo=True) for line in c: end = int(line[1]) + 1 + int(line[2]) if line[5] == '+' else int(line [1]) + 1 start = int(line[1]) + 1 if line[5] == '+' else int(line[1]) + 1 + int( line[2]) side5 = '%s\t5p\t%s' % (line[15], start) side3 = '%s\t3p\t%s' % (line[15], end) data[side5] += 1 data[side3] += 1 counts_reads = op.join(out_dir, 'locus_readpos.counts') with open(counts_reads, 'w') as out_handle: for k in data: print(k, file=out_handle, end='') return counts_reads
reads_per_position
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(bam_in, loci_file, out_dir):\n data = Counter()\n a = pybedtools.BedTool(bam_in)\n b = pybedtools.BedTool(loci_file)\n c = a.intersect(b, s=True, bed=True, wo=True)\n for line in c:\n end = int(line[1]) + 1 + int(line[2]) if line[5] == '+' else int(line\n [1]) + 1\n start = int(line[1]) + 1 if line[5] == '+' else int(line[1]) + 1 + int(\n line[2])\n side5 = '%s\\t5p\\t%s' % (line[15], start)\n side3 = '%s\\t3p\\t%s' % (line[15], end)\n data[side5] += 1\n data[side3] += 1\n counts_reads = op.join(out_dir, 'locus_readpos.counts')\n with open(counts_reads, 'w') as out_handle:\n for k in data:\n print(k, file=out_handle, end='')\n return counts_reads\n", "role": "user" }, { "content": "reads_per_position", "role": "assistant" } ]
def x(bam_in, loci_file, reference, out_dir): lenvec_plus = op.join(out_dir, 'genomic_lenvec.plus') lenvec_minus = op.join(out_dir, 'genomic_lenvec.minus') compute_genomic_cmd = ( 'compute_genomic_lenvectors {bam_in} {lenvec_plus} {lenvec_minus} {min_len} {max_len} ' ) index_genomic_cmd = 'index_genomic_lenvectors {lenvec} ' genomic_lenvec = op.join(out_dir, 'genomic_lenvec') feat_len_file = op.join(out_dir, 'feat_lengths.txt') compute_locus_cmd = ( 'compute_locus_lenvectors {loci_file} {genomic_lenvec} {min_len} {max_len} > {feat_len_file}' ) cov_S_file = op.join(out_dir, 'loci.cov_anti') coverage_anti_cmd = ( 'coverageBed -S -counts -b {bam_in} -a {loci_file} > {cov_S_file}') feat_posentropy = op.join(out_dir, 'feat_posentropy.txt') entropy_cmd = 'compute_locus_entropy.rb {counts_reads} > {feat_posentropy}' with utils.chdir(out_dir): run(compute_genomic_cmd.format(min_len=min_trimmed_read_len, max_len=max_trimmed_read_len, **locals()), 'Run compute_genomic') run(index_genomic_cmd.format(lenvec=lenvec_plus), 'Run index in plus') run(index_genomic_cmd.format(lenvec=lenvec_minus), 'Run index in minus' ) run(compute_locus_cmd.format(min_len=min_trimmed_read_len, max_len= max_trimmed_read_len, **locals()), 'Run compute locus') run(coverage_anti_cmd.format(**locals()), 'Run coverage antisense') feat_antisense = _order_antisense_column(cov_S_file, min_trimmed_read_len) counts_reads = _reads_per_position(bam_in, loci_file, out_dir) run(entropy_cmd.format(**locals()), 'Run entropy') rnafold = calculate_structure(loci_file, reference)
create_features
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(bam_in, loci_file, reference, out_dir):\n lenvec_plus = op.join(out_dir, 'genomic_lenvec.plus')\n lenvec_minus = op.join(out_dir, 'genomic_lenvec.minus')\n compute_genomic_cmd = (\n 'compute_genomic_lenvectors {bam_in} {lenvec_plus} {lenvec_minus} {min_len} {max_len} '\n )\n index_genomic_cmd = 'index_genomic_lenvectors {lenvec} '\n genomic_lenvec = op.join(out_dir, 'genomic_lenvec')\n feat_len_file = op.join(out_dir, 'feat_lengths.txt')\n compute_locus_cmd = (\n 'compute_locus_lenvectors {loci_file} {genomic_lenvec} {min_len} {max_len} > {feat_len_file}'\n )\n cov_S_file = op.join(out_dir, 'loci.cov_anti')\n coverage_anti_cmd = (\n 'coverageBed -S -counts -b {bam_in} -a {loci_file} > {cov_S_file}')\n feat_posentropy = op.join(out_dir, 'feat_posentropy.txt')\n entropy_cmd = 'compute_locus_entropy.rb {counts_reads} > {feat_posentropy}'\n with utils.chdir(out_dir):\n run(compute_genomic_cmd.format(min_len=min_trimmed_read_len,\n max_len=max_trimmed_read_len, **locals()), 'Run compute_genomic')\n run(index_genomic_cmd.format(lenvec=lenvec_plus), 'Run index in plus')\n run(index_genomic_cmd.format(lenvec=lenvec_minus), 'Run index in minus'\n )\n run(compute_locus_cmd.format(min_len=min_trimmed_read_len, max_len=\n max_trimmed_read_len, **locals()), 'Run compute locus')\n run(coverage_anti_cmd.format(**locals()), 'Run coverage antisense')\n feat_antisense = _order_antisense_column(cov_S_file,\n min_trimmed_read_len)\n counts_reads = _reads_per_position(bam_in, loci_file, out_dir)\n run(entropy_cmd.format(**locals()), 'Run entropy')\n rnafold = calculate_structure(loci_file, reference)\n", "role": "user" }, { "content": "create_features", "role": "assistant" } ]
def x(args): logger.info('reading sequeces') data = load_data(args.json) logger.info('create profile') data = make_profile(data, os.path.join(args.out, 'profiles'), args) logger.info('create database') make_database(data, 'seqcluster.db', args.out) logger.info( 'Done. Download https://github.com/lpantano/seqclusterViz/archive/master.zip to browse the output.' )
report
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(args):\n logger.info('reading sequeces')\n data = load_data(args.json)\n logger.info('create profile')\n data = make_profile(data, os.path.join(args.out, 'profiles'), args)\n logger.info('create database')\n make_database(data, 'seqcluster.db', args.out)\n logger.info(\n 'Done. Download https://github.com/lpantano/seqclusterViz/archive/master.zip to browse the output.'\n )\n", "role": "user" }, { "content": "report", "role": "assistant" } ]
def x(peaks): previous = peaks[0] new_peaks = [previous] for pos in peaks: if pos > previous + 10: new_peaks.add(pos) previous = pos return new_peaks
summarize_peaks
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(peaks):\n previous = peaks[0]\n new_peaks = [previous]\n for pos in peaks:\n if pos > previous + 10:\n new_peaks.add(pos)\n previous = pos\n return new_peaks\n", "role": "user" }, { "content": "summarize_peaks", "role": "assistant" } ]
def x(x, y, win=10): previous = min(y) peaks = [] intervals = range(x, y, win) for pos in intervals: if y[pos] > previous * 10: previous = y[pos] peaks.add(pos) peaks = _summarize_peaks(peaks)
find_mature
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(x, y, win=10):\n previous = min(y)\n peaks = []\n intervals = range(x, y, win)\n for pos in intervals:\n if y[pos] > previous * 10:\n previous = y[pos]\n peaks.add(pos)\n peaks = _summarize_peaks(peaks)\n", "role": "user" }, { "content": "find_mature", "role": "assistant" } ]
def x(in_file): keep = Counter() with open_fastq(in_file) as handle: for line in handle: if line.startswith('@'): if line.find('UMI') > -1: logger.info( 'Find UMI tags in read names, collapsing by UMI.') return collapse_umi(in_file) seq = handle.next().strip() handle.next() qual = handle.next().strip() if seq in keep: keep[seq].update(qual) else: keep[seq] = quality(qual) logger.info('Sequences loaded: %s' % len(keep)) return keep
collapse
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(in_file):\n keep = Counter()\n with open_fastq(in_file) as handle:\n for line in handle:\n if line.startswith('@'):\n if line.find('UMI') > -1:\n logger.info(\n 'Find UMI tags in read names, collapsing by UMI.')\n return collapse_umi(in_file)\n seq = handle.next().strip()\n handle.next()\n qual = handle.next().strip()\n if seq in keep:\n keep[seq].update(qual)\n else:\n keep[seq] = quality(qual)\n logger.info('Sequences loaded: %s' % len(keep))\n return keep\n", "role": "user" }, { "content": "collapse", "role": "assistant" } ]
def x(in_file): keep = defaultdict(dict) with open_fastq(in_file) as handle: for line in handle: if line.startswith('@'): m = re.search('UMI_([ATGC]*)', line.strip()) umis = m.group(0) seq = handle.next().strip() handle.next() qual = handle.next().strip() if (umis, seq) in keep: keep[umis, seq][1].update(qual) keep[umis, seq][0].update(seq) else: keep[umis, seq] = [umi(seq), quality(qual)] logger.info('Sequences loaded: %s' % len(keep)) return keep
collapse_umi
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(in_file):\n keep = defaultdict(dict)\n with open_fastq(in_file) as handle:\n for line in handle:\n if line.startswith('@'):\n m = re.search('UMI_([ATGC]*)', line.strip())\n umis = m.group(0)\n seq = handle.next().strip()\n handle.next()\n qual = handle.next().strip()\n if (umis, seq) in keep:\n keep[umis, seq][1].update(qual)\n keep[umis, seq][0].update(seq)\n else:\n keep[umis, seq] = [umi(seq), quality(qual)]\n logger.info('Sequences loaded: %s' % len(keep))\n return keep\n", "role": "user" }, { "content": "collapse_umi", "role": "assistant" } ]
def x(in_file): _, ext = os.path.splitext(in_file) if ext == '.gz': return gzip.open(in_file, 'rb') if ext in ['.fastq', '.fq', '.fasta', '.fa']: return open(in_file, 'r') return ValueError('File needs to be fastq|fasta|fq|fa [.gz]')
open_fastq
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(in_file):\n _, ext = os.path.splitext(in_file)\n if ext == '.gz':\n return gzip.open(in_file, 'rb')\n if ext in ['.fastq', '.fq', '.fasta', '.fa']:\n return open(in_file, 'r')\n return ValueError('File needs to be fastq|fasta|fq|fa [.gz]')\n", "role": "user" }, { "content": "open_fastq", "role": "assistant" } ]
def x(args): try: umi_fn = args.fastq if _is_umi(args.fastq): umis = collapse(args.fastq) umi_fn = os.path.join(args.out, splitext_plus(os.path.basename( args.fastq))[0] + '_umi_trimmed.fastq') write_output(umi_fn, umis, args.minimum) seqs = collapse(umi_fn) out_file = splitext_plus(os.path.basename(args.fastq))[0 ] + '_trimmed.fastq' except IOError as e: logger.error('I/O error({0}): {1}'.format(e.errno, e.strerror)) raise 'Can not read file' out_file = os.path.join(args.out, out_file) write_output(out_file, seqs, args.minimum) return out_file
collapse_fastq
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(args):\n try:\n umi_fn = args.fastq\n if _is_umi(args.fastq):\n umis = collapse(args.fastq)\n umi_fn = os.path.join(args.out, splitext_plus(os.path.basename(\n args.fastq))[0] + '_umi_trimmed.fastq')\n write_output(umi_fn, umis, args.minimum)\n seqs = collapse(umi_fn)\n out_file = splitext_plus(os.path.basename(args.fastq))[0\n ] + '_trimmed.fastq'\n except IOError as e:\n logger.error('I/O error({0}): {1}'.format(e.errno, e.strerror))\n raise 'Can not read file'\n out_file = os.path.join(args.out, out_file)\n write_output(out_file, seqs, args.minimum)\n return out_file\n", "role": "user" }, { "content": "collapse_fastq", "role": "assistant" } ]
def x(doctree): current = 0 num_children = len(doctree.children) while current < num_children: child = doctree.children[current] child.replace_self(child.traverse(no_autoslides_filter)) if len(doctree.children) == num_children: current += 1 else: num_children = len(doctree.children)
filter_doctree_for_slides
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(doctree):\n current = 0\n num_children = len(doctree.children)\n while current < num_children:\n child = doctree.children[current]\n child.replace_self(child.traverse(no_autoslides_filter))\n if len(doctree.children) == num_children:\n current += 1\n else:\n num_children = len(doctree.children)\n", "role": "user" }, { "content": "filter_doctree_for_slides", "role": "assistant" } ]
def x(self, node, increment=True): parent_title_node = node.parent.next_node(nodes.title) nextslide_info = getattr(parent_title_node, 'nextslide_info', ( parent_title_node.deepcopy().children, 1)) nextslide_info = nextslide_info[0], nextslide_info[1] + 1 if node.args: textnodes, messages = node.state.inline_text(node.args[0], 1) new_title = nodes.title(node.args[0], '', *textnodes) else: title_nodes = nextslide_info[0][:] if 'increment' in node.attributes: title_nodes.append(nodes.Text(' (%s)' % nextslide_info[1])) new_title = nodes.title('', '', *title_nodes) new_title.nextslide_info = nextslide_info return new_title
make_title_node
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, node, increment=True):\n parent_title_node = node.parent.next_node(nodes.title)\n nextslide_info = getattr(parent_title_node, 'nextslide_info', (\n parent_title_node.deepcopy().children, 1))\n nextslide_info = nextslide_info[0], nextslide_info[1] + 1\n if node.args:\n textnodes, messages = node.state.inline_text(node.args[0], 1)\n new_title = nodes.title(node.args[0], '', *textnodes)\n else:\n title_nodes = nextslide_info[0][:]\n if 'increment' in node.attributes:\n title_nodes.append(nodes.Text(' (%s)' % nextslide_info[1]))\n new_title = nodes.title('', '', *title_nodes)\n new_title.nextslide_info = nextslide_info\n return new_title\n", "role": "user" }, { "content": "make_title_node", "role": "assistant" } ]
def x(self, builder): if 'theme' in self.attributes: builder.apply_theme(self.attributes['theme'], builder.theme_options)
apply
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, builder):\n if 'theme' in self.attributes:\n builder.apply_theme(self.attributes['theme'], builder.theme_options)\n", "role": "user" }, { "content": "apply", "role": "assistant" } ]
def x(cls, builder, doctree=None): result = {'theme': builder.config.slide_theme, 'autoslides': builder. config.autoslides, 'slide_classes': []} if doctree: conf_node = cls.get(doctree) if conf_node: result.update(conf_node.attributes) return result
get_conf
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, builder, doctree=None):\n result = {'theme': builder.config.slide_theme, 'autoslides': builder.\n config.autoslides, 'slide_classes': []}\n if doctree:\n conf_node = cls.get(doctree)\n if conf_node:\n result.update(conf_node.attributes)\n return result\n", "role": "user" }, { "content": "get_conf", "role": "assistant" } ]
def x(context): COPY_LISTS = 'script_files', 'css_files' for attr in COPY_LISTS: if attr in context: context[attr] = context[attr][:] return context
fix_context
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(context):\n COPY_LISTS = 'script_files', 'css_files'\n for attr in COPY_LISTS:\n if attr in context:\n context[attr] = context[attr][:]\n return context\n", "role": "user" }, { "content": "fix_context", "role": "assistant" } ]
def x(d): msg = bold('Welcome to the Hieroglyph %s quickstart utility.') % (version (),) print(msg) msg = """ This will ask questions for creating a Hieroglyph project, and then ask some basic Sphinx questions. """ print(msg) d.update({'version': datetime.date.today().strftime('%Y.%m.%d'), 'release': datetime.date.today().strftime('%Y.%m.%d'), 'make_mode': True}) if 'project' not in d: print('\nThe presentation title will be included on the title slide.') sphinx.quickstart.do_prompt(d, 'project', 'Presentation title') if 'author' not in d: sphinx.quickstart.do_prompt(d, 'author', 'Author name(s)') theme_entrypoints = pkg_resources.iter_entry_points('hieroglyph.theme') themes = [t.load() for t in theme_entrypoints] msg = '\nAvailable themes:\n\n' for theme in themes: msg += '\n'.join([bold(theme['name']), theme['desc'], '', '']) msg += 'Which theme would you like to use?' print(msg) sphinx.quickstart.do_prompt(d, 'slide_theme', 'Slide Theme', themes[0][ 'name'], sphinx.quickstart.choice(*[t['name'] for t in themes])) print('') sphinx.quickstart.ask_user(d)
ask_user
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(d):\n msg = bold('Welcome to the Hieroglyph %s quickstart utility.') % (version\n (),)\n print(msg)\n msg = \"\"\"\nThis will ask questions for creating a Hieroglyph project, and then ask\nsome basic Sphinx questions.\n\"\"\"\n print(msg)\n d.update({'version': datetime.date.today().strftime('%Y.%m.%d'),\n 'release': datetime.date.today().strftime('%Y.%m.%d'), 'make_mode':\n True})\n if 'project' not in d:\n print('\\nThe presentation title will be included on the title slide.')\n sphinx.quickstart.do_prompt(d, 'project', 'Presentation title')\n if 'author' not in d:\n sphinx.quickstart.do_prompt(d, 'author', 'Author name(s)')\n theme_entrypoints = pkg_resources.iter_entry_points('hieroglyph.theme')\n themes = [t.load() for t in theme_entrypoints]\n msg = '\\nAvailable themes:\\n\\n'\n for theme in themes:\n msg += '\\n'.join([bold(theme['name']), theme['desc'], '', ''])\n msg += 'Which theme would you like to use?'\n print(msg)\n sphinx.quickstart.do_prompt(d, 'slide_theme', 'Slide Theme', themes[0][\n 'name'], sphinx.quickstart.choice(*[t['name'] for t in themes]))\n print('')\n sphinx.quickstart.ask_user(d)\n", "role": "user" }, { "content": "ask_user", "role": "assistant" } ]
def x(self): return {'title': self.title, 'level': self.level, 'content': self. content, 'classes': self.classes, 'slide_classes': self. _filter_classes(exclude='content-'), 'content_classes': self. _filter_classes(include='content-'), 'slide_number': self. slide_number, 'config': self._translator.builder.config, 'id': self.id}
get_slide_context
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return {'title': self.title, 'level': self.level, 'content': self.\n content, 'classes': self.classes, 'slide_classes': self.\n _filter_classes(exclude='content-'), 'content_classes': self.\n _filter_classes(include='content-'), 'slide_number': self.\n slide_number, 'config': self._translator.builder.config, 'id': self.id}\n", "role": "user" }, { "content": "get_slide_context", "role": "assistant" } ]
def x(self, slide_no): if self.builder.config.slide_numbers: self.body.append('\n<div class="slide-no">%s</div>\n' % (slide_no,))
add_slide_number
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, slide_no):\n if self.builder.config.slide_numbers:\n self.body.append('\\n<div class=\"slide-no\">%s</div>\\n' % (slide_no,))\n", "role": "user" }, { "content": "add_slide_number", "role": "assistant" } ]
def x(self, slide_no): if self.builder.config.slide_footer: self.body.append('\n<div class="slide-footer">%s</div>\n' % (self. builder.config.slide_footer,))
add_slide_footer
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, slide_no):\n if self.builder.config.slide_footer:\n self.body.append('\\n<div class=\"slide-footer\">%s</div>\\n' % (self.\n builder.config.slide_footer,))\n", "role": "user" }, { "content": "add_slide_footer", "role": "assistant" } ]
def x(app): from hieroglyph import writer if app.builder.name not in HTML_BUILDERS: return if app.config.slide_link_html_to_slides: app.builder.templates.loaders.append(SphinxFileSystemLoader(os.path .join(os.path.dirname(__file__), 'themes', 'slides'))) if not app.config.html_sidebars: app.config.html_sidebars = {'**': ['localtoc.html', 'relations.html', 'sourcelink.html', SLIDELINK_TEMPLATE, 'searchbox.html']} else: included = False for glob, templates in app.config.html_sidebars: if SLIDELINK_TEMPLATE in templates: included = True break if not included: app.config.html_sidebars.setdefault('**', []).append( SLIDELINK_TEMPLATE) if app.config.slide_link_html_sections_to_slides: if sphinx.version_info >= (1, 6, 0): override_translator = type('SlideLinkTranslator', (app.builder. get_translator_class(), object), {'depart_title': writer. depart_title}) app.set_translator(app.builder, override_translator) else: app.builder.translator_class = type('SlideLinkTranslator', (app .builder.translator_class, object), {'depart_title': writer .depart_title})
inspect_config
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(app):\n from hieroglyph import writer\n if app.builder.name not in HTML_BUILDERS:\n return\n if app.config.slide_link_html_to_slides:\n app.builder.templates.loaders.append(SphinxFileSystemLoader(os.path\n .join(os.path.dirname(__file__), 'themes', 'slides')))\n if not app.config.html_sidebars:\n app.config.html_sidebars = {'**': ['localtoc.html',\n 'relations.html', 'sourcelink.html', SLIDELINK_TEMPLATE,\n 'searchbox.html']}\n else:\n included = False\n for glob, templates in app.config.html_sidebars:\n if SLIDELINK_TEMPLATE in templates:\n included = True\n break\n if not included:\n app.config.html_sidebars.setdefault('**', []).append(\n SLIDELINK_TEMPLATE)\n if app.config.slide_link_html_sections_to_slides:\n if sphinx.version_info >= (1, 6, 0):\n override_translator = type('SlideLinkTranslator', (app.builder.\n get_translator_class(), object), {'depart_title': writer.\n depart_title})\n app.set_translator(app.builder, override_translator)\n else:\n app.builder.translator_class = type('SlideLinkTranslator', (app\n .builder.translator_class, object), {'depart_title': writer\n .depart_title})\n", "role": "user" }, { "content": "inspect_config", "role": "assistant" } ]
def x(builder, pagename=None): return builder.get_relative_uri(pagename or builder.current_docname, os .path.join(builder.app.config.slide_relative_path, pagename or builder.current_docname))
slide_path
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(builder, pagename=None):\n return builder.get_relative_uri(pagename or builder.current_docname, os\n .path.join(builder.app.config.slide_relative_path, pagename or\n builder.current_docname))\n", "role": "user" }, { "content": "slide_path", "role": "assistant" } ]
def x(builder, pagename=None): return builder.get_relative_uri(pagename or builder.current_docname, os .path.join(builder.app.config.slide_html_relative_path, pagename or builder.current_docname))
html_path
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(builder, pagename=None):\n return builder.get_relative_uri(pagename or builder.current_docname, os\n .path.join(builder.app.config.slide_html_relative_path, pagename or\n builder.current_docname))\n", "role": "user" }, { "content": "html_path", "role": "assistant" } ]
def x(app, pagename, templatename, context, doctree): context['show_slidelink' ] = app.config.slide_link_html_to_slides and hasattr(app.builder, 'get_outfilename') if context['show_slidelink']: context['slide_path'] = slide_path(app.builder, pagename)
add_link
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(app, pagename, templatename, context, doctree):\n context['show_slidelink'\n ] = app.config.slide_link_html_to_slides and hasattr(app.builder,\n 'get_outfilename')\n if context['show_slidelink']:\n context['slide_path'] = slide_path(app.builder, pagename)\n", "role": "user" }, { "content": "add_link", "role": "assistant" } ]
def x(self, themename, themeoptions): self._theme_stack.append((self.theme, self.theme_options)) theme_factory = HTMLThemeFactory(self.app) theme_factory.load_additional_themes(self.get_builtin_theme_dirs() + self.config.slide_theme_path) self.theme = theme_factory.create(themename) self.theme_options = themeoptions.copy() self.templates.init(self, self.theme) self.templates.environment.filters['json'] = json.dumps if self.theme not in self._additional_themes: self._additional_themes.append(self.theme)
apply_theme
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, themename, themeoptions):\n self._theme_stack.append((self.theme, self.theme_options))\n theme_factory = HTMLThemeFactory(self.app)\n theme_factory.load_additional_themes(self.get_builtin_theme_dirs() +\n self.config.slide_theme_path)\n self.theme = theme_factory.create(themename)\n self.theme_options = themeoptions.copy()\n self.templates.init(self, self.theme)\n self.templates.environment.filters['json'] = json.dumps\n if self.theme not in self._additional_themes:\n self._additional_themes.append(self.theme)\n", "role": "user" }, { "content": "apply_theme", "role": "assistant" } ]
def x(self, doctree): super(AbstractSlideBuilder, self).post_process_images(doctree) relative_base = ['..'] * doctree.attributes.get('source')[len(self. srcdir) + 1:].count('/') for node in doctree.traverse(nodes.image): if node.get('candidates') is None: node['candidates'] = '*', if node['uri'].startswith(self.outdir): node['uri'] = '/'.join(relative_base + [node['uri'][len(self. outdir) + 1:]])
post_process_images
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, doctree):\n super(AbstractSlideBuilder, self).post_process_images(doctree)\n relative_base = ['..'] * doctree.attributes.get('source')[len(self.\n srcdir) + 1:].count('/')\n for node in doctree.traverse(nodes.image):\n if node.get('candidates') is None:\n node['candidates'] = '*',\n if node['uri'].startswith(self.outdir):\n node['uri'] = '/'.join(relative_base + [node['uri'][len(self.\n outdir) + 1:]])\n", "role": "user" }, { "content": "post_process_images", "role": "assistant" } ]
def x(section): metadata = {} metadata_lines = section.split('\n') for line in metadata_lines: colon_index = line.find(':') if colon_index != -1: key = line[:colon_index].strip() val = line[colon_index + 1:].strip() metadata[key] = val return metadata
parse_metadata
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(section):\n metadata = {}\n metadata_lines = section.split('\\n')\n for line in metadata_lines:\n colon_index = line.find(':')\n if colon_index != -1:\n key = line[:colon_index].strip()\n val = line[colon_index + 1:].strip()\n metadata[key] = val\n return metadata\n", "role": "user" }, { "content": "parse_metadata", "role": "assistant" } ]
def x(html, metadata): if metadata.get('build_lists') and metadata['build_lists'] == 'true': html = html.replace('<ul>', '<ul class="build">') html = html.replace('<ol>', '<ol class="build">') return html
postprocess_html
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(html, metadata):\n if metadata.get('build_lists') and metadata['build_lists'] == 'true':\n html = html.replace('<ul>', '<ul class=\"build\">')\n html = html.replace('<ol>', '<ol class=\"build\">')\n return html\n", "role": "user" }, { "content": "postprocess_html", "role": "assistant" } ]
def x(linter, checker_method, augmentation): if sys.version_info[0] <= 2: checker = get_checker(linter, checker_method.im_class) else: try: checker = get_checker(linter, checker_method.__self__.__class__) except AttributeError: checker = get_checker(linter, get_class(checker_method. __module__, checker_method.__qualname__)) old_method = getattr(checker, checker_method.__name__) def x(node): def x(): old_method(node) augmentation(chain, node) setattr(checker, checker_method.__name__, augment_func)
augment_visit
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(linter, checker_method, augmentation):\n if sys.version_info[0] <= 2:\n checker = get_checker(linter, checker_method.im_class)\n else:\n try:\n checker = get_checker(linter, checker_method.__self__.__class__)\n except AttributeError:\n checker = get_checker(linter, get_class(checker_method.\n __module__, checker_method.__qualname__))\n old_method = getattr(checker, checker_method.__name__)\n\n def x(node):\n\n def x():\n old_method(node)\n augmentation(chain, node)\n setattr(checker, checker_method.__name__, augment_func)\n", "role": "user" }, { "content": "augment_visit", "role": "assistant" } ]
def x(linter, checker_method, message_id_or_symbol, test_func): msgs_store = getattr(linter, 'msgs_store', linter) def x(message_id_or_symbol): if hasattr(msgs_store, 'check_message_id'): return [msgs_store.check_message_id(message_id_or_symbol)] elif hasattr(msgs_store, 'get_message_definition'): return [msgs_store.get_message_definition(message_id_or_symbol)] elif hasattr(msgs_store, 'get_message_definitions'): return msgs_store.get_message_definitions(message_id_or_symbol) else: raise ValueError( 'pylint.utils.MessagesStore does not have a get_message_definition(s) method' ) try: pylint_messages = get_message_definitions(message_id_or_symbol) symbols = [symbol for pylint_message in pylint_messages for symbol in (pylint_message.msgid, pylint_message.symbol) if symbol is not None ] except UnknownMessage: symbols = [message_id_or_symbol] def x(chain, node): with Suppress(linter) as s: if test_func(node): s.suppress(*symbols) chain() augment_visit(linter, checker_method, do_suppress)
suppress_message
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(linter, checker_method, message_id_or_symbol, test_func):\n msgs_store = getattr(linter, 'msgs_store', linter)\n\n def x(message_id_or_symbol):\n if hasattr(msgs_store, 'check_message_id'):\n return [msgs_store.check_message_id(message_id_or_symbol)]\n elif hasattr(msgs_store, 'get_message_definition'):\n return [msgs_store.get_message_definition(message_id_or_symbol)]\n elif hasattr(msgs_store, 'get_message_definitions'):\n return msgs_store.get_message_definitions(message_id_or_symbol)\n else:\n raise ValueError(\n 'pylint.utils.MessagesStore does not have a get_message_definition(s) method'\n )\n try:\n pylint_messages = get_message_definitions(message_id_or_symbol)\n symbols = [symbol for pylint_message in pylint_messages for symbol in\n (pylint_message.msgid, pylint_message.symbol) if symbol is not None\n ]\n except UnknownMessage:\n symbols = [message_id_or_symbol]\n\n def x(chain, node):\n with Suppress(linter) as s:\n if test_func(node):\n s.suppress(*symbols)\n chain()\n augment_visit(linter, checker_method, do_suppress)\n", "role": "user" }, { "content": "suppress_message", "role": "assistant" } ]
def x(self, hash_values): q = """SELECT threat_type,platform_type,threat_entry_type, expires_at < current_timestamp AS has_expired FROM full_hash WHERE value IN ({}) """ output = [] with self.get_cursor() as dbc: placeholders = ','.join(['?'] * len(hash_values)) dbc.execute(q.format(placeholders), [sqlite3.Binary(hv) for hv in hash_values]) for h in dbc.fetchall(): threat_type, platform_type, threat_entry_type, has_expired = h threat_list = ThreatList(threat_type, platform_type, threat_entry_type) output.append((threat_list, has_expired)) return output
lookup_full_hashes
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, hash_values):\n q = \"\"\"SELECT threat_type,platform_type,threat_entry_type, expires_at < current_timestamp AS has_expired\n FROM full_hash WHERE value IN ({})\n \"\"\"\n output = []\n with self.get_cursor() as dbc:\n placeholders = ','.join(['?'] * len(hash_values))\n dbc.execute(q.format(placeholders), [sqlite3.Binary(hv) for hv in\n hash_values])\n for h in dbc.fetchall():\n threat_type, platform_type, threat_entry_type, has_expired = h\n threat_list = ThreatList(threat_type, platform_type,\n threat_entry_type)\n output.append((threat_list, has_expired))\n return output\n", "role": "user" }, { "content": "lookup_full_hashes", "role": "assistant" } ]
def x(self, cues): q = """SELECT value, MAX(negative_expires_at < current_timestamp) AS negative_cache_expired FROM hash_prefix WHERE cue IN ({}) GROUP BY 1 """ output = [] with self.get_cursor() as dbc: dbc.execute(q.format(','.join(['?'] * len(cues))), [sqlite3.Binary( cue) for cue in cues]) for h in dbc.fetchall(): value, negative_cache_expired = h output.append((bytes(value), negative_cache_expired)) return output
lookup_hash_prefix
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, cues):\n q = \"\"\"SELECT value, MAX(negative_expires_at < current_timestamp) AS negative_cache_expired\n FROM hash_prefix WHERE cue IN ({}) GROUP BY 1\n \"\"\"\n output = []\n with self.get_cursor() as dbc:\n dbc.execute(q.format(','.join(['?'] * len(cues))), [sqlite3.Binary(\n cue) for cue in cues])\n for h in dbc.fetchall():\n value, negative_cache_expired = h\n output.append((bytes(value), negative_cache_expired))\n return output\n", "role": "user" }, { "content": "lookup_hash_prefix", "role": "assistant" } ]
def x(self, threat_list, hash_value, cache_duration, malware_threat_type): log.info('Storing full hash %s to list %s with cache duration %s', to_hex(hash_value), str(threat_list), cache_duration) qi = """INSERT OR IGNORE INTO full_hash (value, threat_type, platform_type, threat_entry_type, malware_threat_type, downloaded_at) VALUES (?, ?, ?, ?, ?, current_timestamp) """ qu = ( "UPDATE full_hash SET expires_at=datetime(current_timestamp, '+{} SECONDS') WHERE value=? AND threat_type=? AND platform_type=? AND threat_entry_type=?" ) i_parameters = [sqlite3.Binary(hash_value), threat_list.threat_type, threat_list.platform_type, threat_list.threat_entry_type, malware_threat_type] u_parameters = [sqlite3.Binary(hash_value), threat_list.threat_type, threat_list.platform_type, threat_list.threat_entry_type] with self.get_cursor() as dbc: dbc.execute(qi, i_parameters) dbc.execute(qu.format(int(cache_duration)), u_parameters)
store_full_hash
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, threat_list, hash_value, cache_duration, malware_threat_type):\n log.info('Storing full hash %s to list %s with cache duration %s',\n to_hex(hash_value), str(threat_list), cache_duration)\n qi = \"\"\"INSERT OR IGNORE INTO full_hash\n (value, threat_type, platform_type, threat_entry_type, malware_threat_type, downloaded_at)\n VALUES\n (?, ?, ?, ?, ?, current_timestamp)\n \"\"\"\n qu = (\n \"UPDATE full_hash SET expires_at=datetime(current_timestamp, '+{} SECONDS') WHERE value=? AND threat_type=? AND platform_type=? AND threat_entry_type=?\"\n )\n i_parameters = [sqlite3.Binary(hash_value), threat_list.threat_type,\n threat_list.platform_type, threat_list.threat_entry_type,\n malware_threat_type]\n u_parameters = [sqlite3.Binary(hash_value), threat_list.threat_type,\n threat_list.platform_type, threat_list.threat_entry_type]\n with self.get_cursor() as dbc:\n dbc.execute(qi, i_parameters)\n dbc.execute(qu.format(int(cache_duration)), u_parameters)\n", "role": "user" }, { "content": "store_full_hash", "role": "assistant" } ]
def x(self, keep_expired_for=60 * 60 * 12): q = """DELETE FROM full_hash WHERE expires_at < datetime(current_timestamp, '-{} SECONDS') """ log.info('Cleaning up full_hash entries expired more than {} seconds ago.' .format(keep_expired_for)) with self.get_cursor() as dbc: dbc.execute(q.format(int(keep_expired_for)))
cleanup_full_hashes
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, keep_expired_for=60 * 60 * 12):\n q = \"\"\"DELETE FROM full_hash WHERE expires_at < datetime(current_timestamp, '-{} SECONDS')\n \"\"\"\n log.info('Cleaning up full_hash entries expired more than {} seconds ago.'\n .format(keep_expired_for))\n with self.get_cursor() as dbc:\n dbc.execute(q.format(int(keep_expired_for)))\n", "role": "user" }, { "content": "cleanup_full_hashes", "role": "assistant" } ]
def x(self): q = 'SELECT threat_type,platform_type,threat_entry_type FROM threat_list' output = [] with self.get_cursor() as dbc: dbc.execute(q) for h in dbc.fetchall(): threat_type, platform_type, threat_entry_type = h threat_list = ThreatList(threat_type, platform_type, threat_entry_type) output.append(threat_list) return output
get_threat_lists
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n q = 'SELECT threat_type,platform_type,threat_entry_type FROM threat_list'\n output = []\n with self.get_cursor() as dbc:\n dbc.execute(q)\n for h in dbc.fetchall():\n threat_type, platform_type, threat_entry_type = h\n threat_list = ThreatList(threat_type, platform_type,\n threat_entry_type)\n output.append(threat_list)\n return output\n", "role": "user" }, { "content": "get_threat_lists", "role": "assistant" } ]
def x(self): q = ( 'SELECT threat_type,platform_type,threat_entry_type,client_state FROM threat_list' ) output = {} with self.get_cursor() as dbc: dbc.execute(q) for h in dbc.fetchall(): threat_type, platform_type, threat_entry_type, client_state = h threat_list_tuple = threat_type, platform_type, threat_entry_type output[threat_list_tuple] = client_state return output
get_client_state
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n q = (\n 'SELECT threat_type,platform_type,threat_entry_type,client_state FROM threat_list'\n )\n output = {}\n with self.get_cursor() as dbc:\n dbc.execute(q)\n for h in dbc.fetchall():\n threat_type, platform_type, threat_entry_type, client_state = h\n threat_list_tuple = threat_type, platform_type, threat_entry_type\n output[threat_list_tuple] = client_state\n return output\n", "role": "user" }, { "content": "get_client_state", "role": "assistant" } ]
def x(self, threat_list): q = """INSERT OR IGNORE INTO threat_list (threat_type, platform_type, threat_entry_type, timestamp) VALUES (?, ?, ?, current_timestamp) """ params = [threat_list.threat_type, threat_list.platform_type, threat_list.threat_entry_type] with self.get_cursor() as dbc: dbc.execute(q, params)
add_threat_list
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, threat_list):\n q = \"\"\"INSERT OR IGNORE INTO threat_list\n (threat_type, platform_type, threat_entry_type, timestamp)\n VALUES\n (?, ?, ?, current_timestamp)\n \"\"\"\n params = [threat_list.threat_type, threat_list.platform_type,\n threat_list.threat_entry_type]\n with self.get_cursor() as dbc:\n dbc.execute(q, params)\n", "role": "user" }, { "content": "add_threat_list", "role": "assistant" } ]
def x(self, threat_list): log.info('Deleting cached threat list "{}"'.format(repr(threat_list))) q = """DELETE FROM threat_list WHERE threat_type=? AND platform_type=? AND threat_entry_type=? """ params = [threat_list.threat_type, threat_list.platform_type, threat_list.threat_entry_type] with self.get_cursor() as dbc: dbc.execute(q, params)
delete_threat_list
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, threat_list):\n log.info('Deleting cached threat list \"{}\"'.format(repr(threat_list)))\n q = \"\"\"DELETE FROM threat_list\n WHERE threat_type=? AND platform_type=? AND threat_entry_type=?\n \"\"\"\n params = [threat_list.threat_type, threat_list.platform_type,\n threat_list.threat_entry_type]\n with self.get_cursor() as dbc:\n dbc.execute(q, params)\n", "role": "user" }, { "content": "delete_threat_list", "role": "assistant" } ]
def x(self, threat_list): q = """SELECT value FROM hash_prefix WHERE threat_type=? AND platform_type=? AND threat_entry_type=? ORDER BY value """ params = [threat_list.threat_type, threat_list.platform_type, threat_list.threat_entry_type] with self.get_cursor() as dbc: dbc.execute(q, params) all_hashes = b''.join(bytes(h[0]) for h in dbc.fetchall()) checksum = hashlib.sha256(all_hashes).digest() return checksum
hash_prefix_list_checksum
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, threat_list):\n q = \"\"\"SELECT value FROM hash_prefix\n WHERE threat_type=? AND platform_type=? AND threat_entry_type=?\n ORDER BY value\n \"\"\"\n params = [threat_list.threat_type, threat_list.platform_type,\n threat_list.threat_entry_type]\n with self.get_cursor() as dbc:\n dbc.execute(q, params)\n all_hashes = b''.join(bytes(h[0]) for h in dbc.fetchall())\n checksum = hashlib.sha256(all_hashes).digest()\n return checksum\n", "role": "user" }, { "content": "hash_prefix_list_checksum", "role": "assistant" } ]
def x(self, threat_list, indices): batch_size = 40 q = """DELETE FROM hash_prefix WHERE threat_type=? AND platform_type=? AND threat_entry_type=? AND value IN ({}) """ prefixes_to_remove = self.get_hash_prefix_values_to_remove(threat_list, indices) with self.get_cursor() as dbc: for i in range(0, len(prefixes_to_remove), batch_size): remove_batch = prefixes_to_remove[i:i + batch_size] params = [threat_list.threat_type, threat_list.platform_type, threat_list.threat_entry_type] + [sqlite3.Binary(b) for b in remove_batch] dbc.execute(q.format(','.join(['?'] * len(remove_batch))), params)
remove_hash_prefix_indices
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, threat_list, indices):\n batch_size = 40\n q = \"\"\"DELETE FROM hash_prefix\n WHERE threat_type=? AND platform_type=? AND threat_entry_type=? AND value IN ({})\n \"\"\"\n prefixes_to_remove = self.get_hash_prefix_values_to_remove(threat_list,\n indices)\n with self.get_cursor() as dbc:\n for i in range(0, len(prefixes_to_remove), batch_size):\n remove_batch = prefixes_to_remove[i:i + batch_size]\n params = [threat_list.threat_type, threat_list.platform_type,\n threat_list.threat_entry_type] + [sqlite3.Binary(b) for b in\n remove_batch]\n dbc.execute(q.format(','.join(['?'] * len(remove_batch))), params)\n", "role": "user" }, { "content": "remove_hash_prefix_indices", "role": "assistant" } ]
def x(self): q = 'SELECT distinct value from hash_prefix' output = [] with self.get_cursor() as dbc: dbc.execute(q) output = [bytes(r[0]) for r in dbc.fetchall()] return output
dump_hash_prefix_values
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n q = 'SELECT distinct value from hash_prefix'\n output = []\n with self.get_cursor() as dbc:\n dbc.execute(q)\n output = [bytes(r[0]) for r in dbc.fetchall()]\n return output\n", "role": "user" }, { "content": "dump_hash_prefix_values", "role": "assistant" } ]
def x(tk): used = False try: while True: try: method, args, kwargs, response_queue = (tk.tk._event_queue. get_nowait()) except queue.Empty: break else: used = True if tk.tk._debug >= 2: print('Calling event from main thread:', method. __name__, args, kwargs) try: response_queue.put((False, method(*args, **kwargs))) except SystemExit: raise except Exception: from sys import exc_info ex_type, ex_value, ex_tb = exc_info() response_queue.put((True, (ex_type, ex_value, ex_tb))) finally: if used: tk.after_idle(_check_events, tk) else: tk.after(tk.tk._check_period, _check_events, tk)
check_events
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(tk):\n used = False\n try:\n while True:\n try:\n method, args, kwargs, response_queue = (tk.tk._event_queue.\n get_nowait())\n except queue.Empty:\n break\n else:\n used = True\n if tk.tk._debug >= 2:\n print('Calling event from main thread:', method.\n __name__, args, kwargs)\n try:\n response_queue.put((False, method(*args, **kwargs)))\n except SystemExit:\n raise\n except Exception:\n from sys import exc_info\n ex_type, ex_value, ex_tb = exc_info()\n response_queue.put((True, (ex_type, ex_value, ex_tb)))\n finally:\n if used:\n tk.after_idle(_check_events, tk)\n else:\n tk.after(tk.tk._check_period, _check_events, tk)\n", "role": "user" }, { "content": "check_events", "role": "assistant" } ]
def x(self): try: self.storage.cleanup_full_hashes() self.storage.commit() self._sync_threat_lists() self.storage.commit() self._sync_hash_prefix_cache() except Exception: self.storage.rollback() raise
update_hash_prefix_cache
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n try:\n self.storage.cleanup_full_hashes()\n self.storage.commit()\n self._sync_threat_lists()\n self.storage.commit()\n self._sync_hash_prefix_cache()\n except Exception:\n self.storage.rollback()\n raise\n", "role": "user" }, { "content": "update_hash_prefix_cache", "role": "assistant" } ]
def x(self, hash_prefixes): client_state = self.storage.get_client_state() self.api_client.fair_use_delay() fh_response = self.api_client.get_full_hashes(hash_prefixes, client_state) for m in fh_response.get('matches', []): threat_list = ThreatList(m['threatType'], m['platformType'], m[ 'threatEntryType']) hash_value = b64decode(m['threat']['hash']) cache_duration = int(m['cacheDuration'].rstrip('s')) malware_threat_type = None for metadata in m['threatEntryMetadata'].get('entries', []): k = b64decode(metadata['key']) v = b64decode(metadata['value']) if k == 'malware_threat_type': malware_threat_type = v self.storage.store_full_hash(threat_list, hash_value, cache_duration, malware_threat_type) negative_cache_duration = int(fh_response['negativeCacheDuration']. rstrip('s')) for prefix_value in hash_prefixes: self.storage.update_hash_prefix_expiration(prefix_value, negative_cache_duration)
sync_full_hashes
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, hash_prefixes):\n client_state = self.storage.get_client_state()\n self.api_client.fair_use_delay()\n fh_response = self.api_client.get_full_hashes(hash_prefixes, client_state)\n for m in fh_response.get('matches', []):\n threat_list = ThreatList(m['threatType'], m['platformType'], m[\n 'threatEntryType'])\n hash_value = b64decode(m['threat']['hash'])\n cache_duration = int(m['cacheDuration'].rstrip('s'))\n malware_threat_type = None\n for metadata in m['threatEntryMetadata'].get('entries', []):\n k = b64decode(metadata['key'])\n v = b64decode(metadata['value'])\n if k == 'malware_threat_type':\n malware_threat_type = v\n self.storage.store_full_hash(threat_list, hash_value,\n cache_duration, malware_threat_type)\n negative_cache_duration = int(fh_response['negativeCacheDuration'].\n rstrip('s'))\n for prefix_value in hash_prefixes:\n self.storage.update_hash_prefix_expiration(prefix_value,\n negative_cache_duration)\n", "role": "user" }, { "content": "sync_full_hashes", "role": "assistant" } ]
def x(self, url): if type(url) is not str: url = url.encode('utf8') if not url.strip(): raise ValueError('Empty input string.') url_hashes = URL(url).hashes try: list_names = self._lookup_hashes(url_hashes) self.storage.commit() except Exception: self.storage.rollback() raise if list_names: return list_names return None
lookup_url
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, url):\n if type(url) is not str:\n url = url.encode('utf8')\n if not url.strip():\n raise ValueError('Empty input string.')\n url_hashes = URL(url).hashes\n try:\n list_names = self._lookup_hashes(url_hashes)\n self.storage.commit()\n except Exception:\n self.storage.rollback()\n raise\n if list_names:\n return list_names\n return None\n", "role": "user" }, { "content": "lookup_url", "role": "assistant" } ]
def x(self, full_hashes): full_hashes = list(full_hashes) cues = [fh[0:4] for fh in full_hashes] result = [] matching_prefixes = {} matching_full_hashes = set() is_potential_threat = False for hash_prefix, negative_cache_expired in self.storage.lookup_hash_prefix( cues): for full_hash in full_hashes: if full_hash.startswith(hash_prefix): is_potential_threat = True matching_prefixes[hash_prefix] = matching_prefixes.get( hash_prefix, False) or negative_cache_expired matching_full_hashes.add(full_hash) if not is_potential_threat: return [] matching_expired_threat_lists = set() for threat_list, has_expired in self.storage.lookup_full_hashes( matching_full_hashes): if has_expired: matching_expired_threat_lists.add(threat_list) else: result.append(threat_list) if result: return result if len(matching_expired_threat_lists) == 0 and sum(map(int, matching_prefixes.values())) == 0: log.info('Negative cache hit.') return [] self._sync_full_hashes(matching_prefixes.keys()) for threat_list, has_expired in self.storage.lookup_full_hashes( matching_full_hashes): if not has_expired: result.append(threat_list) return result
lookup_hashes
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, full_hashes):\n full_hashes = list(full_hashes)\n cues = [fh[0:4] for fh in full_hashes]\n result = []\n matching_prefixes = {}\n matching_full_hashes = set()\n is_potential_threat = False\n for hash_prefix, negative_cache_expired in self.storage.lookup_hash_prefix(\n cues):\n for full_hash in full_hashes:\n if full_hash.startswith(hash_prefix):\n is_potential_threat = True\n matching_prefixes[hash_prefix] = matching_prefixes.get(\n hash_prefix, False) or negative_cache_expired\n matching_full_hashes.add(full_hash)\n if not is_potential_threat:\n return []\n matching_expired_threat_lists = set()\n for threat_list, has_expired in self.storage.lookup_full_hashes(\n matching_full_hashes):\n if has_expired:\n matching_expired_threat_lists.add(threat_list)\n else:\n result.append(threat_list)\n if result:\n return result\n if len(matching_expired_threat_lists) == 0 and sum(map(int,\n matching_prefixes.values())) == 0:\n log.info('Negative cache hit.')\n return []\n self._sync_full_hashes(matching_prefixes.keys())\n for threat_list, has_expired in self.storage.lookup_full_hashes(\n matching_full_hashes):\n if not has_expired:\n result.append(threat_list)\n return result\n", "role": "user" }, { "content": "lookup_hashes", "role": "assistant" } ]
def x(self): response = self.service.threatLists().list().execute() self.set_wait_duration(response.get('minimumWaitDuration')) return response['threatLists']
get_threats_lists
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n response = self.service.threatLists().list().execute()\n self.set_wait_duration(response.get('minimumWaitDuration'))\n return response['threatLists']\n", "role": "user" }, { "content": "get_threats_lists", "role": "assistant" } ]
def x(self, client_state): request_body = {'client': {'clientId': self.client_id, 'clientVersion': self.client_version}, 'listUpdateRequests': []} for (threat_type, platform_type, threat_entry_type ), current_state in client_state.items(): request_body['listUpdateRequests'].append({'threatType': threat_type, 'platformType': platform_type, 'threatEntryType': threat_entry_type, 'state': current_state, 'constraints': { 'supportedCompressions': ['RAW']}}) response = self.service.threatListUpdates().fetch(body=request_body ).execute() self.set_wait_duration(response.get('minimumWaitDuration')) return response['listUpdateResponses']
get_threats_update
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, client_state):\n request_body = {'client': {'clientId': self.client_id, 'clientVersion':\n self.client_version}, 'listUpdateRequests': []}\n for (threat_type, platform_type, threat_entry_type\n ), current_state in client_state.items():\n request_body['listUpdateRequests'].append({'threatType':\n threat_type, 'platformType': platform_type, 'threatEntryType':\n threat_entry_type, 'state': current_state, 'constraints': {\n 'supportedCompressions': ['RAW']}})\n response = self.service.threatListUpdates().fetch(body=request_body\n ).execute()\n self.set_wait_duration(response.get('minimumWaitDuration'))\n return response['listUpdateResponses']\n", "role": "user" }, { "content": "get_threats_update", "role": "assistant" } ]
def x(self, prefixes, client_state): request_body = {'client': {'clientId': self.client_id, 'clientVersion': self.client_version}, 'clientStates': [], 'threatInfo': { 'threatTypes': [], 'platformTypes': [], 'threatEntryTypes': [], 'threatEntries': []}} for prefix in prefixes: request_body['threatInfo']['threatEntries'].append({'hash': b64encode(prefix).decode()}) for (threatType, platformType, threatEntryType ), clientState in client_state.items(): request_body['clientStates'].append(clientState) if threatType not in request_body['threatInfo']['threatTypes']: request_body['threatInfo']['threatTypes'].append(threatType) if platformType not in request_body['threatInfo']['platformTypes']: request_body['threatInfo']['platformTypes'].append(platformType) if threatEntryType not in request_body['threatInfo']['threatEntryTypes' ]: request_body['threatInfo']['threatEntryTypes'].append( threatEntryType) response = self.service.fullHashes().find(body=request_body).execute() self.set_wait_duration(response.get('minimumWaitDuration')) return response
get_full_hashes
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, prefixes, client_state):\n request_body = {'client': {'clientId': self.client_id, 'clientVersion':\n self.client_version}, 'clientStates': [], 'threatInfo': {\n 'threatTypes': [], 'platformTypes': [], 'threatEntryTypes': [],\n 'threatEntries': []}}\n for prefix in prefixes:\n request_body['threatInfo']['threatEntries'].append({'hash':\n b64encode(prefix).decode()})\n for (threatType, platformType, threatEntryType\n ), clientState in client_state.items():\n request_body['clientStates'].append(clientState)\n if threatType not in request_body['threatInfo']['threatTypes']:\n request_body['threatInfo']['threatTypes'].append(threatType)\n if platformType not in request_body['threatInfo']['platformTypes']:\n request_body['threatInfo']['platformTypes'].append(platformType)\n if threatEntryType not in request_body['threatInfo']['threatEntryTypes'\n ]:\n request_body['threatInfo']['threatEntryTypes'].append(\n threatEntryType)\n response = self.service.fullHashes().find(body=request_body).execute()\n self.set_wait_duration(response.get('minimumWaitDuration'))\n return response\n", "role": "user" }, { "content": "get_full_hashes", "role": "assistant" } ]
def x(self): for url_variant in self.url_permutations(self.canonical): url_hash = self.digest(url_variant) yield url_hash
hashes
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n for url_variant in self.url_permutations(self.canonical):\n url_hash = self.digest(url_variant)\n yield url_hash\n", "role": "user" }, { "content": "hashes", "role": "assistant" } ]
def x(self): def x(u): uu = urllib.unquote(u) if uu == u: return uu else: return full_unescape(uu) def x(u): uu = urlparse.unquote_to_bytes(u) if uu == u: return uu else: return full_unescape_to_bytes(uu) def x(s): safe_chars = '!"$&\'()*+,-./:;<=>?@[\\]^_`{|}~' return urllib.quote(s, safe=safe_chars) url = self.url.strip() url = url.replace(b'\n', b'').replace(b'\r', b'').replace(b'\t', b'') url = url.split(b'#', 1)[0] if url.startswith(b'//'): url = b'http:' + url if len(url.split(b'://')) <= 1: url = b'http://' + url if self.__py3 and type(url) is bytes: url = quote(full_unescape_to_bytes(url)) else: url = quote(full_unescape(url)) url_parts = urlparse.urlsplit(url) if not url_parts[0]: url = 'http://{}'.format(url) url_parts = urlparse.urlsplit(url) protocol = url_parts.scheme if self.__py3: host = full_unescape_to_bytes(url_parts.hostname) path = full_unescape_to_bytes(url_parts.path) else: host = full_unescape(url_parts.hostname) path = full_unescape(url_parts.path) query = url_parts.query if not query and '?' not in url: query = None if not path: path = b'/' has_trailing_slash = path[-1:] == b'/' path = posixpath.normpath(path).replace(b'//', b'/') if has_trailing_slash and path[-1:] != b'/': path = path + b'/' port = url_parts.port host = host.strip(b'.') host = re.sub(b'\\.+', b'.', host).lower() if host.isdigit(): try: host = socket.inet_ntoa(struct.pack('!I', int(host))) except Exception: pass elif host.startswith(b'0x') and b'.' not in host: try: host = socket.inet_ntoa(struct.pack('!I', int(host, 16))) except Exception: pass quoted_path = quote(path) quoted_host = quote(host) if port is not None: quoted_host = '{}:{}'.format(quoted_host, port) canonical_url = '{}://{}{}'.format(protocol, quoted_host, quoted_path) if query is not None: canonical_url = '{}?{}'.format(canonical_url, query) return canonical_url
canonical
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n\n def x(u):\n uu = urllib.unquote(u)\n if uu == u:\n return uu\n else:\n return full_unescape(uu)\n\n def x(u):\n uu = urlparse.unquote_to_bytes(u)\n if uu == u:\n return uu\n else:\n return full_unescape_to_bytes(uu)\n\n def x(s):\n safe_chars = '!\"$&\\'()*+,-./:;<=>?@[\\\\]^_`{|}~'\n return urllib.quote(s, safe=safe_chars)\n url = self.url.strip()\n url = url.replace(b'\\n', b'').replace(b'\\r', b'').replace(b'\\t', b'')\n url = url.split(b'#', 1)[0]\n if url.startswith(b'//'):\n url = b'http:' + url\n if len(url.split(b'://')) <= 1:\n url = b'http://' + url\n if self.__py3 and type(url) is bytes:\n url = quote(full_unescape_to_bytes(url))\n else:\n url = quote(full_unescape(url))\n url_parts = urlparse.urlsplit(url)\n if not url_parts[0]:\n url = 'http://{}'.format(url)\n url_parts = urlparse.urlsplit(url)\n protocol = url_parts.scheme\n if self.__py3:\n host = full_unescape_to_bytes(url_parts.hostname)\n path = full_unescape_to_bytes(url_parts.path)\n else:\n host = full_unescape(url_parts.hostname)\n path = full_unescape(url_parts.path)\n query = url_parts.query\n if not query and '?' not in url:\n query = None\n if not path:\n path = b'/'\n has_trailing_slash = path[-1:] == b'/'\n path = posixpath.normpath(path).replace(b'//', b'/')\n if has_trailing_slash and path[-1:] != b'/':\n path = path + b'/'\n port = url_parts.port\n host = host.strip(b'.')\n host = re.sub(b'\\\\.+', b'.', host).lower()\n if host.isdigit():\n try:\n host = socket.inet_ntoa(struct.pack('!I', int(host)))\n except Exception:\n pass\n elif host.startswith(b'0x') and b'.' not in host:\n try:\n host = socket.inet_ntoa(struct.pack('!I', int(host, 16)))\n except Exception:\n pass\n quoted_path = quote(path)\n quoted_host = quote(host)\n if port is not None:\n quoted_host = '{}:{}'.format(quoted_host, port)\n canonical_url = '{}://{}{}'.format(protocol, quoted_host, quoted_path)\n if query is not None:\n canonical_url = '{}?{}'.format(canonical_url, query)\n return canonical_url\n", "role": "user" }, { "content": "canonical", "role": "assistant" } ]
def x(url): def x(host): if re.match('\\d+\\.\\d+\\.\\d+\\.\\d+', host): yield host return parts = host.split('.') l = min(len(parts), 5) if l > 4: yield host for i in range(l - 1): yield '.'.join(parts[i - l:]) def x(path): yield path query = None if '?' in path: path, query = path.split('?', 1) if query is not None: yield path path_parts = path.split('/')[0:-1] curr_path = '' for i in range(min(4, len(path_parts))): curr_path = curr_path + path_parts[i] + '/' yield curr_path protocol, address_str = urllib.splittype(url) host, path = urllib.splithost(address_str) user, host = urllib.splituser(host) host, port = urllib.splitport(host) host = host.strip('/') seen_permutations = set() for h in url_host_permutations(host): for p in url_path_permutations(path): u = '{}{}'.format(h, p) if u not in seen_permutations: yield u seen_permutations.add(u)
url_permutations
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(url):\n\n def x(host):\n if re.match('\\\\d+\\\\.\\\\d+\\\\.\\\\d+\\\\.\\\\d+', host):\n yield host\n return\n parts = host.split('.')\n l = min(len(parts), 5)\n if l > 4:\n yield host\n for i in range(l - 1):\n yield '.'.join(parts[i - l:])\n\n def x(path):\n yield path\n query = None\n if '?' in path:\n path, query = path.split('?', 1)\n if query is not None:\n yield path\n path_parts = path.split('/')[0:-1]\n curr_path = ''\n for i in range(min(4, len(path_parts))):\n curr_path = curr_path + path_parts[i] + '/'\n yield curr_path\n protocol, address_str = urllib.splittype(url)\n host, path = urllib.splithost(address_str)\n user, host = urllib.splituser(host)\n host, port = urllib.splitport(host)\n host = host.strip('/')\n seen_permutations = set()\n for h in url_host_permutations(host):\n for p in url_path_permutations(path):\n u = '{}{}'.format(h, p)\n if u not in seen_permutations:\n yield u\n seen_permutations.add(u)\n", "role": "user" }, { "content": "url_permutations", "role": "assistant" } ]
def x(v1, v2): def x(v): v = re.sub('(\\.0+)*$', '', v) result = [] for part in v.split('.'): m = re.match('^(\\d+)$', part) if m: result.append(int(m.group(1))) continue m = re.match('^(\\d+)([a-zA-Z]+)$', part) if m: result.append(int(m.group(1))) result.append(m.group(2)) continue m = re.match('^(\\d+)([a-zA-Z]+)(\\d+)$', part) if m: result.append(int(m.group(1))) result.append(m.group(2)) result.append(int(m.group(3))) continue return tuple(result) n1 = normalize(v1) n2 = normalize(v2) return (n1 > n2) - (n1 < n2)
compare_versions
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(v1, v2):\n\n def x(v):\n v = re.sub('(\\\\.0+)*$', '', v)\n result = []\n for part in v.split('.'):\n m = re.match('^(\\\\d+)$', part)\n if m:\n result.append(int(m.group(1)))\n continue\n m = re.match('^(\\\\d+)([a-zA-Z]+)$', part)\n if m:\n result.append(int(m.group(1)))\n result.append(m.group(2))\n continue\n m = re.match('^(\\\\d+)([a-zA-Z]+)(\\\\d+)$', part)\n if m:\n result.append(int(m.group(1)))\n result.append(m.group(2))\n result.append(int(m.group(3)))\n continue\n return tuple(result)\n n1 = normalize(v1)\n n2 = normalize(v2)\n return (n1 > n2) - (n1 < n2)\n", "role": "user" }, { "content": "compare_versions", "role": "assistant" } ]
def x(spec): m = re.search('([<>=]?=?)?\\s*([0-9.a-zA-Z]+)', spec) return m.group(2), m.group(1)
split_version_specifier
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(spec):\n m = re.search('([<>=]?=?)?\\\\s*([0-9.a-zA-Z]+)', spec)\n return m.group(2), m.group(1)\n", "role": "user" }, { "content": "split_version_specifier", "role": "assistant" } ]
def x(package): pkg_config_exe = os.environ.get('PKG_CONFIG', None) or 'pkg-config' cmd = '{0} --exists {1}'.format(pkg_config_exe, package).split() return call(cmd) == 0
exists
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(package):\n pkg_config_exe = os.environ.get('PKG_CONFIG', None) or 'pkg-config'\n cmd = '{0} --exists {1}'.format(pkg_config_exe, package).split()\n return call(cmd) == 0\n", "role": "user" }, { "content": "exists", "role": "assistant" } ]
def x(package, static=False): _raise_if_not_exists(package) return _query(package, *_build_options('--libs', static=static))
libs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(package, static=False):\n _raise_if_not_exists(package)\n return _query(package, *_build_options('--libs', static=static))\n", "role": "user" }, { "content": "libs", "role": "assistant" } ]
def x(package): _raise_if_not_exists(package) result = _query(package, '--print-variables') names = (x.strip() for x in result.split('\n') if x != '') return dict((x, _query(package, '--variable={0}'.format(x)).strip()) for x in names)
variables
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(package):\n _raise_if_not_exists(package)\n result = _query(package, '--print-variables')\n names = (x.strip() for x in result.split('\\n') if x != '')\n return dict((x, _query(package, '--variable={0}'.format(x)).strip()) for\n x in names)\n", "role": "user" }, { "content": "variables", "role": "assistant" } ]