source_code
stringlengths
45
7.24k
function_name
stringlengths
1
111
messages
list
def x(list_idl): seqs = set() for idl in list_idl.loci2seq: [seqs.add(s) for s in list_idl.loci2seq[idl]] return seqs
get_seqs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(list_idl):\n seqs = set()\n for idl in list_idl.loci2seq:\n [seqs.add(s) for s in list_idl.loci2seq[idl]]\n return seqs\n", "role": "user" ...
def x(s1, s2, i1, i2): c = len(set(s1).intersection(s2)) t = min(len(s1), len(s2)) pct = 1.0 * c / t * t is_gt = up_threshold(pct, t * 1.0, parameters.similar) logger.debug('_common: pct %s of clusters:%s %s = %s' % (1.0 * c / t, i1, i2, is_gt)) if pct < parameters.similar and is_gt and pct > 0: pct = parameters.similar return pct / t
common
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(s1, s2, i1, i2):\n c = len(set(s1).intersection(s2))\n t = min(len(s1), len(s2))\n pct = 1.0 * c / t * t\n is_gt = up_threshold(pct, t * 1.0, param...
def x(pairs, common, clus_seen, loci_similarity): all_true1 = all([all([(common and loci_similarity[p, c] > parameters. similar) for p in pairs if (p, c) in loci_similarity]) for c in clus_seen]) all_true2 = all([all([(common and loci_similarity[c, p] > parameters. similar) for p in pairs if (c, p) in loci_similarity]) for c in clus_seen]) return all_true1 * all_true2
is_consistent
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(pairs, common, clus_seen, loci_similarity):\n all_true1 = all([all([(common and loci_similarity[p, c] > parameters.\n similar) for p in pairs if (p, ...
def x(loci, loci_similarity): n_cluster = 0 internal_cluster = {} clus_seen = {} loci_sorted = sorted(loci_similarity.iteritems(), key=operator. itemgetter(1), reverse=True) for pairs, sim in loci_sorted: common = sim > parameters.similar n_cluster += 1 logger.debug('_merge_similar:try new cluster %s' % n_cluster) new_c = cluster(n_cluster) p_seen, p_unseen = [], [] size = min(len(_get_seqs(loci[pairs[0]])), len(_get_seqs(loci[pairs [1]]))) if common: consistent = _is_consistent(pairs, common, clus_seen, loci_similarity) logger.debug('_merge_similar: clusters seen: %s' % clus_seen) logger.debug( '_merge_similar: id %s common %s|%s total %s consistent %s' % (pairs, sim, common, size, consistent)) if not consistent: continue if pairs[0] in clus_seen: p_seen.append(pairs[0]) p_unseen.append(pairs[1]) if pairs[1] in clus_seen: p_seen.append(pairs[1]) p_unseen.append(pairs[0]) if len(p_seen) == 0: new_c = _merge_cluster(loci[pairs[0]], new_c) new_c = _merge_cluster(loci[pairs[1]], new_c) [clus_seen.update({p: n_cluster}) for p in pairs] internal_cluster[n_cluster] = new_c if len(p_seen) == 1: idc_seen = clus_seen[p_seen[0]] internal_cluster[idc_seen] = _merge_cluster(loci[p_unseen[0 ]], internal_cluster[idc_seen]) clus_seen[p_unseen[0]] = idc_seen else: logger.debug('_merge_similar: id %s %s are different' % pairs) continue internal_cluster.update(_add_unseen(loci, clus_seen, n_cluster)) logger.debug('_merge_similar: total clus %s' % len(internal_cluster.keys()) ) return internal_cluster
merge_similar
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(loci, loci_similarity):\n n_cluster = 0\n internal_cluster = {}\n clus_seen = {}\n loci_sorted = sorted(loci_similarity.iteritems(), key=operator.\...
def x(old, new): logger.debug('_merge_cluster: %s to %s' % (old.id, new.id)) logger.debug('_merge_cluster: add idls %s' % old.loci2seq.keys()) for idl in old.loci2seq: new.add_id_member(old.loci2seq[idl], idl) return new
merge_cluster
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(old, new):\n logger.debug('_merge_cluster: %s to %s' % (old.id, new.id))\n logger.debug('_merge_cluster: add idls %s' % old.loci2seq.keys())\n for idl...
def x(list_c, s2p, n_cluster): logger.debug('_solve_conflict: count once') if parameters.decision_cluster == 'bayes': return decide_by_bayes(list_c, s2p) loci_similarity = _calculate_similarity(list_c) loci_similarity = sorted(loci_similarity.iteritems(), key=operator. itemgetter(1), reverse=True) common = sum([score for p, score in loci_similarity]) while common > 0: n_cluster += 1 logger.debug('_solve_conflict: ma %s' % loci_similarity) pairs = loci_similarity[0][0] score = loci_similarity[0][1] logger.debug('_solve_conflict: common %s, new %s' % (score, n_cluster)) if parameters.decision_cluster.startswith('most-voted'): list_c = _split_cluster_by_most_vote(list_c, pairs) else: list_c = _split_cluster(list_c, pairs, n_cluster) list_c = {k: v for k, v in list_c.iteritems() if len(v.loci2seq) > 0} loci_similarity = _calculate_similarity(list_c) loci_similarity = sorted(loci_similarity.iteritems(), key=operator. itemgetter(1), reverse=True) common = sum([score for p, score in loci_similarity]) logger.debug('_solve_conflict: solved clusters %s' % len(list_c.keys()) ) return list_c
solve_conflict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(list_c, s2p, n_cluster):\n logger.debug('_solve_conflict: count once')\n if parameters.decision_cluster == 'bayes':\n return decide_by_bayes(list_...
def x(c, pairs, n): old = c[p[0]] new = c[p[1]] new_c = cluster(n) common = set(_get_seqs(old)).intersection(_get_seqs(new)) for idl in old.loci2seq: in_common = list(set(common).intersection(old.loci2seq[idl])) if len(in_common) > 0: logger.debug('_split_cluster: in_common %s with pair 1' % len( in_common)) new_c.add_id_member(in_common, idl) old.loci2seq[idl] = list(set(old.loci2seq[idl]) - set(common)) logger.debug('_split_cluster: len old %s with pair 1' % len(old .loci2seq)) for idl in new.loci2seq: in_common = list(set(common).intersection(new.loci2seq[idl])) if len(in_common) > 0: logger.debug('_split_cluster: in_common %s with pair 2' % len( in_common)) new_c.add_id_member(in_common, idl) new.loci2seq[idl] = list(set(new.loci2seq[idl]) - set(common)) logger.debug('_split_cluster: len old %s with pair 2' % len(new .loci2seq)) old.update() new.update() old.loci2seq = {k: v for k, v in old.loci2seq.iteritems() if len(v) > 0} new.loci2seq = {k: v for k, v in new.loci2seq.iteritems() if len(v) > 0} c[n] = new c[p[0]] = old c[p[1]] = new return c
split_cluster
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(c, pairs, n):\n old = c[p[0]]\n new = c[p[1]]\n new_c = cluster(n)\n common = set(_get_seqs(old)).intersection(_get_seqs(new))\n for idl in old....
def x(c, p): old, new = c[p[0]], c[p[1]] old_size = _get_seqs(old) new_size = _get_seqs(new) logger.debug('_most_vote: size of %s with %s - %s with %s' % (old.id, len(old_size), new.id, len(new_size))) if len(old_size) > len(new_size): keep, remove = old, new else: keep, remove = new, old common = list(set(old_size).intersection(new_size)) logger.debug('_most_vote: keep %s remove %s with common %s' % (keep.id, remove.id, len(common))) for idl in remove.loci2seq: if len(common) > 0: remove.loci2seq[idl] = list(set(remove.loci2seq[idl]) - set(common) ) keep.loci2seq = {k: v for k, v in keep.loci2seq.iteritems() if len(v) > 0} remove.loci2seq = {k: v for k, v in remove.loci2seq.iteritems() if len( v) > 0} keep.update() remove.update() c[keep.id] = keep c[remove.id] = remove return c
split_cluster_by_most_vote
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(c, p):\n old, new = c[p[0]], c[p[1]]\n old_size = _get_seqs(old)\n new_size = _get_seqs(new)\n logger.debug('_most_vote: size of %s with %s - %s wi...
def x(list_c): global REMOVED init = len(list_c) list_c = {k: v for k, v in list_c.iteritems() if len(_get_seqs(v)) > parameters.min_seqs} logger.debug('_clean_cluster: number of clusters %s ' % len(list_c.keys())) list_c = {k: _select_loci(v) for k, v in list_c.iteritems()} end = len(list_c) REMOVED += init - end return list_c
clean_cluster
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(list_c):\n global REMOVED\n init = len(list_c)\n list_c = {k: v for k, v in list_c.iteritems() if len(_get_seqs(v)) >\n parameters.min_seqs}\n ...
def x(c): loci_len = {k: len(v) for k, v in c.loci2seq.iteritems()} logger.debug('_select_loci: number of loci %s' % len(c.loci2seq.keys())) loci_len_sort = sorted(loci_len.iteritems(), key=operator.itemgetter(1), reverse=True) max_size = loci_len_sort[0][1] logger.debug('_select_loci: max size %s' % max_size) loci_clean = {locus: c.loci2seq[locus] for locus, size in loci_len_sort if size > 0.8 * max_size} c.loci2seq = loci_clean removed = list(set(c.idmembers.keys()) - set(_get_seqs(c))) c.add_id_member(removed, loci_len_sort[0][0]) logger.debug('_select_loci: number of loci %s after cleaning' % len(c. loci2seq.keys())) return c
select_loci
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(c):\n loci_len = {k: len(v) for k, v in c.loci2seq.iteritems()}\n logger.debug('_select_loci: number of loci %s' % len(c.loci2seq.keys()))\n loci_len_...
def x(c, locilen_sorted, seen_seqs, filtered, maxseq, n_cluster): first_run = 0 seen_seqs = list() n_cluster += 1 logger.debug('_solve_loci:new cluster %s' % n_cluster) new_c = cluster(n_cluster) for idl, lenl in locilen_sorted: locus_seqs = c.loci2seq[idl] if first_run == 0: seen_seqs = locus_seqs first_run = 1 first_idl = idl intersect = list(set(seen_seqs).intersection(locus_seqs)) common = 0 if intersect: common = len(intersect) * 1.0 / min(len(seen_seqs), len(locus_seqs) ) logger.debug( '_sole_loci:id %s idl %s len %s max %s seen %s inter %s common %s ' % (c.id, idl, lenl, maxseq, len(seen_seqs), len(intersect), common)) if common * 1.0 >= 0.6: if lenl * 1.0 >= 0.6 * maxseq: c, new_c, seen_seqs = _merge_loci_in_cluster(c, new_c, idl, seen_seqs) else: c, new_c, seen_seqs = _merge_with_first_loci(c, new_c, first_idl, idl, seen_seqs) else: c = _remove_seqs_from_loci(c, idl, seen_seqs) filtered[n_cluster] = new_c return c, seen_seqs, filtered, n_cluster
solve_loci_deprecated
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(c, locilen_sorted, seen_seqs, filtered, maxseq, n_cluster):\n first_run = 0\n seen_seqs = list()\n n_cluster += 1\n logger.debug('_solve_loci:new c...
def x(string): ann = set() if not string: return 'This cluster is inter-genic.' for item in string: for db in item: ann = ann.union(set(item[db])) return 'annotated as: %s ...' % ','.join(list(ann)[:3])
get_description
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(string):\n ann = set()\n if not string:\n return 'This cluster is inter-genic.'\n for item in string:\n for db in item:\n ann...
def x(profile): x = set() for sample in profile: x = x.union(set(profile[sample].keys())) if not x: return '' end, start = max(x), min(x) x = range(start, end, 4) scaled_profile = defaultdict(list) for pos in x: for sample in profile: y = _get_closer(profile[sample], pos) if y: scaled_profile[sample].append(profile[sample][y]) else: scaled_profile[sample].append(0) return {'x': list(x), 'y': scaled_profile, 'names': scaled_profile.keys()}
set_format
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(profile):\n x = set()\n for sample in profile:\n x = x.union(set(profile[sample].keys()))\n if not x:\n return ''\n end, start = max(...
def x(con, data): with con: cur = con.cursor() cur.execute('DROP TABLE IF EXISTS clusters;') cur.execute( 'CREATE TABLE clusters(Id INT, Description TEXT, Locus TEXT, Annotation TEXT, Sequences TEXT, Profile TXT, Precursor TXT)' ) for c in data[0]: locus = json.dumps(data[0][c]['loci']) annotation = json.dumps(data[0][c]['ann']) description = _get_description(data[0][c]['ann']) sequences = json.dumps(_get_sequences(data[0][c])) keys = data[0][c]['freq'][0].values()[0].keys() profile = 'Not available.' if 'profile' in data[0][c]: profile = json.dumps(_set_format(data[0][c]['profile'])) precursor = json.dumps(data[0][c].get('precursor')) cur.execute( "INSERT INTO clusters VALUES(%s, '%s', '%s', '%s', '%s', '%s', '%s')" % (c, description, locus, annotation, sequences, profile, precursor))
insert_data
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(con, data):\n with con:\n cur = con.cursor()\n cur.execute('DROP TABLE IF EXISTS clusters;')\n cur.execute(\n 'CREATE TABLE ...
def x(file_in): loc_id = 1 bedfile_clusters = '' bamfile = pybedtools.BedTool(file_in) bed = pybedtools.BedTool.bam_to_bed(bamfile) for c, start, end, name, q, strand in bed: loc_id += 1 bedfile_clusters += '%s\t%s\t%s\t%s\t%s\t%s\n' % (c, start, end, name, loc_id, strand) return bedfile_clusters
parse_align_file
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(file_in):\n loc_id = 1\n bedfile_clusters = ''\n bamfile = pybedtools.BedTool(file_in)\n bed = pybedtools.BedTool.bam_to_bed(bamfile)\n for c, s...
def x(seq_obj, in_file): name = '' index = 1 total = defaultdict(int) with open(in_file) as handle_in: line = handle_in.readline().strip() cols = line.split('\t') samples = cols[2:] for line in handle_in: line = line.strip() cols = line.split('\t') name = int(cols[0].replace('seq_', '')) seq = cols[1] exp = {} for i in range(len(samples)): exp[samples[i]] = int(cols[i + 2]) total[samples[i]] += int(cols[i + 2]) index = index + 1 if name in seq_obj: seq_obj[name].set_freq(exp) seq_obj[name].set_seq(seq) seq_obj = _normalize_seqs(seq_obj, total) return seq_obj, total, index
parse_ma_file
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(seq_obj, in_file):\n name = ''\n index = 1\n total = defaultdict(int)\n with open(in_file) as handle_in:\n line = handle_in.readline().strip...
def x(cols, field='name'): field = field.lower() try: group = cols[2] attrs = cols[8].split(';') name = [attr.strip().split(' ')[1] for attr in attrs if attr.strip( ).split(' ')[0].lower().endswith(field)] if not name: name = [attr.strip().split(' ')[1] for attr in attrs if attr. strip().split(' ')[0].lower().endswith('gene_id')] if not name: name = ['None'] biotype = [attr.strip().split(' ')[1] for attr in attrs if attr. strip().split(' ')[0].lower().endswith('biotype')] if biotype: group = biotype[0] c = cols[0] s = int(cols[3]) e = int(cols[4]) st = cols[6] return [c, s, e, st, group, name[0]] except (Exception, e): logger.error(cols) logger.error('File is not in correct format') logger.error('Expect chr source feature start end . strand attributes') logger.error("Attributes are 'gene_name SNCA; gene_id ENSG; '") logger.error('The 3rd column is used as type of small RNA (like miRNA)' ) logger.error("at least should contains '; *name NAME; '") logger.error(e) raise
read_gtf_line
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cols, field='name'):\n field = field.lower()\n try:\n group = cols[2]\n attrs = cols[8].split(';')\n name = [attr.strip().split(' ')...
def x(pos_a, pos_b): strd = '-' if pos_a[2] in pos_b[2]: strd = '+' if pos_a[2] in '+' and pos_b[2] in '+': lento5 = pos_a[0] - pos_b[1] + 1 lento3 = pos_a[1] - pos_b[1] + 1 if pos_a[2] in '+' and pos_b[2] in '-': lento5 = pos_a[1] - pos_b[0] + 1 lento3 = pos_a[0] - pos_b[1] + 1 if pos_a[2] in '-' and pos_b[2] in '+': lento5 = pos_a[0] - pos_b[1] + 1 lento3 = pos_a[1] - pos_b[0] + 1 if pos_a[2] in '-' and pos_b[2] in '-': lento3 = pos_a[0] - pos_b[0] + 1 lento5 = pos_a[1] - pos_b[1] + 1 else: lento5 = pos_a[0] - pos_b[0] + 1 lento3 = pos_a[1] - pos_b[1] + 1 return lento5, lento3, strd
position_in_feature
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(pos_a, pos_b):\n strd = '-'\n if pos_a[2] in pos_b[2]:\n strd = '+'\n if pos_a[2] in '+' and pos_b[2] in '+':\n lento5 = pos_a[0] - pos_...
def x(c, clus_obj, db, type_ann, feature_id='name'): id_sa, id_ea, id_id, id_idl, id_sta = 1, 2, 3, 4, 5 if type_ann == 'bed': id_sb = 7 id_eb = 8 id_stb = 11 id_tag = 9 ida = 0 clus_id = clus_obj.clus loci_id = clus_obj.loci db = os.path.splitext(db)[0] logger.debug('Type:%s\n' % type_ann) for cols in c.features(): if type_ann == 'gtf': cb, sb, eb, stb, db, tag = read_gtf_line(cols[6:], feature_id) else: sb = int(cols[id_sb]) eb = int(cols[id_eb]) stb = cols[id_stb] tag = cols[id_tag] id = int(cols[id_id]) idl = int(cols[id_idl]) if id in clus_id: clus = clus_id[id] sa = int(cols[id_sa]) ea = int(cols[id_ea]) ida += 1 lento5, lento3, strd = _position_in_feature([sa, ea, cols[ id_sta]], [sb, eb, stb]) if db in loci_id[idl].db_ann: ann = annotation(db, tag, strd, lento5, lento3) tdb = loci_id[idl].db_ann[db] tdb.add_db_ann(ida, ann) loci_id[idl].add_db(db, tdb) else: ann = annotation(db, tag, strd, lento5, lento3) tdb = dbannotation(1) tdb.add_db_ann(ida, ann) loci_id[idl].add_db(db, tdb) clus_id[id] = clus clus_obj.clus = clus_id clus_obj.loci = loci_id return clus_obj
anncluster
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(c, clus_obj, db, type_ann, feature_id='name'):\n id_sa, id_ea, id_id, id_idl, id_sta = 1, 2, 3, 4, 5\n if type_ann == 'bed':\n id_sb = 7\n ...
def x(bam_in, genome, out): if not genome: logger.info('No genome given. skipping.') return None out_file = op.join(out, op.basename(bam_in) + '_cov.tsv') if file_exists(out_file): return None fai = genome + '.fai' cov = pybedtools.BedTool(bam_in).genome_coverage(g=fai, max=1) cov.saveas(out_file) total = 0 for region in cov: if region[0] == 'genome' and int(region[1]) != 0: total += float(region[4]) logger.info('Total genome with sequences: %s ' % total)
detect_complexity
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(bam_in, genome, out):\n if not genome:\n logger.info('No genome given. skipping.')\n return None\n out_file = op.join(out, op.basename(bam_...
def x(bam_in, mask=None): seq_obj = defaultdict(int) if mask: mask_file = op.splitext(bam_in)[0] + '_mask.bam' if not file_exists(mask_file): pybedtools.BedTool(bam_file).intersect(b=mask, v=True).saveas( mask_file) bam_in = mask_file out_file = op.splitext(bam_in)[0] + '_rmlw.bam' run('samtools index %s' % bam_in) if not file_exists(bam_in + '.bai'): raise IOError( 'Failed to created bam index of %s. Try to do it manually' % bam_in ) bam_handle = pysam.AlignmentFile(bam_in, 'rb') with pysam.AlignmentFile(out_file, 'wb', template=bam_handle ) as out_handle: for read in bam_handle.fetch(): seq_name = int(read.query_name.replace('seq_', '')) match_size = [nts for oper, nts in read.cigartuples if oper == 0] subs_size = [nts for oper, nts in read.cigartuples if oper == 4] if match_size[0] < 17: continue if subs_size: if subs_size[0] > 3: continue try: nh = read.get_tag('NH') except KeyError: nh = 1 seq_obj[seq_name] = sequence(seq_name) seq_obj[seq_name].align = nh out_handle.write(read) return out_file, seq_obj
clean_bam_file
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(bam_in, mask=None):\n seq_obj = defaultdict(int)\n if mask:\n mask_file = op.splitext(bam_in)[0] + '_mask.bam'\n if not file_exists(mask_fi...
def x(c, current_seq, MIN_SEQ, non_un_gl=False): current_loci = {} current_clus = {} sequence2clusters = defaultdict(set) lindex = 0 eindex = 0 previous_id = 0 for line in c.features(): c, start, end, name, score, strand, c_id = line name = int(name.replace('seq_', '')) pos = int(start) if strand == '+' else int(end) if name not in current_seq: continue if c.find('Un_gl') > -1 and non_un_gl: continue if c_id != previous_id: if previous_id > 0: if len(current_clus[eindex].idmembers) < MIN_SEQ: for s in current_clus[eindex].idmembers: sequence2clusters[s] = sequence2clusters[s] - set([ eindex]) del current_clus[eindex] logger.debug('detect_cluster: %s %s %s' % (c_id, previous_id, name) ) lindex += 1 eindex += 1 current_clus[eindex] = cluster(eindex) newpos = position(lindex, c, start, end, strand) current_loci[lindex] = newpos current_loci[lindex].end = int(end) current_loci[lindex].coverage[pos] += 1 size = range(pos, pos + current_seq[name].len) current_loci[lindex].counts.update(dict(zip(size, [current_seq[name ].total()] * current_seq[name].len))) current_clus[eindex].idmembers[name] = 1 current_clus[eindex].add_id_member([name], lindex) current_seq[name].add_pos(lindex, pos) previous_id = c_id sequence2clusters[name].add(eindex) logger.info('%s Clusters read' % eindex) metacluster_obj, cluster_id = _find_metaclusters(current_clus, sequence2clusters, current_seq, MIN_SEQ) return cluster_info_obj(current_clus, metacluster_obj, current_loci, current_seq)
detect_clusters
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(c, current_seq, MIN_SEQ, non_un_gl=False):\n current_loci = {}\n current_clus = {}\n sequence2clusters = defaultdict(set)\n lindex = 0\n eindex ...
def x(clus_obj, sequence2clusters, current_seq, min_seqs): seen = defaultdict(int) metacluster = defaultdict(set) c_index = len(sequence2clusters) logger.info('Creating meta-clusters based on shared sequences: %s' % c_index) meta_idx = 1 bar = ProgressBar(maxval=c_index) bar.start() bar.update() for itern, name in enumerate(sequence2clusters): clusters = sequence2clusters[name] if len(clusters) == 0: c_index -= 1 continue current_seq[name].align = 1 meta_idx += 1 bar.update(itern) already_in = _common(clusters, seen) _update(clusters, meta_idx, seen) metacluster[meta_idx] = metacluster[meta_idx].union(clusters) if already_in: for seen_metacluster in already_in: clusters2merge = metacluster[seen_metacluster] metacluster[meta_idx] = metacluster[meta_idx].union( clusters2merge) _update(clusters2merge, meta_idx, seen) del metacluster[seen_metacluster] logger.info('%s metaclusters from %s sequences' % (len(metacluster), c_index)) return metacluster, seen
find_metaclusters
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(clus_obj, sequence2clusters, current_seq, min_seqs):\n seen = defaultdict(int)\n metacluster = defaultdict(set)\n c_index = len(sequence2clusters)\n ...
def x(clus_obj, min_seqs): logger.info('Creating meta-clusters based on shared sequences.') seen = defaultdict() metacluster = defaultdict(list) c_index = clus_obj.keys() meta_idx = 0 with ProgressBar(maxval=len(c_index), redirect_stdout=True) as p: for itern, c in enumerate(c_index): p.update(itern) clus = clus_obj[c] if len(clus.idmembers.keys()) < min_seqs: del clus_obj[c] continue logger.debug('reading cluster %s' % c) logger.debug('loci2seq %s' % clus.loci2seq) already_in, not_in = _get_seqs_from_cluster(clus.idmembers.keys (), seen) logger.debug('seen %s news %s' % (already_in, not_in)) meta_idx += 1 metacluster[meta_idx].append(c) seen.update(dict(zip(not_in, [meta_idx] * len(not_in)))) if len(already_in) > 0: logger.debug('seen in %s' % already_in) for eindex in already_in: for cluster in metacluster[eindex]: metacluster[meta_idx].append(cluster) prev_clus = clus_obj[cluster] logger.debug('_find_families: prev %s current %s' % (eindex, clus.id)) seqs_in = prev_clus.idmembers.keys() seen.update(dict(zip(seqs_in, [meta_idx] * len( seqs_in)))) del metacluster[eindex] logger.info('%s clusters merged' % len(metacluster)) return metacluster, seen
find_families_deprecated
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(clus_obj, min_seqs):\n logger.info('Creating meta-clusters based on shared sequences.')\n seen = defaultdict()\n metacluster = defaultdict(list)\n ...
def x(clus_obj): new_cluster = {} for cid in clus_obj.clus: cluster = clus_obj.clus[cid] cluster.update() logger.debug('peak calling for %s' % cid) bigger = cluster.locimaxid if bigger in clus_obj.loci: s, e = min(clus_obj.loci[bigger].counts.keys()), max(clus_obj. loci[bigger].counts.keys()) scale = s if clus_obj.loci[bigger].strand == '-': scale = e logger.debug('bigger %s at %s-%s' % (bigger, s, e)) dt = np.array([0] * (abs(e - s) + 12)) for pos in clus_obj.loci[bigger].counts: ss = abs(int(pos) - scale) + 5 dt[ss] += clus_obj.loci[bigger].counts[pos] x = np.array(range(0, len(dt))) logger.debug('x %s and y %s' % (x, dt)) if len(x) > 35 + 12: peaks = list(np.array(pysen.pysenMMean(x, dt)) - 5) logger.debug(peaks) else: peaks = ['short'] cluster.peaks = peaks new_cluster[cid] = cluster clus_obj.clus = new_cluster return clus_obj
peak_calling
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(clus_obj):\n new_cluster = {}\n for cid in clus_obj.clus:\n cluster = clus_obj.clus[cid]\n cluster.update()\n logger.debug('peak cal...
def x(args): if args.fasta: name = None seq = '' reads = dict() with open(args.fasta) as in_handle: for line in in_handle: if line.startswith('>'): if name: reads.update(_generate_reads(seq, name)) seq = '' name = line[1:-1] else: seq += line.strip() reads.update(_generate_reads(seq, name)) _write_reads(reads, args.out)
simulate
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(args):\n if args.fasta:\n name = None\n seq = ''\n reads = dict()\n with open(args.fasta) as in_handle:\n for line in...
def x(seq, name): reads = dict() if len(seq) < 130 and len(seq) > 70: reads.update(_mature(seq[:40], 0, name)) reads.update(_mature(seq[-40:], len(seq) - 40, name)) reads.update(_noise(seq, name)) reads.update(_noise(seq, name, 25)) return reads
generate_reads
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(seq, name):\n reads = dict()\n if len(seq) < 130 and len(seq) > 70:\n reads.update(_mature(seq[:40], 0, name))\n reads.update(_mature(seq[-...
def x(subseq, absolute, c, size=33, total=5000): reads = dict() probs = [0.1, 0.2, 0.4, 0.2, 0.1] end = 5 + size error = [-2, -1, 0, 1, 2] for error5 in error: for error3 in error: s = 5 - error5 e = end - error3 seen = subseq[s:e] counts = int(probs[error5 + 2] * probs[error3 + 2] * total) + 1 name = 'seq_%s_%s_%s_x%s' % (c, s + absolute, e + absolute, counts) reads[name] = seen, counts return reads
mature
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(subseq, absolute, c, size=33, total=5000):\n reads = dict()\n probs = [0.1, 0.2, 0.4, 0.2, 0.1]\n end = 5 + size\n error = [-2, -1, 0, 1, 2]\n f...
def x(seq, c, size=33, total=1000): reads = dict() seen = 0 while seen < total: s = random.randint(0, len(seq) - size) e = s + size + random.randint(-5, 5) p = random.uniform(0, 0.1) counts = int(p * total) + 1 seen += counts name = 'seq_%s_%s_%s_x%s' % (c, s, e, counts) reads[name] = seq[s:e], counts return reads
noise
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(seq, c, size=33, total=1000):\n reads = dict()\n seen = 0\n while seen < total:\n s = random.randint(0, len(seq) - size)\n e = s + size ...
def x(reads, prefix): out_ma = prefix + '.ma' out_fasta = prefix + '.fasta' out_real = prefix + '.txt' with open(out_ma, 'w') as ma_handle: print('id\tseq\tsample', file=ma_handle, end='') with open(out_fasta, 'w') as fa_handle: with open(out_real, 'w') as read_handle: for idx, r in enumerate(reads): info = r.split('_') print('seq_%s\t%s\t%s' % (idx, reads[r][0], reads[r][1] ), file=ma_handle, end='') print('>seq_%s\n%s' % (idx, reads[r][0]), file= fa_handle, end='') print('%s\t%s\t%s\t%s\t%s\t%s\t%s' % (idx, r, reads[r][ 0], reads[r][1], info[1], info[2], info[3]), file= read_handle, end='')
write_reads
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(reads, prefix):\n out_ma = prefix + '.ma'\n out_fasta = prefix + '.fasta'\n out_real = prefix + '.txt'\n with open(out_ma, 'w') as ma_handle:\n ...
def x(args): logger.info('Reading sequeces') data = parse_ma_file(args.ma) logger.info('Get sequences from sam') is_align = _read_sam(args.sam) is_json, is_db = _read_json(args.json) res = _summarise_sam(data, is_align, is_json, is_db) _write_suma(res, os.path.join(args.out, 'stats_align.dat')) logger.info('Done')
stats
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(args):\n logger.info('Reading sequeces')\n data = parse_ma_file(args.ma)\n logger.info('Get sequences from sam')\n is_align = _read_sam(args.sam)\n...
def x(fn_json): is_json = set() is_db = {} with open(fn_json) as handle: data = json.load(handle) for item in list(data[0].values()): seqs_name = [list(x.keys()) for x in item['seqs']] db_name = item['valid'] if 'valid' in item else None [is_json.add(name[0]) for name in seqs_name] if db_name: [is_db.update({name[0]: ','.join(db_name)}) for name in seqs_name] return is_json, is_db
read_json
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(fn_json):\n is_json = set()\n is_db = {}\n with open(fn_json) as handle:\n data = json.load(handle)\n for item in list(data[0].values())...
def x(cmd, data=None, checks=None, region=None, log_error=True, log_stdout= False): try: logger.debug(' '.join(str(x) for x in cmd) if not isinstance(cmd, basestring) else cmd) _do_run(cmd, checks, log_stdout) except: if log_error: logger.info('error at command') raise
run
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cmd, data=None, checks=None, region=None, log_error=True, log_stdout=\n False):\n try:\n logger.debug(' '.join(str(x) for x in cmd) if not isinsta...
def x(cmd): if isinstance(cmd, basestring): if cmd.find(' | ') > 0 or cmd.find('>(') or cmd.find('<('): return 'set -o pipefail; ' + cmd, True, find_bash() else: return cmd, True, None else: return [str(x) for x in cmd], False, None
normalize_cmd_args
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cmd):\n if isinstance(cmd, basestring):\n if cmd.find(' | ') > 0 or cmd.find('>(') or cmd.find('<('):\n return 'set -o pipefail; ' + cmd, ...
def x(cmd, checks, log_stdout=False): cmd, shell_arg, executable_arg = _normalize_cmd_args(cmd) s = subprocess.Popen(cmd, shell=shell_arg, executable=executable_arg, stdout=subprocess.PIPE, stderr=subprocess.STDOUT, close_fds=True) debug_stdout = collections.deque(maxlen=100) while 1: line = s.stdout.readline() if line: debug_stdout.append(line) if log_stdout: logger.debug(line.rstrip()) else: logger.debug(line.rstrip()) exitcode = s.poll() if exitcode is not None: for line in s.stdout: debug_stdout.append(line) if exitcode is not None and exitcode != 0: error_msg = ' '.join(cmd) if not isinstance(cmd, basestring ) else cmd error_msg += '\n' error_msg += ''.join(debug_stdout) s.communicate() s.stdout.close() raise subprocess.CalledProcessError(exitcode, error_msg) else: break s.communicate() s.stdout.close() if checks: for check in checks: if not check(): raise IOError('External command failed')
do_run
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cmd, checks, log_stdout=False):\n cmd, shell_arg, executable_arg = _normalize_cmd_args(cmd)\n s = subprocess.Popen(cmd, shell=shell_arg, executable=execu...
def x(list_c, loci_obj, seq_obj): seqs = defaultdict(set) for c in list_c.values(): for l in c.loci2seq: [seqs[s].add(c.id) for s in c.loci2seq[l]] common = [s for s in seqs if len(seqs[s]) > 1] seqs_in_c = defaultdict(float) for c in list_c.values(): for l in c.loci2seq: for s in c.loci2seq[l]: if s in common: pos = seq_obj[s].pos[l] cov = 1.0 * loci_obj[l].coverage[pos] if seqs_in_c[s, c.id] < cov: seqs_in_c[s, c.id] = cov seqs_in_c = _transform(seqs_in_c) return seqs_in_c
dict_seq_locus
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(list_c, loci_obj, seq_obj):\n seqs = defaultdict(set)\n for c in list_c.values():\n for l in c.loci2seq:\n [seqs[s].add(c.id) for s in ...
def x(self, data): for hypo in self.Values(): like = self.Likelihood(data, hypo) self.Mult(hypo, like) self.Normalize()
Update
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, data):\n for hypo in self.Values():\n like = self.Likelihood(data, hypo)\n self.Mult(hypo, like)\n self.Normalize()\n", "role": "...
def x(self, data, hypo): mix = self.loci[hypo] like = mix[data] return like
Likelihood
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, data, hypo):\n mix = self.loci[hypo]\n like = mix[data]\n return like\n", "role": "user" }, { "content": "Likelihood", "role": "...
def x(clus_obj, index): current = clus_obj.clus clus_seqt = clus_obj.seq clus_locit = clus_obj.loci itern = 0 for idc in current.keys(): itern += 1 timestamp = str(idc) seqListTemp = () f = open('/tmp/' + timestamp + '.fa', 'w') for idl in current[idc].loci2seq.keys(): seqListTemp = list(set(seqListTemp).union(current[idc].loci2seq [idl])) maxscore = 0 for s in seqListTemp: score = calculate_size(clus_seqt[s].freq) maxscore = max(maxscore, score) clus_seqt[s].score = score seq = clus_seqt[s] f.write('>' + s + '\n' + seq.seq + '\n') f.close() locilen_sorted = sorted(current[idc].locilen.iteritems(), key= operator.itemgetter(1), reverse=True) lmax = clus_locit[locilen_sorted[0][0]] f = open('/tmp/' + timestamp + '.bed', 'w') f.write('%s\t%s\t%s\t.\t.\t%s\n' % (lmax.chr, lmax.start, lmax.end, lmax.strand)) f.close() os.system('bedtools getfasta -s -fi ' + index + ' -bed /tmp/' + timestamp + '.bed -fo /tmp/' + timestamp + '.pre.fa') os.system('bowtie2-build /tmp/' + timestamp + '.pre.fa /tmp/' + timestamp + '.pre.ind >/dev/null 2>&1') os.system( 'bowtie2 --rdg 7,3 --mp 4 --end-to-end --no-head --no-sq -D 20 -R 3 -N 0 -i S,1,0.8 -L 3 -f /tmp/' + timestamp + '.pre.ind /tmp/' + timestamp + '.fa -S /tmp/' + timestamp + '.map >>bowtie.log 2>&1') f = open('/tmp/' + timestamp + '.map', 'r') seqpos = {} minv = 10000000 for line in f: line = line.strip() cols = line.split('\t') seqpos[cols[0]] = int(cols[3]) if minv > int(cols[3]): minv = int(cols[3]) f.close() seqpos_sorted = sorted(seqpos.iteritems(), key=operator.itemgetter( 1), reverse=False) showseq = '' showseq_plain = '' for s, pos in seqpos_sorted: ratio = clus_seqt[s].score * 1.0 / maxscore * 100.0 realScore = math.log(ratio, 2) * 2 if realScore < 0: realScore = 0 showseq_plain += ( '<br>%s<a style = "font-size:%.0fpx;"href = javascript:loadSeq("%s")>%s</a>' % (''.join('.' for i in range(pos - 1)), realScore + 10, s, clus_seqt[s].seq)) current[idc].showseq_plain = showseq_plain os.system('rm /tmp/' + timestamp + '*') clus_obj.clus = current clus_obj.seq = clus_seqt return clus_obj
show_seq
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(clus_obj, index):\n current = clus_obj.clus\n clus_seqt = clus_obj.seq\n clus_locit = clus_obj.loci\n itern = 0\n for idc in current.keys():\n ...
def x(s, t): for ids in s: obj = s[ids] [obj.norm_freq.update({sample: 1.0 * obj.freq[sample] / (t[sample] + 1) * 1000000}) for sample in obj.norm_freq] s[ids] = obj return s
normalize_seqs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(s, t):\n for ids in s:\n obj = s[ids]\n [obj.norm_freq.update({sample: 1.0 * obj.freq[sample] / (t[sample] +\n 1) * 1000000}) for s...
def x(bam_in, precursors): a = pybedtools.BedTool(bam_in) b = pybedtools.BedTool(precursors) c = a.intersect(b, u=True) out_file = utils.splitext_plus(op.basename(bam_in))[0] + '_clean.bam' c.saveas(out_file) return op.abspath(out_file)
prepare_bam
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(bam_in, precursors):\n a = pybedtools.BedTool(bam_in)\n b = pybedtools.BedTool(precursors)\n c = a.intersect(b, u=True)\n out_file = utils.splitext...
def x(bed_file): new_bed = utils.splitext_plus(bed_file)[0] + '_order.bed' with open(bed_file) as in_handle: with open(new_bed, 'w') as out_handle: for line in in_handle: cols = line.strip().split('\t') cols[3] = _select_anno(cols[3]) + '_' + cols[4] cols[4] = '0' print('\t'.join(cols), file=out_handle, end='') return new_bed
reorder_columns
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(bed_file):\n new_bed = utils.splitext_plus(bed_file)[0] + '_order.bed'\n with open(bed_file) as in_handle:\n with open(new_bed, 'w') as out_handle...
def x(cov_file): new_cov = utils.splitext_plus(cov_file)[0] + '_fix.cov' with open(cov_file) as in_handle: with open(new_cov, 'w') as out_handle: for line in in_handle: cols = line.strip().split('\t') cols[4] = cols[6] print('\t'.join(cols[0:6]), file=out_handle, end='') return new_cov
fix_score_column
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cov_file):\n new_cov = utils.splitext_plus(cov_file)[0] + '_fix.cov'\n with open(cov_file) as in_handle:\n with open(new_cov, 'w') as out_handle:\...
def x(bam_in, bed_file, out_dir, prefix): bed_file = _reorder_columns(bed_file) counts_reads_cmd = ( 'coverageBed -s -counts -b {bam_in} -a {bed_file} | sort -k4,4 > {out_dir}/loci.cov' ) with utils.chdir(out_dir): run(counts_reads_cmd.format(min_trimmed_read_len= min_trimmed_read_len, max_trimmed_read_len=max_trimmed_read_len, **locals()), 'Run counts_reads') loci_file = _fix_score_column(op.join(out_dir, 'loci.cov')) return loci_file
detect_regions
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(bam_in, bed_file, out_dir, prefix):\n bed_file = _reorder_columns(bed_file)\n counts_reads_cmd = (\n 'coverageBed -s -counts -b {bam_in} -a {bed_f...
def x(cov_file, min_reads): new_cov = op.join(op.dirname(cov_file), 'feat_antisense.txt') with open(cov_file) as in_handle: with open(new_cov, 'w') as out_handle: print('name\tantisense', file=out_handle, end='') for line in in_handle: cols = line.strip().split('\t') cols[6] = 0 if cols[6] < min_reads else cols[6] print('%s\t%s' % (cols[3], cols[6]), file=out_handle, end='') return new_cov
order_antisense_column
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cov_file, min_reads):\n new_cov = op.join(op.dirname(cov_file), 'feat_antisense.txt')\n with open(cov_file) as in_handle:\n with open(new_cov, 'w'...
def x(bam_in, loci_file, out_dir): data = Counter() a = pybedtools.BedTool(bam_in) b = pybedtools.BedTool(loci_file) c = a.intersect(b, s=True, bed=True, wo=True) for line in c: end = int(line[1]) + 1 + int(line[2]) if line[5] == '+' else int(line [1]) + 1 start = int(line[1]) + 1 if line[5] == '+' else int(line[1]) + 1 + int( line[2]) side5 = '%s\t5p\t%s' % (line[15], start) side3 = '%s\t3p\t%s' % (line[15], end) data[side5] += 1 data[side3] += 1 counts_reads = op.join(out_dir, 'locus_readpos.counts') with open(counts_reads, 'w') as out_handle: for k in data: print(k, file=out_handle, end='') return counts_reads
reads_per_position
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(bam_in, loci_file, out_dir):\n data = Counter()\n a = pybedtools.BedTool(bam_in)\n b = pybedtools.BedTool(loci_file)\n c = a.intersect(b, s=True, b...
def x(bam_in, loci_file, reference, out_dir): lenvec_plus = op.join(out_dir, 'genomic_lenvec.plus') lenvec_minus = op.join(out_dir, 'genomic_lenvec.minus') compute_genomic_cmd = ( 'compute_genomic_lenvectors {bam_in} {lenvec_plus} {lenvec_minus} {min_len} {max_len} ' ) index_genomic_cmd = 'index_genomic_lenvectors {lenvec} ' genomic_lenvec = op.join(out_dir, 'genomic_lenvec') feat_len_file = op.join(out_dir, 'feat_lengths.txt') compute_locus_cmd = ( 'compute_locus_lenvectors {loci_file} {genomic_lenvec} {min_len} {max_len} > {feat_len_file}' ) cov_S_file = op.join(out_dir, 'loci.cov_anti') coverage_anti_cmd = ( 'coverageBed -S -counts -b {bam_in} -a {loci_file} > {cov_S_file}') feat_posentropy = op.join(out_dir, 'feat_posentropy.txt') entropy_cmd = 'compute_locus_entropy.rb {counts_reads} > {feat_posentropy}' with utils.chdir(out_dir): run(compute_genomic_cmd.format(min_len=min_trimmed_read_len, max_len=max_trimmed_read_len, **locals()), 'Run compute_genomic') run(index_genomic_cmd.format(lenvec=lenvec_plus), 'Run index in plus') run(index_genomic_cmd.format(lenvec=lenvec_minus), 'Run index in minus' ) run(compute_locus_cmd.format(min_len=min_trimmed_read_len, max_len= max_trimmed_read_len, **locals()), 'Run compute locus') run(coverage_anti_cmd.format(**locals()), 'Run coverage antisense') feat_antisense = _order_antisense_column(cov_S_file, min_trimmed_read_len) counts_reads = _reads_per_position(bam_in, loci_file, out_dir) run(entropy_cmd.format(**locals()), 'Run entropy') rnafold = calculate_structure(loci_file, reference)
create_features
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(bam_in, loci_file, reference, out_dir):\n lenvec_plus = op.join(out_dir, 'genomic_lenvec.plus')\n lenvec_minus = op.join(out_dir, 'genomic_lenvec.minus')...
def x(args): logger.info('reading sequeces') data = load_data(args.json) logger.info('create profile') data = make_profile(data, os.path.join(args.out, 'profiles'), args) logger.info('create database') make_database(data, 'seqcluster.db', args.out) logger.info( 'Done. Download https://github.com/lpantano/seqclusterViz/archive/master.zip to browse the output.' )
report
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(args):\n logger.info('reading sequeces')\n data = load_data(args.json)\n logger.info('create profile')\n data = make_profile(data, os.path.join(arg...
def x(peaks): previous = peaks[0] new_peaks = [previous] for pos in peaks: if pos > previous + 10: new_peaks.add(pos) previous = pos return new_peaks
summarize_peaks
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(peaks):\n previous = peaks[0]\n new_peaks = [previous]\n for pos in peaks:\n if pos > previous + 10:\n new_peaks.add(pos)\n p...
def x(x, y, win=10): previous = min(y) peaks = [] intervals = range(x, y, win) for pos in intervals: if y[pos] > previous * 10: previous = y[pos] peaks.add(pos) peaks = _summarize_peaks(peaks)
find_mature
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(x, y, win=10):\n previous = min(y)\n peaks = []\n intervals = range(x, y, win)\n for pos in intervals:\n if y[pos] > previous * 10:\n ...
def x(in_file): keep = Counter() with open_fastq(in_file) as handle: for line in handle: if line.startswith('@'): if line.find('UMI') > -1: logger.info( 'Find UMI tags in read names, collapsing by UMI.') return collapse_umi(in_file) seq = handle.next().strip() handle.next() qual = handle.next().strip() if seq in keep: keep[seq].update(qual) else: keep[seq] = quality(qual) logger.info('Sequences loaded: %s' % len(keep)) return keep
collapse
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(in_file):\n keep = Counter()\n with open_fastq(in_file) as handle:\n for line in handle:\n if line.startswith('@'):\n if...
def x(in_file): keep = defaultdict(dict) with open_fastq(in_file) as handle: for line in handle: if line.startswith('@'): m = re.search('UMI_([ATGC]*)', line.strip()) umis = m.group(0) seq = handle.next().strip() handle.next() qual = handle.next().strip() if (umis, seq) in keep: keep[umis, seq][1].update(qual) keep[umis, seq][0].update(seq) else: keep[umis, seq] = [umi(seq), quality(qual)] logger.info('Sequences loaded: %s' % len(keep)) return keep
collapse_umi
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(in_file):\n keep = defaultdict(dict)\n with open_fastq(in_file) as handle:\n for line in handle:\n if line.startswith('@'):\n ...
def x(in_file): _, ext = os.path.splitext(in_file) if ext == '.gz': return gzip.open(in_file, 'rb') if ext in ['.fastq', '.fq', '.fasta', '.fa']: return open(in_file, 'r') return ValueError('File needs to be fastq|fasta|fq|fa [.gz]')
open_fastq
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(in_file):\n _, ext = os.path.splitext(in_file)\n if ext == '.gz':\n return gzip.open(in_file, 'rb')\n if ext in ['.fastq', '.fq', '.fasta', '.f...
def x(args): try: umi_fn = args.fastq if _is_umi(args.fastq): umis = collapse(args.fastq) umi_fn = os.path.join(args.out, splitext_plus(os.path.basename( args.fastq))[0] + '_umi_trimmed.fastq') write_output(umi_fn, umis, args.minimum) seqs = collapse(umi_fn) out_file = splitext_plus(os.path.basename(args.fastq))[0 ] + '_trimmed.fastq' except IOError as e: logger.error('I/O error({0}): {1}'.format(e.errno, e.strerror)) raise 'Can not read file' out_file = os.path.join(args.out, out_file) write_output(out_file, seqs, args.minimum) return out_file
collapse_fastq
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(args):\n try:\n umi_fn = args.fastq\n if _is_umi(args.fastq):\n umis = collapse(args.fastq)\n umi_fn = os.path.join(args...
def x(doctree): current = 0 num_children = len(doctree.children) while current < num_children: child = doctree.children[current] child.replace_self(child.traverse(no_autoslides_filter)) if len(doctree.children) == num_children: current += 1 else: num_children = len(doctree.children)
filter_doctree_for_slides
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(doctree):\n current = 0\n num_children = len(doctree.children)\n while current < num_children:\n child = doctree.children[current]\n chi...
def x(self, node, increment=True): parent_title_node = node.parent.next_node(nodes.title) nextslide_info = getattr(parent_title_node, 'nextslide_info', ( parent_title_node.deepcopy().children, 1)) nextslide_info = nextslide_info[0], nextslide_info[1] + 1 if node.args: textnodes, messages = node.state.inline_text(node.args[0], 1) new_title = nodes.title(node.args[0], '', *textnodes) else: title_nodes = nextslide_info[0][:] if 'increment' in node.attributes: title_nodes.append(nodes.Text(' (%s)' % nextslide_info[1])) new_title = nodes.title('', '', *title_nodes) new_title.nextslide_info = nextslide_info return new_title
make_title_node
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, node, increment=True):\n parent_title_node = node.parent.next_node(nodes.title)\n nextslide_info = getattr(parent_title_node, 'nextslide_info', (\n...
def x(self, builder): if 'theme' in self.attributes: builder.apply_theme(self.attributes['theme'], builder.theme_options)
apply
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, builder):\n if 'theme' in self.attributes:\n builder.apply_theme(self.attributes['theme'], builder.theme_options)\n", "role": "user" }, ...
def x(cls, builder, doctree=None): result = {'theme': builder.config.slide_theme, 'autoslides': builder. config.autoslides, 'slide_classes': []} if doctree: conf_node = cls.get(doctree) if conf_node: result.update(conf_node.attributes) return result
get_conf
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(cls, builder, doctree=None):\n result = {'theme': builder.config.slide_theme, 'autoslides': builder.\n config.autoslides, 'slide_classes': []}\n i...
def x(context): COPY_LISTS = 'script_files', 'css_files' for attr in COPY_LISTS: if attr in context: context[attr] = context[attr][:] return context
fix_context
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(context):\n COPY_LISTS = 'script_files', 'css_files'\n for attr in COPY_LISTS:\n if attr in context:\n context[attr] = context[attr][:]...
def x(d): msg = bold('Welcome to the Hieroglyph %s quickstart utility.') % (version (),) print(msg) msg = """ This will ask questions for creating a Hieroglyph project, and then ask some basic Sphinx questions. """ print(msg) d.update({'version': datetime.date.today().strftime('%Y.%m.%d'), 'release': datetime.date.today().strftime('%Y.%m.%d'), 'make_mode': True}) if 'project' not in d: print('\nThe presentation title will be included on the title slide.') sphinx.quickstart.do_prompt(d, 'project', 'Presentation title') if 'author' not in d: sphinx.quickstart.do_prompt(d, 'author', 'Author name(s)') theme_entrypoints = pkg_resources.iter_entry_points('hieroglyph.theme') themes = [t.load() for t in theme_entrypoints] msg = '\nAvailable themes:\n\n' for theme in themes: msg += '\n'.join([bold(theme['name']), theme['desc'], '', '']) msg += 'Which theme would you like to use?' print(msg) sphinx.quickstart.do_prompt(d, 'slide_theme', 'Slide Theme', themes[0][ 'name'], sphinx.quickstart.choice(*[t['name'] for t in themes])) print('') sphinx.quickstart.ask_user(d)
ask_user
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(d):\n msg = bold('Welcome to the Hieroglyph %s quickstart utility.') % (version\n (),)\n print(msg)\n msg = \"\"\"\nThis will ask questions for...
def x(self): return {'title': self.title, 'level': self.level, 'content': self. content, 'classes': self.classes, 'slide_classes': self. _filter_classes(exclude='content-'), 'content_classes': self. _filter_classes(include='content-'), 'slide_number': self. slide_number, 'config': self._translator.builder.config, 'id': self.id}
get_slide_context
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return {'title': self.title, 'level': self.level, 'content': self.\n content, 'classes': self.classes, 'slide_classes': self.\n _filt...
def x(self, slide_no): if self.builder.config.slide_numbers: self.body.append('\n<div class="slide-no">%s</div>\n' % (slide_no,))
add_slide_number
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, slide_no):\n if self.builder.config.slide_numbers:\n self.body.append('\\n<div class=\"slide-no\">%s</div>\\n' % (slide_no,))\n", "role": "...
def x(self, slide_no): if self.builder.config.slide_footer: self.body.append('\n<div class="slide-footer">%s</div>\n' % (self. builder.config.slide_footer,))
add_slide_footer
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, slide_no):\n if self.builder.config.slide_footer:\n self.body.append('\\n<div class=\"slide-footer\">%s</div>\\n' % (self.\n builder...
def x(app): from hieroglyph import writer if app.builder.name not in HTML_BUILDERS: return if app.config.slide_link_html_to_slides: app.builder.templates.loaders.append(SphinxFileSystemLoader(os.path .join(os.path.dirname(__file__), 'themes', 'slides'))) if not app.config.html_sidebars: app.config.html_sidebars = {'**': ['localtoc.html', 'relations.html', 'sourcelink.html', SLIDELINK_TEMPLATE, 'searchbox.html']} else: included = False for glob, templates in app.config.html_sidebars: if SLIDELINK_TEMPLATE in templates: included = True break if not included: app.config.html_sidebars.setdefault('**', []).append( SLIDELINK_TEMPLATE) if app.config.slide_link_html_sections_to_slides: if sphinx.version_info >= (1, 6, 0): override_translator = type('SlideLinkTranslator', (app.builder. get_translator_class(), object), {'depart_title': writer. depart_title}) app.set_translator(app.builder, override_translator) else: app.builder.translator_class = type('SlideLinkTranslator', (app .builder.translator_class, object), {'depart_title': writer .depart_title})
inspect_config
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(app):\n from hieroglyph import writer\n if app.builder.name not in HTML_BUILDERS:\n return\n if app.config.slide_link_html_to_slides:\n ...
def x(builder, pagename=None): return builder.get_relative_uri(pagename or builder.current_docname, os .path.join(builder.app.config.slide_relative_path, pagename or builder.current_docname))
slide_path
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(builder, pagename=None):\n return builder.get_relative_uri(pagename or builder.current_docname, os\n .path.join(builder.app.config.slide_relative_pat...
def x(builder, pagename=None): return builder.get_relative_uri(pagename or builder.current_docname, os .path.join(builder.app.config.slide_html_relative_path, pagename or builder.current_docname))
html_path
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(builder, pagename=None):\n return builder.get_relative_uri(pagename or builder.current_docname, os\n .path.join(builder.app.config.slide_html_relativ...
def x(app, pagename, templatename, context, doctree): context['show_slidelink' ] = app.config.slide_link_html_to_slides and hasattr(app.builder, 'get_outfilename') if context['show_slidelink']: context['slide_path'] = slide_path(app.builder, pagename)
add_link
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(app, pagename, templatename, context, doctree):\n context['show_slidelink'\n ] = app.config.slide_link_html_to_slides and hasattr(app.builder,\n ...
def x(self, themename, themeoptions): self._theme_stack.append((self.theme, self.theme_options)) theme_factory = HTMLThemeFactory(self.app) theme_factory.load_additional_themes(self.get_builtin_theme_dirs() + self.config.slide_theme_path) self.theme = theme_factory.create(themename) self.theme_options = themeoptions.copy() self.templates.init(self, self.theme) self.templates.environment.filters['json'] = json.dumps if self.theme not in self._additional_themes: self._additional_themes.append(self.theme)
apply_theme
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, themename, themeoptions):\n self._theme_stack.append((self.theme, self.theme_options))\n theme_factory = HTMLThemeFactory(self.app)\n theme_fact...
def x(self, doctree): super(AbstractSlideBuilder, self).post_process_images(doctree) relative_base = ['..'] * doctree.attributes.get('source')[len(self. srcdir) + 1:].count('/') for node in doctree.traverse(nodes.image): if node.get('candidates') is None: node['candidates'] = '*', if node['uri'].startswith(self.outdir): node['uri'] = '/'.join(relative_base + [node['uri'][len(self. outdir) + 1:]])
post_process_images
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, doctree):\n super(AbstractSlideBuilder, self).post_process_images(doctree)\n relative_base = ['..'] * doctree.attributes.get('source')[len(self.\n ...
def x(section): metadata = {} metadata_lines = section.split('\n') for line in metadata_lines: colon_index = line.find(':') if colon_index != -1: key = line[:colon_index].strip() val = line[colon_index + 1:].strip() metadata[key] = val return metadata
parse_metadata
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(section):\n metadata = {}\n metadata_lines = section.split('\\n')\n for line in metadata_lines:\n colon_index = line.find(':')\n if colo...
def x(html, metadata): if metadata.get('build_lists') and metadata['build_lists'] == 'true': html = html.replace('<ul>', '<ul class="build">') html = html.replace('<ol>', '<ol class="build">') return html
postprocess_html
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(html, metadata):\n if metadata.get('build_lists') and metadata['build_lists'] == 'true':\n html = html.replace('<ul>', '<ul class=\"build\">')\n ...
def x(linter, checker_method, augmentation): if sys.version_info[0] <= 2: checker = get_checker(linter, checker_method.im_class) else: try: checker = get_checker(linter, checker_method.__self__.__class__) except AttributeError: checker = get_checker(linter, get_class(checker_method. __module__, checker_method.__qualname__)) old_method = getattr(checker, checker_method.__name__) def x(node): def x(): old_method(node) augmentation(chain, node) setattr(checker, checker_method.__name__, augment_func)
augment_visit
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(linter, checker_method, augmentation):\n if sys.version_info[0] <= 2:\n checker = get_checker(linter, checker_method.im_class)\n else:\n tr...
def x(linter, checker_method, message_id_or_symbol, test_func): msgs_store = getattr(linter, 'msgs_store', linter) def x(message_id_or_symbol): if hasattr(msgs_store, 'check_message_id'): return [msgs_store.check_message_id(message_id_or_symbol)] elif hasattr(msgs_store, 'get_message_definition'): return [msgs_store.get_message_definition(message_id_or_symbol)] elif hasattr(msgs_store, 'get_message_definitions'): return msgs_store.get_message_definitions(message_id_or_symbol) else: raise ValueError( 'pylint.utils.MessagesStore does not have a get_message_definition(s) method' ) try: pylint_messages = get_message_definitions(message_id_or_symbol) symbols = [symbol for pylint_message in pylint_messages for symbol in (pylint_message.msgid, pylint_message.symbol) if symbol is not None ] except UnknownMessage: symbols = [message_id_or_symbol] def x(chain, node): with Suppress(linter) as s: if test_func(node): s.suppress(*symbols) chain() augment_visit(linter, checker_method, do_suppress)
suppress_message
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(linter, checker_method, message_id_or_symbol, test_func):\n msgs_store = getattr(linter, 'msgs_store', linter)\n\n def x(message_id_or_symbol):\n ...
def x(self, hash_values): q = """SELECT threat_type,platform_type,threat_entry_type, expires_at < current_timestamp AS has_expired FROM full_hash WHERE value IN ({}) """ output = [] with self.get_cursor() as dbc: placeholders = ','.join(['?'] * len(hash_values)) dbc.execute(q.format(placeholders), [sqlite3.Binary(hv) for hv in hash_values]) for h in dbc.fetchall(): threat_type, platform_type, threat_entry_type, has_expired = h threat_list = ThreatList(threat_type, platform_type, threat_entry_type) output.append((threat_list, has_expired)) return output
lookup_full_hashes
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, hash_values):\n q = \"\"\"SELECT threat_type,platform_type,threat_entry_type, expires_at < current_timestamp AS has_expired\n FROM full...
def x(self, cues): q = """SELECT value, MAX(negative_expires_at < current_timestamp) AS negative_cache_expired FROM hash_prefix WHERE cue IN ({}) GROUP BY 1 """ output = [] with self.get_cursor() as dbc: dbc.execute(q.format(','.join(['?'] * len(cues))), [sqlite3.Binary( cue) for cue in cues]) for h in dbc.fetchall(): value, negative_cache_expired = h output.append((bytes(value), negative_cache_expired)) return output
lookup_hash_prefix
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, cues):\n q = \"\"\"SELECT value, MAX(negative_expires_at < current_timestamp) AS negative_cache_expired\n FROM hash_prefix WHERE cue IN...
def x(self, threat_list, hash_value, cache_duration, malware_threat_type): log.info('Storing full hash %s to list %s with cache duration %s', to_hex(hash_value), str(threat_list), cache_duration) qi = """INSERT OR IGNORE INTO full_hash (value, threat_type, platform_type, threat_entry_type, malware_threat_type, downloaded_at) VALUES (?, ?, ?, ?, ?, current_timestamp) """ qu = ( "UPDATE full_hash SET expires_at=datetime(current_timestamp, '+{} SECONDS') WHERE value=? AND threat_type=? AND platform_type=? AND threat_entry_type=?" ) i_parameters = [sqlite3.Binary(hash_value), threat_list.threat_type, threat_list.platform_type, threat_list.threat_entry_type, malware_threat_type] u_parameters = [sqlite3.Binary(hash_value), threat_list.threat_type, threat_list.platform_type, threat_list.threat_entry_type] with self.get_cursor() as dbc: dbc.execute(qi, i_parameters) dbc.execute(qu.format(int(cache_duration)), u_parameters)
store_full_hash
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, threat_list, hash_value, cache_duration, malware_threat_type):\n log.info('Storing full hash %s to list %s with cache duration %s',\n to_hex(ha...
def x(self, keep_expired_for=60 * 60 * 12): q = """DELETE FROM full_hash WHERE expires_at < datetime(current_timestamp, '-{} SECONDS') """ log.info('Cleaning up full_hash entries expired more than {} seconds ago.' .format(keep_expired_for)) with self.get_cursor() as dbc: dbc.execute(q.format(int(keep_expired_for)))
cleanup_full_hashes
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, keep_expired_for=60 * 60 * 12):\n q = \"\"\"DELETE FROM full_hash WHERE expires_at < datetime(current_timestamp, '-{} SECONDS')\n \"\"\"\n l...
def x(self): q = 'SELECT threat_type,platform_type,threat_entry_type FROM threat_list' output = [] with self.get_cursor() as dbc: dbc.execute(q) for h in dbc.fetchall(): threat_type, platform_type, threat_entry_type = h threat_list = ThreatList(threat_type, platform_type, threat_entry_type) output.append(threat_list) return output
get_threat_lists
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n q = 'SELECT threat_type,platform_type,threat_entry_type FROM threat_list'\n output = []\n with self.get_cursor() as dbc:\n dbc.execute...
def x(self): q = ( 'SELECT threat_type,platform_type,threat_entry_type,client_state FROM threat_list' ) output = {} with self.get_cursor() as dbc: dbc.execute(q) for h in dbc.fetchall(): threat_type, platform_type, threat_entry_type, client_state = h threat_list_tuple = threat_type, platform_type, threat_entry_type output[threat_list_tuple] = client_state return output
get_client_state
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n q = (\n 'SELECT threat_type,platform_type,threat_entry_type,client_state FROM threat_list'\n )\n output = {}\n with self.get_cu...
def x(self, threat_list): q = """INSERT OR IGNORE INTO threat_list (threat_type, platform_type, threat_entry_type, timestamp) VALUES (?, ?, ?, current_timestamp) """ params = [threat_list.threat_type, threat_list.platform_type, threat_list.threat_entry_type] with self.get_cursor() as dbc: dbc.execute(q, params)
add_threat_list
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, threat_list):\n q = \"\"\"INSERT OR IGNORE INTO threat_list\n (threat_type, platform_type, threat_entry_type, timestamp)\n ...
def x(self, threat_list): log.info('Deleting cached threat list "{}"'.format(repr(threat_list))) q = """DELETE FROM threat_list WHERE threat_type=? AND platform_type=? AND threat_entry_type=? """ params = [threat_list.threat_type, threat_list.platform_type, threat_list.threat_entry_type] with self.get_cursor() as dbc: dbc.execute(q, params)
delete_threat_list
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, threat_list):\n log.info('Deleting cached threat list \"{}\"'.format(repr(threat_list)))\n q = \"\"\"DELETE FROM threat_list\n W...
def x(self, threat_list): q = """SELECT value FROM hash_prefix WHERE threat_type=? AND platform_type=? AND threat_entry_type=? ORDER BY value """ params = [threat_list.threat_type, threat_list.platform_type, threat_list.threat_entry_type] with self.get_cursor() as dbc: dbc.execute(q, params) all_hashes = b''.join(bytes(h[0]) for h in dbc.fetchall()) checksum = hashlib.sha256(all_hashes).digest() return checksum
hash_prefix_list_checksum
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, threat_list):\n q = \"\"\"SELECT value FROM hash_prefix\n WHERE threat_type=? AND platform_type=? AND threat_entry_type=?\n ...
def x(self, threat_list, indices): batch_size = 40 q = """DELETE FROM hash_prefix WHERE threat_type=? AND platform_type=? AND threat_entry_type=? AND value IN ({}) """ prefixes_to_remove = self.get_hash_prefix_values_to_remove(threat_list, indices) with self.get_cursor() as dbc: for i in range(0, len(prefixes_to_remove), batch_size): remove_batch = prefixes_to_remove[i:i + batch_size] params = [threat_list.threat_type, threat_list.platform_type, threat_list.threat_entry_type] + [sqlite3.Binary(b) for b in remove_batch] dbc.execute(q.format(','.join(['?'] * len(remove_batch))), params)
remove_hash_prefix_indices
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, threat_list, indices):\n batch_size = 40\n q = \"\"\"DELETE FROM hash_prefix\n WHERE threat_type=? AND platform_type=? AND threat_en...
def x(self): q = 'SELECT distinct value from hash_prefix' output = [] with self.get_cursor() as dbc: dbc.execute(q) output = [bytes(r[0]) for r in dbc.fetchall()] return output
dump_hash_prefix_values
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n q = 'SELECT distinct value from hash_prefix'\n output = []\n with self.get_cursor() as dbc:\n dbc.execute(q)\n output = [bytes(...
def x(tk): used = False try: while True: try: method, args, kwargs, response_queue = (tk.tk._event_queue. get_nowait()) except queue.Empty: break else: used = True if tk.tk._debug >= 2: print('Calling event from main thread:', method. __name__, args, kwargs) try: response_queue.put((False, method(*args, **kwargs))) except SystemExit: raise except Exception: from sys import exc_info ex_type, ex_value, ex_tb = exc_info() response_queue.put((True, (ex_type, ex_value, ex_tb))) finally: if used: tk.after_idle(_check_events, tk) else: tk.after(tk.tk._check_period, _check_events, tk)
check_events
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(tk):\n used = False\n try:\n while True:\n try:\n method, args, kwargs, response_queue = (tk.tk._event_queue.\n ...
def x(self): try: self.storage.cleanup_full_hashes() self.storage.commit() self._sync_threat_lists() self.storage.commit() self._sync_hash_prefix_cache() except Exception: self.storage.rollback() raise
update_hash_prefix_cache
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n try:\n self.storage.cleanup_full_hashes()\n self.storage.commit()\n self._sync_threat_lists()\n self.storage.commit()\n...
def x(self, hash_prefixes): client_state = self.storage.get_client_state() self.api_client.fair_use_delay() fh_response = self.api_client.get_full_hashes(hash_prefixes, client_state) for m in fh_response.get('matches', []): threat_list = ThreatList(m['threatType'], m['platformType'], m[ 'threatEntryType']) hash_value = b64decode(m['threat']['hash']) cache_duration = int(m['cacheDuration'].rstrip('s')) malware_threat_type = None for metadata in m['threatEntryMetadata'].get('entries', []): k = b64decode(metadata['key']) v = b64decode(metadata['value']) if k == 'malware_threat_type': malware_threat_type = v self.storage.store_full_hash(threat_list, hash_value, cache_duration, malware_threat_type) negative_cache_duration = int(fh_response['negativeCacheDuration']. rstrip('s')) for prefix_value in hash_prefixes: self.storage.update_hash_prefix_expiration(prefix_value, negative_cache_duration)
sync_full_hashes
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, hash_prefixes):\n client_state = self.storage.get_client_state()\n self.api_client.fair_use_delay()\n fh_response = self.api_client.get_full_has...
def x(self, url): if type(url) is not str: url = url.encode('utf8') if not url.strip(): raise ValueError('Empty input string.') url_hashes = URL(url).hashes try: list_names = self._lookup_hashes(url_hashes) self.storage.commit() except Exception: self.storage.rollback() raise if list_names: return list_names return None
lookup_url
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, url):\n if type(url) is not str:\n url = url.encode('utf8')\n if not url.strip():\n raise ValueError('Empty input string.')\n url_...
def x(self, full_hashes): full_hashes = list(full_hashes) cues = [fh[0:4] for fh in full_hashes] result = [] matching_prefixes = {} matching_full_hashes = set() is_potential_threat = False for hash_prefix, negative_cache_expired in self.storage.lookup_hash_prefix( cues): for full_hash in full_hashes: if full_hash.startswith(hash_prefix): is_potential_threat = True matching_prefixes[hash_prefix] = matching_prefixes.get( hash_prefix, False) or negative_cache_expired matching_full_hashes.add(full_hash) if not is_potential_threat: return [] matching_expired_threat_lists = set() for threat_list, has_expired in self.storage.lookup_full_hashes( matching_full_hashes): if has_expired: matching_expired_threat_lists.add(threat_list) else: result.append(threat_list) if result: return result if len(matching_expired_threat_lists) == 0 and sum(map(int, matching_prefixes.values())) == 0: log.info('Negative cache hit.') return [] self._sync_full_hashes(matching_prefixes.keys()) for threat_list, has_expired in self.storage.lookup_full_hashes( matching_full_hashes): if not has_expired: result.append(threat_list) return result
lookup_hashes
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, full_hashes):\n full_hashes = list(full_hashes)\n cues = [fh[0:4] for fh in full_hashes]\n result = []\n matching_prefixes = {}\n matching...
def x(self): response = self.service.threatLists().list().execute() self.set_wait_duration(response.get('minimumWaitDuration')) return response['threatLists']
get_threats_lists
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n response = self.service.threatLists().list().execute()\n self.set_wait_duration(response.get('minimumWaitDuration'))\n return response['threa...
def x(self, client_state): request_body = {'client': {'clientId': self.client_id, 'clientVersion': self.client_version}, 'listUpdateRequests': []} for (threat_type, platform_type, threat_entry_type ), current_state in client_state.items(): request_body['listUpdateRequests'].append({'threatType': threat_type, 'platformType': platform_type, 'threatEntryType': threat_entry_type, 'state': current_state, 'constraints': { 'supportedCompressions': ['RAW']}}) response = self.service.threatListUpdates().fetch(body=request_body ).execute() self.set_wait_duration(response.get('minimumWaitDuration')) return response['listUpdateResponses']
get_threats_update
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, client_state):\n request_body = {'client': {'clientId': self.client_id, 'clientVersion':\n self.client_version}, 'listUpdateRequests': []}\n ...
def x(self, prefixes, client_state): request_body = {'client': {'clientId': self.client_id, 'clientVersion': self.client_version}, 'clientStates': [], 'threatInfo': { 'threatTypes': [], 'platformTypes': [], 'threatEntryTypes': [], 'threatEntries': []}} for prefix in prefixes: request_body['threatInfo']['threatEntries'].append({'hash': b64encode(prefix).decode()}) for (threatType, platformType, threatEntryType ), clientState in client_state.items(): request_body['clientStates'].append(clientState) if threatType not in request_body['threatInfo']['threatTypes']: request_body['threatInfo']['threatTypes'].append(threatType) if platformType not in request_body['threatInfo']['platformTypes']: request_body['threatInfo']['platformTypes'].append(platformType) if threatEntryType not in request_body['threatInfo']['threatEntryTypes' ]: request_body['threatInfo']['threatEntryTypes'].append( threatEntryType) response = self.service.fullHashes().find(body=request_body).execute() self.set_wait_duration(response.get('minimumWaitDuration')) return response
get_full_hashes
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, prefixes, client_state):\n request_body = {'client': {'clientId': self.client_id, 'clientVersion':\n self.client_version}, 'clientStates': [], ...
def x(self): for url_variant in self.url_permutations(self.canonical): url_hash = self.digest(url_variant) yield url_hash
hashes
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n for url_variant in self.url_permutations(self.canonical):\n url_hash = self.digest(url_variant)\n yield url_hash\n", "role": "use...
def x(self): def x(u): uu = urllib.unquote(u) if uu == u: return uu else: return full_unescape(uu) def x(u): uu = urlparse.unquote_to_bytes(u) if uu == u: return uu else: return full_unescape_to_bytes(uu) def x(s): safe_chars = '!"$&\'()*+,-./:;<=>?@[\\]^_`{|}~' return urllib.quote(s, safe=safe_chars) url = self.url.strip() url = url.replace(b'\n', b'').replace(b'\r', b'').replace(b'\t', b'') url = url.split(b'#', 1)[0] if url.startswith(b'//'): url = b'http:' + url if len(url.split(b'://')) <= 1: url = b'http://' + url if self.__py3 and type(url) is bytes: url = quote(full_unescape_to_bytes(url)) else: url = quote(full_unescape(url)) url_parts = urlparse.urlsplit(url) if not url_parts[0]: url = 'http://{}'.format(url) url_parts = urlparse.urlsplit(url) protocol = url_parts.scheme if self.__py3: host = full_unescape_to_bytes(url_parts.hostname) path = full_unescape_to_bytes(url_parts.path) else: host = full_unescape(url_parts.hostname) path = full_unescape(url_parts.path) query = url_parts.query if not query and '?' not in url: query = None if not path: path = b'/' has_trailing_slash = path[-1:] == b'/' path = posixpath.normpath(path).replace(b'//', b'/') if has_trailing_slash and path[-1:] != b'/': path = path + b'/' port = url_parts.port host = host.strip(b'.') host = re.sub(b'\\.+', b'.', host).lower() if host.isdigit(): try: host = socket.inet_ntoa(struct.pack('!I', int(host))) except Exception: pass elif host.startswith(b'0x') and b'.' not in host: try: host = socket.inet_ntoa(struct.pack('!I', int(host, 16))) except Exception: pass quoted_path = quote(path) quoted_host = quote(host) if port is not None: quoted_host = '{}:{}'.format(quoted_host, port) canonical_url = '{}://{}{}'.format(protocol, quoted_host, quoted_path) if query is not None: canonical_url = '{}?{}'.format(canonical_url, query) return canonical_url
canonical
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n\n def x(u):\n uu = urllib.unquote(u)\n if uu == u:\n return uu\n else:\n return full_unescape(uu)\n\n ...
def x(url): def x(host): if re.match('\\d+\\.\\d+\\.\\d+\\.\\d+', host): yield host return parts = host.split('.') l = min(len(parts), 5) if l > 4: yield host for i in range(l - 1): yield '.'.join(parts[i - l:]) def x(path): yield path query = None if '?' in path: path, query = path.split('?', 1) if query is not None: yield path path_parts = path.split('/')[0:-1] curr_path = '' for i in range(min(4, len(path_parts))): curr_path = curr_path + path_parts[i] + '/' yield curr_path protocol, address_str = urllib.splittype(url) host, path = urllib.splithost(address_str) user, host = urllib.splituser(host) host, port = urllib.splitport(host) host = host.strip('/') seen_permutations = set() for h in url_host_permutations(host): for p in url_path_permutations(path): u = '{}{}'.format(h, p) if u not in seen_permutations: yield u seen_permutations.add(u)
url_permutations
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(url):\n\n def x(host):\n if re.match('\\\\d+\\\\.\\\\d+\\\\.\\\\d+\\\\.\\\\d+', host):\n yield host\n return\n parts = h...
def x(v1, v2): def x(v): v = re.sub('(\\.0+)*$', '', v) result = [] for part in v.split('.'): m = re.match('^(\\d+)$', part) if m: result.append(int(m.group(1))) continue m = re.match('^(\\d+)([a-zA-Z]+)$', part) if m: result.append(int(m.group(1))) result.append(m.group(2)) continue m = re.match('^(\\d+)([a-zA-Z]+)(\\d+)$', part) if m: result.append(int(m.group(1))) result.append(m.group(2)) result.append(int(m.group(3))) continue return tuple(result) n1 = normalize(v1) n2 = normalize(v2) return (n1 > n2) - (n1 < n2)
compare_versions
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(v1, v2):\n\n def x(v):\n v = re.sub('(\\\\.0+)*$', '', v)\n result = []\n for part in v.split('.'):\n m = re.match('^(\\\\d+...
def x(spec): m = re.search('([<>=]?=?)?\\s*([0-9.a-zA-Z]+)', spec) return m.group(2), m.group(1)
split_version_specifier
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(spec):\n m = re.search('([<>=]?=?)?\\\\s*([0-9.a-zA-Z]+)', spec)\n return m.group(2), m.group(1)\n", "role": "user" }, { "content": "split_ve...
def x(package): pkg_config_exe = os.environ.get('PKG_CONFIG', None) or 'pkg-config' cmd = '{0} --exists {1}'.format(pkg_config_exe, package).split() return call(cmd) == 0
exists
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(package):\n pkg_config_exe = os.environ.get('PKG_CONFIG', None) or 'pkg-config'\n cmd = '{0} --exists {1}'.format(pkg_config_exe, package).split()\n r...
def x(package, static=False): _raise_if_not_exists(package) return _query(package, *_build_options('--libs', static=static))
libs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(package, static=False):\n _raise_if_not_exists(package)\n return _query(package, *_build_options('--libs', static=static))\n", "role": "user" }, ...
def x(package): _raise_if_not_exists(package) result = _query(package, '--print-variables') names = (x.strip() for x in result.split('\n') if x != '') return dict((x, _query(package, '--variable={0}'.format(x)).strip()) for x in names)
variables
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(package):\n _raise_if_not_exists(package)\n result = _query(package, '--print-variables')\n names = (x.strip() for x in result.split('\\n') if x != ''...