id_within_dataset
int64
1
55.5k
snippet
stringlengths
19
14.2k
tokens
listlengths
6
1.63k
nl
stringlengths
6
352
split_within_dataset
stringclasses
1 value
is_duplicated
bool
2 classes
48,559
def _accumulateFrequencyCounts(values, freqCounts=None): values = numpy.array(values) numEntries = (values.max() + 1) if (freqCounts is not None): numEntries = max(numEntries, freqCounts.size) if (freqCounts is not None): if (freqCounts.size != numEntries): newCounts = numpy.zeros(numEntries, dtype='int32') newCounts[0:freqCounts.size] = freqCounts else: newCounts = freqCounts else: newCounts = numpy.zeros(numEntries, dtype='int32') for v in values: newCounts[v] += 1 return newCounts
[ "def", "_accumulateFrequencyCounts", "(", "values", ",", "freqCounts", "=", "None", ")", ":", "values", "=", "numpy", ".", "array", "(", "values", ")", "numEntries", "=", "(", "values", ".", "max", "(", ")", "+", "1", ")", "if", "(", "freqCounts", "is", "not", "None", ")", ":", "numEntries", "=", "max", "(", "numEntries", ",", "freqCounts", ".", "size", ")", "if", "(", "freqCounts", "is", "not", "None", ")", ":", "if", "(", "freqCounts", ".", "size", "!=", "numEntries", ")", ":", "newCounts", "=", "numpy", ".", "zeros", "(", "numEntries", ",", "dtype", "=", "'int32'", ")", "newCounts", "[", "0", ":", "freqCounts", ".", "size", "]", "=", "freqCounts", "else", ":", "newCounts", "=", "freqCounts", "else", ":", "newCounts", "=", "numpy", ".", "zeros", "(", "numEntries", ",", "dtype", "=", "'int32'", ")", "for", "v", "in", "values", ":", "newCounts", "[", "v", "]", "+=", "1", "return", "newCounts" ]
accumulate a list of values values into the frequency counts freqcounts .
train
true
48,564
def is_spam(request, form, url): for spam_filter_path in settings.SPAM_FILTERS: spam_filter = import_dotted_path(spam_filter_path) if spam_filter(request, form, url): return True
[ "def", "is_spam", "(", "request", ",", "form", ",", "url", ")", ":", "for", "spam_filter_path", "in", "settings", ".", "SPAM_FILTERS", ":", "spam_filter", "=", "import_dotted_path", "(", "spam_filter_path", ")", "if", "spam_filter", "(", "request", ",", "form", ",", "url", ")", ":", "return", "True" ]
main entry point for spam handling - called from the comment view and page processor for mezzanine .
train
true
48,565
def mask_color(clip, color=[0, 0, 0], thr=0, s=1): hill = (lambda x: ((1.0 * (x != 0)) if (thr == 0) else ((x ** s) / ((thr ** s) + (x ** s))))) color = np.array(color) flim = (lambda im: hill(np.sqrt(((im - color) ** 2).sum(axis=2)))) mask = clip.fl_image(flim) mask.ismask = True newclip = clip.set_mask(mask) return newclip
[ "def", "mask_color", "(", "clip", ",", "color", "=", "[", "0", ",", "0", ",", "0", "]", ",", "thr", "=", "0", ",", "s", "=", "1", ")", ":", "hill", "=", "(", "lambda", "x", ":", "(", "(", "1.0", "*", "(", "x", "!=", "0", ")", ")", "if", "(", "thr", "==", "0", ")", "else", "(", "(", "x", "**", "s", ")", "/", "(", "(", "thr", "**", "s", ")", "+", "(", "x", "**", "s", ")", ")", ")", ")", ")", "color", "=", "np", ".", "array", "(", "color", ")", "flim", "=", "(", "lambda", "im", ":", "hill", "(", "np", ".", "sqrt", "(", "(", "(", "im", "-", "color", ")", "**", "2", ")", ".", "sum", "(", "axis", "=", "2", ")", ")", ")", ")", "mask", "=", "clip", ".", "fl_image", "(", "flim", ")", "mask", ".", "ismask", "=", "True", "newclip", "=", "clip", ".", "set_mask", "(", "mask", ")", "return", "newclip" ]
returns a new clip with a mask for transparency where the original clip is of the given color .
train
false
48,567
def _is_in_range(value, min_value, max_value): if (value is None): return False if ((not (min_value or max_value)) or (min_value == max_value == value)): return True if (((not min_value) or (value > min_value)) and ((max_value is None) or (value <= max_value))): return True return False
[ "def", "_is_in_range", "(", "value", ",", "min_value", ",", "max_value", ")", ":", "if", "(", "value", "is", "None", ")", ":", "return", "False", "if", "(", "(", "not", "(", "min_value", "or", "max_value", ")", ")", "or", "(", "min_value", "==", "max_value", "==", "value", ")", ")", ":", "return", "True", "if", "(", "(", "(", "not", "min_value", ")", "or", "(", "value", ">", "min_value", ")", ")", "and", "(", "(", "max_value", "is", "None", ")", "or", "(", "value", "<=", "max_value", ")", ")", ")", ":", "return", "True", "return", "False" ]
help function to check if the range matches with value if min_value is none the max_value determines if the range matches .
train
false
48,568
def is_pubkey(string): pgp_regex = '.*?(-----BEGIN PGP PUBLIC KEY BLOCK-----.*?-----END PGP PUBLIC KEY BLOCK-----).*' return re.match(pgp_regex, string, re.DOTALL)
[ "def", "is_pubkey", "(", "string", ")", ":", "pgp_regex", "=", "'.*?(-----BEGIN PGP PUBLIC KEY BLOCK-----.*?-----END PGP PUBLIC KEY BLOCK-----).*'", "return", "re", ".", "match", "(", "pgp_regex", ",", "string", ",", "re", ".", "DOTALL", ")" ]
verifies if string is a pubkey .
train
false
48,570
def update_comment_in_doc(doc): if (doc.communication_type not in (u'Comment', u'Communication')): return if ((doc.communication_type == u'Comment') and (doc.comment_type != u'Comment')): return def get_content(doc): return ((doc.content[:97] + u'...') if (len(doc.content) > 100) else doc.content) if (doc.reference_doctype and doc.reference_name and doc.content): _comments = get_comments_from_parent(doc) updated = False for c in _comments: if (c.get(u'name') == doc.name): c[u'comment'] = get_content(doc) updated = True if (not updated): _comments.append({u'comment': get_content(doc), u'by': (doc.sender or doc.owner), u'name': doc.name}) update_comments_in_parent(doc.reference_doctype, doc.reference_name, _comments)
[ "def", "update_comment_in_doc", "(", "doc", ")", ":", "if", "(", "doc", ".", "communication_type", "not", "in", "(", "u'Comment'", ",", "u'Communication'", ")", ")", ":", "return", "if", "(", "(", "doc", ".", "communication_type", "==", "u'Comment'", ")", "and", "(", "doc", ".", "comment_type", "!=", "u'Comment'", ")", ")", ":", "return", "def", "get_content", "(", "doc", ")", ":", "return", "(", "(", "doc", ".", "content", "[", ":", "97", "]", "+", "u'...'", ")", "if", "(", "len", "(", "doc", ".", "content", ")", ">", "100", ")", "else", "doc", ".", "content", ")", "if", "(", "doc", ".", "reference_doctype", "and", "doc", ".", "reference_name", "and", "doc", ".", "content", ")", ":", "_comments", "=", "get_comments_from_parent", "(", "doc", ")", "updated", "=", "False", "for", "c", "in", "_comments", ":", "if", "(", "c", ".", "get", "(", "u'name'", ")", "==", "doc", ".", "name", ")", ":", "c", "[", "u'comment'", "]", "=", "get_content", "(", "doc", ")", "updated", "=", "True", "if", "(", "not", "updated", ")", ":", "_comments", ".", "append", "(", "{", "u'comment'", ":", "get_content", "(", "doc", ")", ",", "u'by'", ":", "(", "doc", ".", "sender", "or", "doc", ".", "owner", ")", ",", "u'name'", ":", "doc", ".", "name", "}", ")", "update_comments_in_parent", "(", "doc", ".", "reference_doctype", ",", "doc", ".", "reference_name", ",", "_comments", ")" ]
updates _comments property in parent document .
train
false
48,571
@step(u'The "([^"]*)" problem displays a "([^"]*)" answer') def assert_problem_has_answer(step, problem_type, answer_class): assert (answer_class in ['correct', 'incorrect', 'blank']) assert (problem_type in PROBLEM_DICT) problem_has_answer(world.scenario_dict['COURSE'].number, problem_type, answer_class)
[ "@", "step", "(", "u'The \"([^\"]*)\" problem displays a \"([^\"]*)\" answer'", ")", "def", "assert_problem_has_answer", "(", "step", ",", "problem_type", ",", "answer_class", ")", ":", "assert", "(", "answer_class", "in", "[", "'correct'", ",", "'incorrect'", ",", "'blank'", "]", ")", "assert", "(", "problem_type", "in", "PROBLEM_DICT", ")", "problem_has_answer", "(", "world", ".", "scenario_dict", "[", "'COURSE'", "]", ".", "number", ",", "problem_type", ",", "answer_class", ")" ]
assert that the problem is displaying a particular answer .
train
false
48,572
def approximate_fst(desired_fst, simulated_fst, parameter_fst, max_run_fst=1, min_run_fst=0, limit=0.005): if (abs((simulated_fst - desired_fst)) < limit): return (parameter_fst, max_run_fst, min_run_fst) if (simulated_fst > desired_fst): max_run_fst = parameter_fst next_parameter_fst = ((min_run_fst + parameter_fst) / 2) else: min_run_fst = parameter_fst next_parameter_fst = ((max_run_fst + parameter_fst) / 2) return (next_parameter_fst, max_run_fst, min_run_fst)
[ "def", "approximate_fst", "(", "desired_fst", ",", "simulated_fst", ",", "parameter_fst", ",", "max_run_fst", "=", "1", ",", "min_run_fst", "=", "0", ",", "limit", "=", "0.005", ")", ":", "if", "(", "abs", "(", "(", "simulated_fst", "-", "desired_fst", ")", ")", "<", "limit", ")", ":", "return", "(", "parameter_fst", ",", "max_run_fst", ",", "min_run_fst", ")", "if", "(", "simulated_fst", ">", "desired_fst", ")", ":", "max_run_fst", "=", "parameter_fst", "next_parameter_fst", "=", "(", "(", "min_run_fst", "+", "parameter_fst", ")", "/", "2", ")", "else", ":", "min_run_fst", "=", "parameter_fst", "next_parameter_fst", "=", "(", "(", "max_run_fst", "+", "parameter_fst", ")", "/", "2", ")", "return", "(", "next_parameter_fst", ",", "max_run_fst", ",", "min_run_fst", ")" ]
calculates the next fst attempt in order to approximate a desired fst .
train
false
48,573
def random_y_given_x(x): return (x + roll_a_die())
[ "def", "random_y_given_x", "(", "x", ")", ":", "return", "(", "x", "+", "roll_a_die", "(", ")", ")" ]
equally likely to be x + 1 .
train
false
48,574
def genBhLobe(x): N = 512 f = (((x * np.pi) * 2) / N) df = ((2 * np.pi) / N) y = np.zeros(x.size) consts = [0.35875, 0.48829, 0.14128, 0.01168] for m in range(0, 4): y += ((consts[m] / 2) * (sinc((f - (df * m)), N) + sinc((f + (df * m)), N))) y = ((y / N) / consts[0]) return y
[ "def", "genBhLobe", "(", "x", ")", ":", "N", "=", "512", "f", "=", "(", "(", "(", "x", "*", "np", ".", "pi", ")", "*", "2", ")", "/", "N", ")", "df", "=", "(", "(", "2", "*", "np", ".", "pi", ")", "/", "N", ")", "y", "=", "np", ".", "zeros", "(", "x", ".", "size", ")", "consts", "=", "[", "0.35875", ",", "0.48829", ",", "0.14128", ",", "0.01168", "]", "for", "m", "in", "range", "(", "0", ",", "4", ")", ":", "y", "+=", "(", "(", "consts", "[", "m", "]", "/", "2", ")", "*", "(", "sinc", "(", "(", "f", "-", "(", "df", "*", "m", ")", ")", ",", "N", ")", "+", "sinc", "(", "(", "f", "+", "(", "df", "*", "m", ")", ")", ",", "N", ")", ")", ")", "y", "=", "(", "(", "y", "/", "N", ")", "/", "consts", "[", "0", "]", ")", "return", "y" ]
generate the main lobe of a blackman-harris window x: bin positions to compute returns y: main lobe os spectrum of a blackman-harris window .
train
false
48,575
def _close_to_num(o1, o2, epsilon=5): delta = abs(((o2 - o1) * MUSECONDS_PER_DAY)) assert (delta < epsilon)
[ "def", "_close_to_num", "(", "o1", ",", "o2", ",", "epsilon", "=", "5", ")", ":", "delta", "=", "abs", "(", "(", "(", "o2", "-", "o1", ")", "*", "MUSECONDS_PER_DAY", ")", ")", "assert", "(", "delta", "<", "epsilon", ")" ]
assert that float ordinals *o1* and *o2* are within *epsilon* microseconds .
train
false
48,576
def make_attendance_records(student, student_name, status, course_schedule=None, student_batch=None, date=None): student_attendance = frappe.new_doc(u'Student Attendance') student_attendance.student = student student_attendance.student_name = student_name student_attendance.course_schedule = course_schedule student_attendance.student_batch = student_batch student_attendance.date = date student_attendance.status = status student_attendance.submit()
[ "def", "make_attendance_records", "(", "student", ",", "student_name", ",", "status", ",", "course_schedule", "=", "None", ",", "student_batch", "=", "None", ",", "date", "=", "None", ")", ":", "student_attendance", "=", "frappe", ".", "new_doc", "(", "u'Student Attendance'", ")", "student_attendance", ".", "student", "=", "student", "student_attendance", ".", "student_name", "=", "student_name", "student_attendance", ".", "course_schedule", "=", "course_schedule", "student_attendance", ".", "student_batch", "=", "student_batch", "student_attendance", ".", "date", "=", "date", "student_attendance", ".", "status", "=", "status", "student_attendance", ".", "submit", "(", ")" ]
creates attendance record .
train
false
48,578
def _get_window(start, end): from scipy.signal import hann window = (1 - np.r_[(hann(4)[:2], np.ones((np.abs((end - start)) - 4)), hann(4)[(-2):])].T) return window
[ "def", "_get_window", "(", "start", ",", "end", ")", ":", "from", "scipy", ".", "signal", "import", "hann", "window", "=", "(", "1", "-", "np", ".", "r_", "[", "(", "hann", "(", "4", ")", "[", ":", "2", "]", ",", "np", ".", "ones", "(", "(", "np", ".", "abs", "(", "(", "end", "-", "start", ")", ")", "-", "4", ")", ")", ",", "hann", "(", "4", ")", "[", "(", "-", "2", ")", ":", "]", ")", "]", ".", "T", ")", "return", "window" ]
return window which has length as much as parameter start - end .
train
false
48,579
@pytest.mark.parametrize('value', [['foo\\bar'], ['fo\xc3\xb6\r\nb\xc3\xa4r'], ['fo\xc3\xb6\\r\\nb\xc3\xa4r'], ['fo\xc3\xb6\r\n\\r\\nb\xc3\xa4r', 'b\xc3\xa4r\r\n\\r\\nb\xc3\xa4z'], ['nfo\xc3\xb6\nb\xc3\xa4r'], ['nfo\xc3\xb6\\nb\xc3\xa4r'], ['fo\xc3\xb6\n\\nb\xc3\xa4r', 'b\xc3\xa4r\n\\nb\xc3\xa4z']]) def test_multistringwidget_decompress_list_of_values(value): widget = MultiStringWidget() assert (widget.decompress(value) == value)
[ "@", "pytest", ".", "mark", ".", "parametrize", "(", "'value'", ",", "[", "[", "'foo\\\\bar'", "]", ",", "[", "'fo\\xc3\\xb6\\r\\nb\\xc3\\xa4r'", "]", ",", "[", "'fo\\xc3\\xb6\\\\r\\\\nb\\xc3\\xa4r'", "]", ",", "[", "'fo\\xc3\\xb6\\r\\n\\\\r\\\\nb\\xc3\\xa4r'", ",", "'b\\xc3\\xa4r\\r\\n\\\\r\\\\nb\\xc3\\xa4z'", "]", ",", "[", "'nfo\\xc3\\xb6\\nb\\xc3\\xa4r'", "]", ",", "[", "'nfo\\xc3\\xb6\\\\nb\\xc3\\xa4r'", "]", ",", "[", "'fo\\xc3\\xb6\\n\\\\nb\\xc3\\xa4r'", ",", "'b\\xc3\\xa4r\\n\\\\nb\\xc3\\xa4z'", "]", "]", ")", "def", "test_multistringwidget_decompress_list_of_values", "(", "value", ")", ":", "widget", "=", "MultiStringWidget", "(", ")", "assert", "(", "widget", ".", "decompress", "(", "value", ")", "==", "value", ")" ]
tests units multistringwidget decompresses a list of values .
train
false
48,580
def _local_server_get(url, session): request = HttpRequest() request.method = 'GET' request.session = session (view, args, kwargs) = resolve(url) response = view(request, *args, **kwargs) return response.content
[ "def", "_local_server_get", "(", "url", ",", "session", ")", ":", "request", "=", "HttpRequest", "(", ")", "request", ".", "method", "=", "'GET'", "request", ".", "session", "=", "session", "(", "view", ",", "args", ",", "kwargs", ")", "=", "resolve", "(", "url", ")", "response", "=", "view", "(", "request", ",", "*", "args", ",", "**", "kwargs", ")", "return", "response", ".", "content" ]
simulate a server-server get request for an in-process api .
train
false
48,581
def walk_outputs(object): out = [] if isinstance(object, dict): for (key, val) in sorted(object.items()): if isdefined(val): out.extend(walk_outputs(val)) elif isinstance(object, (list, tuple)): for val in object: if isdefined(val): out.extend(walk_outputs(val)) elif (isdefined(object) and isinstance(object, (str, bytes))): if (os.path.islink(object) or os.path.isfile(object)): out = [(filename, u'f') for filename in get_all_files(object)] elif os.path.isdir(object): out = [(object, u'd')] return out
[ "def", "walk_outputs", "(", "object", ")", ":", "out", "=", "[", "]", "if", "isinstance", "(", "object", ",", "dict", ")", ":", "for", "(", "key", ",", "val", ")", "in", "sorted", "(", "object", ".", "items", "(", ")", ")", ":", "if", "isdefined", "(", "val", ")", ":", "out", ".", "extend", "(", "walk_outputs", "(", "val", ")", ")", "elif", "isinstance", "(", "object", ",", "(", "list", ",", "tuple", ")", ")", ":", "for", "val", "in", "object", ":", "if", "isdefined", "(", "val", ")", ":", "out", ".", "extend", "(", "walk_outputs", "(", "val", ")", ")", "elif", "(", "isdefined", "(", "object", ")", "and", "isinstance", "(", "object", ",", "(", "str", ",", "bytes", ")", ")", ")", ":", "if", "(", "os", ".", "path", ".", "islink", "(", "object", ")", "or", "os", ".", "path", ".", "isfile", "(", "object", ")", ")", ":", "out", "=", "[", "(", "filename", ",", "u'f'", ")", "for", "filename", "in", "get_all_files", "(", "object", ")", "]", "elif", "os", ".", "path", ".", "isdir", "(", "object", ")", ":", "out", "=", "[", "(", "object", ",", "u'd'", ")", "]", "return", "out" ]
extract every file and directory from a python structure .
train
false
48,582
def install_reactor(explicit_reactor=None, verbose=False): import sys log = txaio.make_logger() if explicit_reactor: from twisted.application.reactors import installReactor log.info("Trying to install explicitly specified Twisted reactor '{reactor}'", reactor=explicit_reactor) try: installReactor(explicit_reactor) except: log.failure('Could not install Twisted reactor {reactor}\n{log_failure.value}', reactor=explicit_reactor) sys.exit(1) else: log.debug('Automatically choosing optimal Twisted reactor') install_optimal_reactor(verbose) from twisted.internet import reactor txaio.config.loop = reactor if verbose: from twisted.python.reflect import qual log.debug('Running Twisted reactor {reactor}', reactor=qual(reactor.__class__)) return reactor
[ "def", "install_reactor", "(", "explicit_reactor", "=", "None", ",", "verbose", "=", "False", ")", ":", "import", "sys", "log", "=", "txaio", ".", "make_logger", "(", ")", "if", "explicit_reactor", ":", "from", "twisted", ".", "application", ".", "reactors", "import", "installReactor", "log", ".", "info", "(", "\"Trying to install explicitly specified Twisted reactor '{reactor}'\"", ",", "reactor", "=", "explicit_reactor", ")", "try", ":", "installReactor", "(", "explicit_reactor", ")", "except", ":", "log", ".", "failure", "(", "'Could not install Twisted reactor {reactor}\\n{log_failure.value}'", ",", "reactor", "=", "explicit_reactor", ")", "sys", ".", "exit", "(", "1", ")", "else", ":", "log", ".", "debug", "(", "'Automatically choosing optimal Twisted reactor'", ")", "install_optimal_reactor", "(", "verbose", ")", "from", "twisted", ".", "internet", "import", "reactor", "txaio", ".", "config", ".", "loop", "=", "reactor", "if", "verbose", ":", "from", "twisted", ".", "python", ".", "reflect", "import", "qual", "log", ".", "debug", "(", "'Running Twisted reactor {reactor}'", ",", "reactor", "=", "qual", "(", "reactor", ".", "__class__", ")", ")", "return", "reactor" ]
install twisted reactor .
train
false
48,584
def cookie_date(epoch_seconds=None): rfcdate = formatdate(epoch_seconds) return ('%s-%s-%s GMT' % (rfcdate[:7], rfcdate[8:11], rfcdate[12:25]))
[ "def", "cookie_date", "(", "epoch_seconds", "=", "None", ")", ":", "rfcdate", "=", "formatdate", "(", "epoch_seconds", ")", "return", "(", "'%s-%s-%s GMT'", "%", "(", "rfcdate", "[", ":", "7", "]", ",", "rfcdate", "[", "8", ":", "11", "]", ",", "rfcdate", "[", "12", ":", "25", "]", ")", ")" ]
formats the time to ensure compatibility with netscapes cookie standard .
train
false
48,585
@not_implemented_for('undirected') def number_attracting_components(G): n = len(list(attracting_components(G))) return n
[ "@", "not_implemented_for", "(", "'undirected'", ")", "def", "number_attracting_components", "(", "G", ")", ":", "n", "=", "len", "(", "list", "(", "attracting_components", "(", "G", ")", ")", ")", "return", "n" ]
returns the number of attracting components in g .
train
false
48,586
def _find_guids(guid_string): guids = [] for found_guid in re.finditer(GUID_REGEX, guid_string): if found_guid.groups(): guids.append(found_guid.group(0).strip('{}')) return sorted(list(set(guids)))
[ "def", "_find_guids", "(", "guid_string", ")", ":", "guids", "=", "[", "]", "for", "found_guid", "in", "re", ".", "finditer", "(", "GUID_REGEX", ",", "guid_string", ")", ":", "if", "found_guid", ".", "groups", "(", ")", ":", "guids", ".", "append", "(", "found_guid", ".", "group", "(", "0", ")", ".", "strip", "(", "'{}'", ")", ")", "return", "sorted", "(", "list", "(", "set", "(", "guids", ")", ")", ")" ]
return the set of guids found in guid_string .
train
true
48,587
def parent_is_init(): if (os.getppid() == 1): return True return False
[ "def", "parent_is_init", "(", ")", ":", "if", "(", "os", ".", "getppid", "(", ")", "==", "1", ")", ":", "return", "True", "return", "False" ]
check if parent is init check if the parent process is init .
train
false
48,588
def is_extension_type(value): if is_categorical(value): return True elif is_sparse(value): return True elif is_datetimetz(value): return True return False
[ "def", "is_extension_type", "(", "value", ")", ":", "if", "is_categorical", "(", "value", ")", ":", "return", "True", "elif", "is_sparse", "(", "value", ")", ":", "return", "True", "elif", "is_datetimetz", "(", "value", ")", ":", "return", "True", "return", "False" ]
if we are a klass that is preserved by the internals these are internal klasses that we represent .
train
false
48,589
def add_doc_link(app, pagename, templatename, context, doctree): if ((not app.config.github_user) and app.config.github_project): return source_suffix = app.config.source_suffix source_suffix = (source_suffix if isinstance(source_suffix, basestring) else source_suffix[0]) context['github_link'] = (lambda mode='edit': make_github_link(app, ('doc/%s%s' % (pagename, source_suffix)), mode=mode))
[ "def", "add_doc_link", "(", "app", ",", "pagename", ",", "templatename", ",", "context", ",", "doctree", ")", ":", "if", "(", "(", "not", "app", ".", "config", ".", "github_user", ")", "and", "app", ".", "config", ".", "github_project", ")", ":", "return", "source_suffix", "=", "app", ".", "config", ".", "source_suffix", "source_suffix", "=", "(", "source_suffix", "if", "isinstance", "(", "source_suffix", ",", "basestring", ")", "else", "source_suffix", "[", "0", "]", ")", "context", "[", "'github_link'", "]", "=", "(", "lambda", "mode", "=", "'edit'", ":", "make_github_link", "(", "app", ",", "(", "'doc/%s%s'", "%", "(", "pagename", ",", "source_suffix", ")", ")", ",", "mode", "=", "mode", ")", ")" ]
add github_link function linking to the current page on github .
train
false
48,590
def search_python(python_code, template_name): retval = [] for (fn, content) in python_code: for (ln, line) in enumerate(content): if (((u'"%s"' % template_name) in line) or ((u"'%s'" % template_name) in line)): retval.append((fn, (ln + 1))) return retval
[ "def", "search_python", "(", "python_code", ",", "template_name", ")", ":", "retval", "=", "[", "]", "for", "(", "fn", ",", "content", ")", "in", "python_code", ":", "for", "(", "ln", ",", "line", ")", "in", "enumerate", "(", "content", ")", ":", "if", "(", "(", "(", "u'\"%s\"'", "%", "template_name", ")", "in", "line", ")", "or", "(", "(", "u\"'%s'\"", "%", "template_name", ")", "in", "line", ")", ")", ":", "retval", ".", "append", "(", "(", "fn", ",", "(", "ln", "+", "1", ")", ")", ")", "return", "retval" ]
searches python code for a template name .
train
false
48,591
@pytest.fixture(scope=u'function') def remove_test_dir(request): def fin_remove_test_dir(): if os.path.exists(u'test_copy_without_render'): utils.rmtree(u'test_copy_without_render') request.addfinalizer(fin_remove_test_dir)
[ "@", "pytest", ".", "fixture", "(", "scope", "=", "u'function'", ")", "def", "remove_test_dir", "(", "request", ")", ":", "def", "fin_remove_test_dir", "(", ")", ":", "if", "os", ".", "path", ".", "exists", "(", "u'test_copy_without_render'", ")", ":", "utils", ".", "rmtree", "(", "u'test_copy_without_render'", ")", "request", ".", "addfinalizer", "(", "fin_remove_test_dir", ")" ]
remove the folder that is created by the test .
train
false
48,592
def _clean_order(order): order = re.findall('\\([0-9 ]*?\\)', order) clean = (lambda x: tuple(map(int, re.sub('[()]', '', x).split(' ')))) if (len(order) > 1): (order, sorder) = map(clean, order) else: order = clean(order[0]) sorder = (0, 0, 0) return (order, sorder)
[ "def", "_clean_order", "(", "order", ")", ":", "order", "=", "re", ".", "findall", "(", "'\\\\([0-9 ]*?\\\\)'", ",", "order", ")", "clean", "=", "(", "lambda", "x", ":", "tuple", "(", "map", "(", "int", ",", "re", ".", "sub", "(", "'[()]'", ",", "''", ",", "x", ")", ".", "split", "(", "' '", ")", ")", ")", ")", "if", "(", "len", "(", "order", ")", ">", "1", ")", ":", "(", "order", ",", "sorder", ")", "=", "map", "(", "clean", ",", "order", ")", "else", ":", "order", "=", "clean", "(", "order", "[", "0", "]", ")", "sorder", "=", "(", "0", ",", "0", ",", "0", ")", "return", "(", "order", ",", "sorder", ")" ]
takes something like and returns a arma order .
train
false
48,593
def test_printwcs(): h = get_pkg_data_contents(u'spectra/orion-freq-1.hdr', encoding=u'binary') w = wcs.WCS(h) w.printwcs() h = get_pkg_data_contents(u'data/3d_cd.hdr', encoding=u'binary') w = wcs.WCS(h) w.printwcs()
[ "def", "test_printwcs", "(", ")", ":", "h", "=", "get_pkg_data_contents", "(", "u'spectra/orion-freq-1.hdr'", ",", "encoding", "=", "u'binary'", ")", "w", "=", "wcs", ".", "WCS", "(", "h", ")", "w", ".", "printwcs", "(", ")", "h", "=", "get_pkg_data_contents", "(", "u'data/3d_cd.hdr'", ",", "encoding", "=", "u'binary'", ")", "w", "=", "wcs", ".", "WCS", "(", "h", ")", "w", ".", "printwcs", "(", ")" ]
just make sure that it runs .
train
false
48,594
def get_switchport(port, module): command = 'show interface {0} switchport'.format(port) body = execute_show_command(command, module) try: body = execute_show_command(command, module)[0] except IndexError: body = [] if body: key_map = {'interface': 'interface', 'oper_mode': 'mode', 'switchport': 'switchport', 'access_vlan': 'access_vlan', 'access_vlan_name': 'access_vlan_name', 'native_vlan': 'native_vlan', 'native_vlan_name': 'native_vlan_name', 'trunk_vlans': 'trunk_vlans'} sp_table = body['TABLE_interface']['ROW_interface'] sp = apply_key_map(key_map, sp_table) return sp else: return {}
[ "def", "get_switchport", "(", "port", ",", "module", ")", ":", "command", "=", "'show interface {0} switchport'", ".", "format", "(", "port", ")", "body", "=", "execute_show_command", "(", "command", ",", "module", ")", "try", ":", "body", "=", "execute_show_command", "(", "command", ",", "module", ")", "[", "0", "]", "except", "IndexError", ":", "body", "=", "[", "]", "if", "body", ":", "key_map", "=", "{", "'interface'", ":", "'interface'", ",", "'oper_mode'", ":", "'mode'", ",", "'switchport'", ":", "'switchport'", ",", "'access_vlan'", ":", "'access_vlan'", ",", "'access_vlan_name'", ":", "'access_vlan_name'", ",", "'native_vlan'", ":", "'native_vlan'", ",", "'native_vlan_name'", ":", "'native_vlan_name'", ",", "'trunk_vlans'", ":", "'trunk_vlans'", "}", "sp_table", "=", "body", "[", "'TABLE_interface'", "]", "[", "'ROW_interface'", "]", "sp", "=", "apply_key_map", "(", "key_map", ",", "sp_table", ")", "return", "sp", "else", ":", "return", "{", "}" ]
gets current config of l2 switchport args: device : this is the device object of an nx-api enabled device using the device class within device .
train
false
48,595
def OutHeader(text, dash): n = (64 - len(text)) Output() Output('/* %s %s %s */', (dash * (n / 2)), text, (dash * (n - (n / 2)))) Output()
[ "def", "OutHeader", "(", "text", ",", "dash", ")", ":", "n", "=", "(", "64", "-", "len", "(", "text", ")", ")", "Output", "(", ")", "Output", "(", "'/* %s %s %s */'", ",", "(", "dash", "*", "(", "n", "/", "2", ")", ")", ",", "text", ",", "(", "dash", "*", "(", "n", "-", "(", "n", "/", "2", ")", ")", ")", ")", "Output", "(", ")" ]
output a header comment using a given dash character .
train
false
48,596
def format_exc(limit=None): try: (etype, value, tb) = sys.exc_info() return ''.join(traceback.format_exception(etype, value, tb, limit)) finally: etype = value = tb = None
[ "def", "format_exc", "(", "limit", "=", "None", ")", ":", "try", ":", "(", "etype", ",", "value", ",", "tb", ")", "=", "sys", ".", "exc_info", "(", ")", "return", "''", ".", "join", "(", "traceback", ".", "format_exception", "(", "etype", ",", "value", ",", "tb", ",", "limit", ")", ")", "finally", ":", "etype", "=", "value", "=", "tb", "=", "None" ]
like print_exc() but return a string .
train
true
48,597
def _aggregate(input, options, output, timeFieldName): aggregator = Aggregator(aggregationInfo=options, inputFields=input.getFields(), timeFieldName=timeFieldName) while True: inRecord = input.getNextRecord() print 'Feeding in: ', inRecord (outRecord, aggBookmark) = aggregator.next(record=inRecord, curInputBookmark=None) print 'Record out: ', outRecord if (outRecord is not None): output.appendRecord(outRecord, None) if ((inRecord is None) and (outRecord is None)): break
[ "def", "_aggregate", "(", "input", ",", "options", ",", "output", ",", "timeFieldName", ")", ":", "aggregator", "=", "Aggregator", "(", "aggregationInfo", "=", "options", ",", "inputFields", "=", "input", ".", "getFields", "(", ")", ",", "timeFieldName", "=", "timeFieldName", ")", "while", "True", ":", "inRecord", "=", "input", ".", "getNextRecord", "(", ")", "print", "'Feeding in: '", ",", "inRecord", "(", "outRecord", ",", "aggBookmark", ")", "=", "aggregator", ".", "next", "(", "record", "=", "inRecord", ",", "curInputBookmark", "=", "None", ")", "print", "'Record out: '", ",", "outRecord", "if", "(", "outRecord", "is", "not", "None", ")", ":", "output", ".", "appendRecord", "(", "outRecord", ",", "None", ")", "if", "(", "(", "inRecord", "is", "None", ")", "and", "(", "outRecord", "is", "None", ")", ")", ":", "break" ]
aggregate the input stream and write aggregated records to the output stream .
train
false
48,598
@frappe.whitelist() def apply_price_list(args, as_doc=False): args = process_args(args) parent = get_price_list_currency_and_exchange_rate(args) children = [] if (u'items' in args): item_list = args.get(u'items') args.update(parent) for item in item_list: args_copy = frappe._dict(args.copy()) args_copy.update(item) item_details = apply_price_list_on_item(args_copy) children.append(item_details) if as_doc: args.price_list_currency = parent.price_list_currency args.plc_conversion_rate = parent.plc_conversion_rate if args.get(u'items'): for (i, item) in enumerate(args.get(u'items')): for fieldname in children[i]: if ((fieldname in item) and (fieldname not in (u'name', u'doctype'))): item[fieldname] = children[i][fieldname] return args else: return {u'parent': parent, u'children': children}
[ "@", "frappe", ".", "whitelist", "(", ")", "def", "apply_price_list", "(", "args", ",", "as_doc", "=", "False", ")", ":", "args", "=", "process_args", "(", "args", ")", "parent", "=", "get_price_list_currency_and_exchange_rate", "(", "args", ")", "children", "=", "[", "]", "if", "(", "u'items'", "in", "args", ")", ":", "item_list", "=", "args", ".", "get", "(", "u'items'", ")", "args", ".", "update", "(", "parent", ")", "for", "item", "in", "item_list", ":", "args_copy", "=", "frappe", ".", "_dict", "(", "args", ".", "copy", "(", ")", ")", "args_copy", ".", "update", "(", "item", ")", "item_details", "=", "apply_price_list_on_item", "(", "args_copy", ")", "children", ".", "append", "(", "item_details", ")", "if", "as_doc", ":", "args", ".", "price_list_currency", "=", "parent", ".", "price_list_currency", "args", ".", "plc_conversion_rate", "=", "parent", ".", "plc_conversion_rate", "if", "args", ".", "get", "(", "u'items'", ")", ":", "for", "(", "i", ",", "item", ")", "in", "enumerate", "(", "args", ".", "get", "(", "u'items'", ")", ")", ":", "for", "fieldname", "in", "children", "[", "i", "]", ":", "if", "(", "(", "fieldname", "in", "item", ")", "and", "(", "fieldname", "not", "in", "(", "u'name'", ",", "u'doctype'", ")", ")", ")", ":", "item", "[", "fieldname", "]", "=", "children", "[", "i", "]", "[", "fieldname", "]", "return", "args", "else", ":", "return", "{", "u'parent'", ":", "parent", ",", "u'children'", ":", "children", "}" ]
apply pricelist on a document-like dict object and return as {parent: dict .
train
false
48,599
def _fake_is_smpl(*args): return True
[ "def", "_fake_is_smpl", "(", "*", "args", ")", ":", "return", "True" ]
assume the shadowimage pair status is smpl .
train
false
48,600
def split_low_tag(tag): (state, id_, name, fun) = tag.split('_|-') return {'state': state, '__id__': id_, 'name': name, 'fun': fun}
[ "def", "split_low_tag", "(", "tag", ")", ":", "(", "state", ",", "id_", ",", "name", ",", "fun", ")", "=", "tag", ".", "split", "(", "'_|-'", ")", "return", "{", "'state'", ":", "state", ",", "'__id__'", ":", "id_", ",", "'name'", ":", "name", ",", "'fun'", ":", "fun", "}" ]
take a low tag and split it back into the low dict that it came from .
train
true
48,601
def list_bans(banlist): if (not banlist): return 'No active bans were found.' table = prettytable.PrettyTable(['{wid', '{wname/ip', '{wdate', '{wreason']) for (inum, ban) in enumerate(banlist): table.add_row([str((inum + 1)), ((ban[0] and ban[0]) or ban[1]), ban[3], ban[4]]) string = ('{wActive bans:{n\n%s' % table) return string
[ "def", "list_bans", "(", "banlist", ")", ":", "if", "(", "not", "banlist", ")", ":", "return", "'No active bans were found.'", "table", "=", "prettytable", ".", "PrettyTable", "(", "[", "'{wid'", ",", "'{wname/ip'", ",", "'{wdate'", ",", "'{wreason'", "]", ")", "for", "(", "inum", ",", "ban", ")", "in", "enumerate", "(", "banlist", ")", ":", "table", ".", "add_row", "(", "[", "str", "(", "(", "inum", "+", "1", ")", ")", ",", "(", "(", "ban", "[", "0", "]", "and", "ban", "[", "0", "]", ")", "or", "ban", "[", "1", "]", ")", ",", "ban", "[", "3", "]", ",", "ban", "[", "4", "]", "]", ")", "string", "=", "(", "'{wActive bans:{n\\n%s'", "%", "table", ")", "return", "string" ]
helper function to display a list of active bans .
train
false
48,602
@pytest.mark.django_db def test_service_methods_with_long_name(rf): MAX_LENGTH = 100 long_name = ('X' * MAX_LENGTH) assert (len(long_name) == MAX_LENGTH) sm = ShippingMethod.objects.language('en').create(shop=get_default_shop(), name=long_name, enabled=True, tax_class=get_default_tax_class()) pm = PaymentMethod.objects.language('en').create(shop=get_default_shop(), name=long_name, enabled=True, tax_class=get_default_tax_class()) order = create_empty_order() order.shipping_method = sm order.payment_method = pm order.full_clean() order.save()
[ "@", "pytest", ".", "mark", ".", "django_db", "def", "test_service_methods_with_long_name", "(", "rf", ")", ":", "MAX_LENGTH", "=", "100", "long_name", "=", "(", "'X'", "*", "MAX_LENGTH", ")", "assert", "(", "len", "(", "long_name", ")", "==", "MAX_LENGTH", ")", "sm", "=", "ShippingMethod", ".", "objects", ".", "language", "(", "'en'", ")", ".", "create", "(", "shop", "=", "get_default_shop", "(", ")", ",", "name", "=", "long_name", ",", "enabled", "=", "True", ",", "tax_class", "=", "get_default_tax_class", "(", ")", ")", "pm", "=", "PaymentMethod", ".", "objects", ".", "language", "(", "'en'", ")", ".", "create", "(", "shop", "=", "get_default_shop", "(", ")", ",", "name", "=", "long_name", ",", "enabled", "=", "True", ",", "tax_class", "=", "get_default_tax_class", "(", ")", ")", "order", "=", "create_empty_order", "(", ")", "order", ".", "shipping_method", "=", "sm", "order", ".", "payment_method", "=", "pm", "order", ".", "full_clean", "(", ")", "order", ".", "save", "(", ")" ]
make sure that service methods with long names dont cause exceptions when creating an order .
train
false
48,604
def salt_refs(data): proto = 'salt://' ret = [] if isinstance(data, str): if data.startswith(proto): return [data] if isinstance(data, list): for comp in data: if isinstance(comp, str): if comp.startswith(proto): ret.append(comp) return ret
[ "def", "salt_refs", "(", "data", ")", ":", "proto", "=", "'salt://'", "ret", "=", "[", "]", "if", "isinstance", "(", "data", ",", "str", ")", ":", "if", "data", ".", "startswith", "(", "proto", ")", ":", "return", "[", "data", "]", "if", "isinstance", "(", "data", ",", "list", ")", ":", "for", "comp", "in", "data", ":", "if", "isinstance", "(", "comp", ",", "str", ")", ":", "if", "comp", ".", "startswith", "(", "proto", ")", ":", "ret", ".", "append", "(", "comp", ")", "return", "ret" ]
pull salt file references out of the states .
train
false
48,608
def test_statsd_fail(): logger = Statsd(Config()) logger.sock = MockSocket(True) logger.info('No impact on logging') logger.debug('No impact on logging') logger.critical('No impact on logging') logger.error('No impact on logging') logger.warning('No impact on logging') logger.exception('No impact on logging')
[ "def", "test_statsd_fail", "(", ")", ":", "logger", "=", "Statsd", "(", "Config", "(", ")", ")", "logger", ".", "sock", "=", "MockSocket", "(", "True", ")", "logger", ".", "info", "(", "'No impact on logging'", ")", "logger", ".", "debug", "(", "'No impact on logging'", ")", "logger", ".", "critical", "(", "'No impact on logging'", ")", "logger", ".", "error", "(", "'No impact on logging'", ")", "logger", ".", "warning", "(", "'No impact on logging'", ")", "logger", ".", "exception", "(", "'No impact on logging'", ")" ]
udp socket fails .
train
false
48,609
def quote_slashes(text): return re.sub(u'[;/]', _quote_slashes, text)
[ "def", "quote_slashes", "(", "text", ")", ":", "return", "re", ".", "sub", "(", "u'[;/]'", ",", "_quote_slashes", ",", "text", ")" ]
quote / characters so that they arent visible to djangos url quoting .
train
false
48,610
def migrate_uuid(node, wname): db = share_db() old_sharejs_uuid = get_sharejs_uuid(node, wname) broadcast_to_sharejs('lock', old_sharejs_uuid) generate_private_uuid(node, wname) new_sharejs_uuid = get_sharejs_uuid(node, wname) doc_item = db['docs'].find_one({'_id': old_sharejs_uuid}) if doc_item: doc_item['_id'] = new_sharejs_uuid db['docs'].insert(doc_item) db['docs'].remove({'_id': old_sharejs_uuid}) ops_items = [item for item in db['docs_ops'].find({'name': old_sharejs_uuid})] if ops_items: for item in ops_items: item['_id'] = item['_id'].replace(old_sharejs_uuid, new_sharejs_uuid) item['name'] = new_sharejs_uuid db['docs_ops'].insert(ops_items) db['docs_ops'].remove({'name': old_sharejs_uuid}) write_contributors = [user._id for user in node.contributors if node.has_permission(user, 'write')] broadcast_to_sharejs('unlock', old_sharejs_uuid, data=write_contributors)
[ "def", "migrate_uuid", "(", "node", ",", "wname", ")", ":", "db", "=", "share_db", "(", ")", "old_sharejs_uuid", "=", "get_sharejs_uuid", "(", "node", ",", "wname", ")", "broadcast_to_sharejs", "(", "'lock'", ",", "old_sharejs_uuid", ")", "generate_private_uuid", "(", "node", ",", "wname", ")", "new_sharejs_uuid", "=", "get_sharejs_uuid", "(", "node", ",", "wname", ")", "doc_item", "=", "db", "[", "'docs'", "]", ".", "find_one", "(", "{", "'_id'", ":", "old_sharejs_uuid", "}", ")", "if", "doc_item", ":", "doc_item", "[", "'_id'", "]", "=", "new_sharejs_uuid", "db", "[", "'docs'", "]", ".", "insert", "(", "doc_item", ")", "db", "[", "'docs'", "]", ".", "remove", "(", "{", "'_id'", ":", "old_sharejs_uuid", "}", ")", "ops_items", "=", "[", "item", "for", "item", "in", "db", "[", "'docs_ops'", "]", ".", "find", "(", "{", "'name'", ":", "old_sharejs_uuid", "}", ")", "]", "if", "ops_items", ":", "for", "item", "in", "ops_items", ":", "item", "[", "'_id'", "]", "=", "item", "[", "'_id'", "]", ".", "replace", "(", "old_sharejs_uuid", ",", "new_sharejs_uuid", ")", "item", "[", "'name'", "]", "=", "new_sharejs_uuid", "db", "[", "'docs_ops'", "]", ".", "insert", "(", "ops_items", ")", "db", "[", "'docs_ops'", "]", ".", "remove", "(", "{", "'name'", ":", "old_sharejs_uuid", "}", ")", "write_contributors", "=", "[", "user", ".", "_id", "for", "user", "in", "node", ".", "contributors", "if", "node", ".", "has_permission", "(", "user", ",", "'write'", ")", "]", "broadcast_to_sharejs", "(", "'unlock'", ",", "old_sharejs_uuid", ",", "data", "=", "write_contributors", ")" ]
migrates uuid to new namespace .
train
false
48,612
def _series_inversion1(p, x, prec): if rs_is_puiseux(p, x): return rs_puiseux(_series_inversion1, p, x, prec) R = p.ring zm = R.zero_monom c = p[zm] if (prec == int(prec)): prec = int(prec) if (zm not in p): raise ValueError('No constant term in series') if _has_constant_term((p - c), x): raise ValueError('p cannot contain a constant term depending on parameters') one = R(1) if (R.domain is EX): one = 1 if (c != one): p1 = (R(1) / c) else: p1 = R(1) for precx in _giant_steps(prec): t = (1 - rs_mul(p1, p, x, precx)) p1 = (p1 + rs_mul(p1, t, x, precx)) return p1
[ "def", "_series_inversion1", "(", "p", ",", "x", ",", "prec", ")", ":", "if", "rs_is_puiseux", "(", "p", ",", "x", ")", ":", "return", "rs_puiseux", "(", "_series_inversion1", ",", "p", ",", "x", ",", "prec", ")", "R", "=", "p", ".", "ring", "zm", "=", "R", ".", "zero_monom", "c", "=", "p", "[", "zm", "]", "if", "(", "prec", "==", "int", "(", "prec", ")", ")", ":", "prec", "=", "int", "(", "prec", ")", "if", "(", "zm", "not", "in", "p", ")", ":", "raise", "ValueError", "(", "'No constant term in series'", ")", "if", "_has_constant_term", "(", "(", "p", "-", "c", ")", ",", "x", ")", ":", "raise", "ValueError", "(", "'p cannot contain a constant term depending on parameters'", ")", "one", "=", "R", "(", "1", ")", "if", "(", "R", ".", "domain", "is", "EX", ")", ":", "one", "=", "1", "if", "(", "c", "!=", "one", ")", ":", "p1", "=", "(", "R", "(", "1", ")", "/", "c", ")", "else", ":", "p1", "=", "R", "(", "1", ")", "for", "precx", "in", "_giant_steps", "(", "prec", ")", ":", "t", "=", "(", "1", "-", "rs_mul", "(", "p1", ",", "p", ",", "x", ",", "precx", ")", ")", "p1", "=", "(", "p1", "+", "rs_mul", "(", "p1", ",", "t", ",", "x", ",", "precx", ")", ")", "return", "p1" ]
univariate series inversion 1/p modulo o .
train
false
48,613
def _InternalDeclareKeyFlags(flag_names, flag_values=FLAGS, key_flag_values=None): key_flag_values = (key_flag_values or flag_values) module = _GetCallingModule() for flag_name in flag_names: if (flag_name not in flag_values): raise UnrecognizedFlagError(flag_name) flag = flag_values.FlagDict()[flag_name] key_flag_values._RegisterKeyFlagForModule(module, flag)
[ "def", "_InternalDeclareKeyFlags", "(", "flag_names", ",", "flag_values", "=", "FLAGS", ",", "key_flag_values", "=", "None", ")", ":", "key_flag_values", "=", "(", "key_flag_values", "or", "flag_values", ")", "module", "=", "_GetCallingModule", "(", ")", "for", "flag_name", "in", "flag_names", ":", "if", "(", "flag_name", "not", "in", "flag_values", ")", ":", "raise", "UnrecognizedFlagError", "(", "flag_name", ")", "flag", "=", "flag_values", ".", "FlagDict", "(", ")", "[", "flag_name", "]", "key_flag_values", ".", "_RegisterKeyFlagForModule", "(", "module", ",", "flag", ")" ]
declares a flag as key for the calling module .
train
false
48,614
def default_from_email(): global _default_from_email if (_default_from_email is None): try: fqdn = socket.getfqdn() except IOError: fqdn = 'localhost' _default_from_email = ('hue@' + fqdn) return _default_from_email
[ "def", "default_from_email", "(", ")", ":", "global", "_default_from_email", "if", "(", "_default_from_email", "is", "None", ")", ":", "try", ":", "fqdn", "=", "socket", ".", "getfqdn", "(", ")", "except", "IOError", ":", "fqdn", "=", "'localhost'", "_default_from_email", "=", "(", "'hue@'", "+", "fqdn", ")", "return", "_default_from_email" ]
email for hue@<host-fqdn> .
train
false
48,616
def r_squared(alpha, beta, x, y): return (1.0 - (sum_of_squared_errors(alpha, beta, x, y) / total_sum_of_squares(y)))
[ "def", "r_squared", "(", "alpha", ",", "beta", ",", "x", ",", "y", ")", ":", "return", "(", "1.0", "-", "(", "sum_of_squared_errors", "(", "alpha", ",", "beta", ",", "x", ",", "y", ")", "/", "total_sum_of_squares", "(", "y", ")", ")", ")" ]
the fraction of variation in y captured by the model .
train
false
48,617
def get_pixels(): return _sensehat.get_pixels()
[ "def", "get_pixels", "(", ")", ":", "return", "_sensehat", ".", "get_pixels", "(", ")" ]
returns a list of 64 smaller lists of [r .
train
false
48,618
def get_yaml_loader(argline): def yaml_loader(*args): return SaltYamlSafeLoader(dictclass=OrderedDict, *args) return yaml_loader
[ "def", "get_yaml_loader", "(", "argline", ")", ":", "def", "yaml_loader", "(", "*", "args", ")", ":", "return", "SaltYamlSafeLoader", "(", "dictclass", "=", "OrderedDict", ",", "*", "args", ")", "return", "yaml_loader" ]
return the ordered dict yaml loader .
train
false
48,619
def remove_image(image): client = _get_client() status = base_status.copy() try: infos = _get_image_infos(image) if infos: status['id'] = infos['Id'] try: client.remove_image(infos['Id']) except Exception: _invalid(status, id_=image, out=traceback.format_exc(), comment='Image could not be deleted') try: infos = _get_image_infos(image) _invalid(status, comment='Image marked to be deleted but not deleted yet') except Exception: _valid(status, id_=image, comment='Image deleted') else: _invalid(status) except Exception: _invalid(status, out=traceback.format_exc(), comment='Image does not exist: {0}'.format(image)) return status
[ "def", "remove_image", "(", "image", ")", ":", "client", "=", "_get_client", "(", ")", "status", "=", "base_status", ".", "copy", "(", ")", "try", ":", "infos", "=", "_get_image_infos", "(", "image", ")", "if", "infos", ":", "status", "[", "'id'", "]", "=", "infos", "[", "'Id'", "]", "try", ":", "client", ".", "remove_image", "(", "infos", "[", "'Id'", "]", ")", "except", "Exception", ":", "_invalid", "(", "status", ",", "id_", "=", "image", ",", "out", "=", "traceback", ".", "format_exc", "(", ")", ",", "comment", "=", "'Image could not be deleted'", ")", "try", ":", "infos", "=", "_get_image_infos", "(", "image", ")", "_invalid", "(", "status", ",", "comment", "=", "'Image marked to be deleted but not deleted yet'", ")", "except", "Exception", ":", "_valid", "(", "status", ",", "id_", "=", "image", ",", "comment", "=", "'Image deleted'", ")", "else", ":", "_invalid", "(", "status", ")", "except", "Exception", ":", "_invalid", "(", "status", ",", "out", "=", "traceback", ".", "format_exc", "(", ")", ",", "comment", "=", "'Image does not exist: {0}'", ".", "format", "(", "image", ")", ")", "return", "status" ]
remove an image from a system .
train
false
48,621
def syncloop(agents, sleep=2): while True: for agent in agents: agent.sync() time.sleep(sleep)
[ "def", "syncloop", "(", "agents", ",", "sleep", "=", "2", ")", ":", "while", "True", ":", "for", "agent", "in", "agents", ":", "agent", ".", "sync", "(", ")", "time", ".", "sleep", "(", "sleep", ")" ]
synchronize periodically the agents .
train
false
48,622
def dup_prs_resultant(f, g, K): if ((not f) or (not g)): return (K.zero, []) (R, S) = dup_inner_subresultants(f, g, K) if (dup_degree(R[(-1)]) > 0): return (K.zero, R) return (S[(-1)], R)
[ "def", "dup_prs_resultant", "(", "f", ",", "g", ",", "K", ")", ":", "if", "(", "(", "not", "f", ")", "or", "(", "not", "g", ")", ")", ":", "return", "(", "K", ".", "zero", ",", "[", "]", ")", "(", "R", ",", "S", ")", "=", "dup_inner_subresultants", "(", "f", ",", "g", ",", "K", ")", "if", "(", "dup_degree", "(", "R", "[", "(", "-", "1", ")", "]", ")", ">", "0", ")", ":", "return", "(", "K", ".", "zero", ",", "R", ")", "return", "(", "S", "[", "(", "-", "1", ")", "]", ",", "R", ")" ]
resultant algorithm in k[x] using subresultant prs .
train
false
48,623
def _pattern_of(index): return np.array([[(index & (2 ** 0)), (index & (2 ** 1)), (index & (2 ** 2))], [(index & (2 ** 3)), (index & (2 ** 4)), (index & (2 ** 5))], [(index & (2 ** 6)), (index & (2 ** 7)), (index & (2 ** 8))]], bool)
[ "def", "_pattern_of", "(", "index", ")", ":", "return", "np", ".", "array", "(", "[", "[", "(", "index", "&", "(", "2", "**", "0", ")", ")", ",", "(", "index", "&", "(", "2", "**", "1", ")", ")", ",", "(", "index", "&", "(", "2", "**", "2", ")", ")", "]", ",", "[", "(", "index", "&", "(", "2", "**", "3", ")", ")", ",", "(", "index", "&", "(", "2", "**", "4", ")", ")", ",", "(", "index", "&", "(", "2", "**", "5", ")", ")", "]", ",", "[", "(", "index", "&", "(", "2", "**", "6", ")", ")", ",", "(", "index", "&", "(", "2", "**", "7", ")", ")", ",", "(", "index", "&", "(", "2", "**", "8", ")", ")", "]", "]", ",", "bool", ")" ]
return the pattern represented by an index value byte decomposition of index .
train
false
48,624
def mathjax_for_rst(pelicanobj, mathjax_script): docutils_settings = pelicanobj.settings.get('DOCUTILS_SETTINGS', {}) docutils_settings['math_output'] = 'MathJax' pelicanobj.settings['DOCUTILS_SETTINGS'] = docutils_settings rst_add_mathjax.mathjax_script = mathjax_script
[ "def", "mathjax_for_rst", "(", "pelicanobj", ",", "mathjax_script", ")", ":", "docutils_settings", "=", "pelicanobj", ".", "settings", ".", "get", "(", "'DOCUTILS_SETTINGS'", ",", "{", "}", ")", "docutils_settings", "[", "'math_output'", "]", "=", "'MathJax'", "pelicanobj", ".", "settings", "[", "'DOCUTILS_SETTINGS'", "]", "=", "docutils_settings", "rst_add_mathjax", ".", "mathjax_script", "=", "mathjax_script" ]
setup math for rst .
train
false
48,628
def angular_momentum(point, frame, *body): if (not isinstance(frame, ReferenceFrame)): raise TypeError('Please enter a valid ReferenceFrame') if (not isinstance(point, Point)): raise TypeError('Please specify a valid Point') else: angular_momentum_sys = Vector(0) for e in body: if isinstance(e, (RigidBody, Particle)): angular_momentum_sys += e.angular_momentum(point, frame) else: raise TypeError('*body must have only Particle or RigidBody') return angular_momentum_sys
[ "def", "angular_momentum", "(", "point", ",", "frame", ",", "*", "body", ")", ":", "if", "(", "not", "isinstance", "(", "frame", ",", "ReferenceFrame", ")", ")", ":", "raise", "TypeError", "(", "'Please enter a valid ReferenceFrame'", ")", "if", "(", "not", "isinstance", "(", "point", ",", "Point", ")", ")", ":", "raise", "TypeError", "(", "'Please specify a valid Point'", ")", "else", ":", "angular_momentum_sys", "=", "Vector", "(", "0", ")", "for", "e", "in", "body", ":", "if", "isinstance", "(", "e", ",", "(", "RigidBody", ",", "Particle", ")", ")", ":", "angular_momentum_sys", "+=", "e", ".", "angular_momentum", "(", "point", ",", "frame", ")", "else", ":", "raise", "TypeError", "(", "'*body must have only Particle or RigidBody'", ")", "return", "angular_momentum_sys" ]
angular momentum of a system this function returns the angular momentum of a system of particles and/or rigidbodys .
train
false
48,629
def hypersimp(f, k): f = sympify(f) g = (f.subs(k, (k + 1)) / f) g = g.rewrite(gamma) g = expand_func(g) g = powsimp(g, deep=True, combine='exp') if g.is_rational_function(k): return simplify(g, ratio=S.Infinity) else: return None
[ "def", "hypersimp", "(", "f", ",", "k", ")", ":", "f", "=", "sympify", "(", "f", ")", "g", "=", "(", "f", ".", "subs", "(", "k", ",", "(", "k", "+", "1", ")", ")", "/", "f", ")", "g", "=", "g", ".", "rewrite", "(", "gamma", ")", "g", "=", "expand_func", "(", "g", ")", "g", "=", "powsimp", "(", "g", ",", "deep", "=", "True", ",", "combine", "=", "'exp'", ")", "if", "g", ".", "is_rational_function", "(", "k", ")", ":", "return", "simplify", "(", "g", ",", "ratio", "=", "S", ".", "Infinity", ")", "else", ":", "return", "None" ]
given combinatorial term f(k) simplify its consecutive term ratio i .
train
false
48,630
@utils.arg('address', metavar='<address>', help=_('IP of Floating IP.')) @deprecated_network def do_floating_ip_delete(cs, args): floating_ips = cs.floating_ips.list() for floating_ip in floating_ips: if (floating_ip.ip == args.address): return cs.floating_ips.delete(floating_ip.id) raise exceptions.CommandError((_('Floating IP %s not found.') % args.address))
[ "@", "utils", ".", "arg", "(", "'address'", ",", "metavar", "=", "'<address>'", ",", "help", "=", "_", "(", "'IP of Floating IP.'", ")", ")", "@", "deprecated_network", "def", "do_floating_ip_delete", "(", "cs", ",", "args", ")", ":", "floating_ips", "=", "cs", ".", "floating_ips", ".", "list", "(", ")", "for", "floating_ip", "in", "floating_ips", ":", "if", "(", "floating_ip", ".", "ip", "==", "args", ".", "address", ")", ":", "return", "cs", ".", "floating_ips", ".", "delete", "(", "floating_ip", ".", "id", ")", "raise", "exceptions", ".", "CommandError", "(", "(", "_", "(", "'Floating IP %s not found.'", ")", "%", "args", ".", "address", ")", ")" ]
de-allocate a floating ip .
train
false
48,631
def nice_exit(ret=0): if (ret and is_bare_console()): print print 'Press [Enter] to close this window.' raw_input() sys.exit(ret)
[ "def", "nice_exit", "(", "ret", "=", "0", ")", ":", "if", "(", "ret", "and", "is_bare_console", "(", ")", ")", ":", "print", "print", "'Press [Enter] to close this window.'", "raw_input", "(", ")", "sys", ".", "exit", "(", "ret", ")" ]
drop-in replacement for sys .
train
false
48,632
def test_kde_integer_input(): x1 = np.arange(5) kde = stats.gaussian_kde(x1) y_expected = [0.13480721, 0.18222869, 0.19514935, 0.18222869, 0.13480721] assert_array_almost_equal(kde(x1), y_expected, decimal=6)
[ "def", "test_kde_integer_input", "(", ")", ":", "x1", "=", "np", ".", "arange", "(", "5", ")", "kde", "=", "stats", ".", "gaussian_kde", "(", "x1", ")", "y_expected", "=", "[", "0.13480721", ",", "0.18222869", ",", "0.19514935", ",", "0.18222869", ",", "0.13480721", "]", "assert_array_almost_equal", "(", "kde", "(", "x1", ")", ",", "y_expected", ",", "decimal", "=", "6", ")" ]
regression test for #1181 .
train
false
48,633
def resume(vm_): with _get_xapi_session() as xapi: vm_uuid = _get_label_uuid(xapi, 'VM', vm_) if (vm_uuid is False): return False try: xapi.VM.unpause(vm_uuid) return True except Exception: return False
[ "def", "resume", "(", "vm_", ")", ":", "with", "_get_xapi_session", "(", ")", "as", "xapi", ":", "vm_uuid", "=", "_get_label_uuid", "(", "xapi", ",", "'VM'", ",", "vm_", ")", "if", "(", "vm_uuid", "is", "False", ")", ":", "return", "False", "try", ":", "xapi", ".", "VM", ".", "unpause", "(", "vm_uuid", ")", "return", "True", "except", "Exception", ":", "return", "False" ]
resumes processing of queues .
train
true
48,635
def sh_margin_details(date='', symbol='', start='', end='', retry_count=3, pause=0.001): date = (date if (date == '') else date.replace('-', '')) start = (start if (start == '') else start.replace('-', '')) end = (end if (end == '') else end.replace('-', '')) if ((start != '') & (end != '')): date = '' data = pd.DataFrame() ct._write_head() df = _sh_mx(data, date=date, start=start, end=end, symbol=symbol, retry_count=retry_count, pause=pause) return df
[ "def", "sh_margin_details", "(", "date", "=", "''", ",", "symbol", "=", "''", ",", "start", "=", "''", ",", "end", "=", "''", ",", "retry_count", "=", "3", ",", "pause", "=", "0.001", ")", ":", "date", "=", "(", "date", "if", "(", "date", "==", "''", ")", "else", "date", ".", "replace", "(", "'-'", ",", "''", ")", ")", "start", "=", "(", "start", "if", "(", "start", "==", "''", ")", "else", "start", ".", "replace", "(", "'-'", ",", "''", ")", ")", "end", "=", "(", "end", "if", "(", "end", "==", "''", ")", "else", "end", ".", "replace", "(", "'-'", ",", "''", ")", ")", "if", "(", "(", "start", "!=", "''", ")", "&", "(", "end", "!=", "''", ")", ")", ":", "date", "=", "''", "data", "=", "pd", ".", "DataFrame", "(", ")", "ct", ".", "_write_head", "(", ")", "df", "=", "_sh_mx", "(", "data", ",", "date", "=", "date", ",", "start", "=", "start", ",", "end", "=", "end", ",", "symbol", "=", "symbol", ",", "retry_count", "=", "retry_count", ",", "pause", "=", "pause", ")", "return", "df" ]
parameters date:string 明细数据日期 format:yyyy-mm-dd 默认为空 symbol:string 标的代码,6位数字e .
train
false
48,636
def format_log_context(msg, connection=None, keyspace=None): connection_info = (connection or 'DEFAULT_CONNECTION') if keyspace: msg = '[Connection: {0}, Keyspace: {1}] {2}'.format(connection_info, keyspace, msg) else: msg = '[Connection: {0}] {1}'.format(connection_info, msg) return msg
[ "def", "format_log_context", "(", "msg", ",", "connection", "=", "None", ",", "keyspace", "=", "None", ")", ":", "connection_info", "=", "(", "connection", "or", "'DEFAULT_CONNECTION'", ")", "if", "keyspace", ":", "msg", "=", "'[Connection: {0}, Keyspace: {1}] {2}'", ".", "format", "(", "connection_info", ",", "keyspace", ",", "msg", ")", "else", ":", "msg", "=", "'[Connection: {0}] {1}'", ".", "format", "(", "connection_info", ",", "msg", ")", "return", "msg" ]
format log message to add keyspace and connection context .
train
true
48,637
def cms_to_token(cms_text): start_delim = '-----BEGIN CMS-----' end_delim = '-----END CMS-----' signed_text = cms_text signed_text = signed_text.replace('/', '-') signed_text = signed_text.replace(start_delim, '') signed_text = signed_text.replace(end_delim, '') signed_text = signed_text.replace('\n', '') return signed_text
[ "def", "cms_to_token", "(", "cms_text", ")", ":", "start_delim", "=", "'-----BEGIN CMS-----'", "end_delim", "=", "'-----END CMS-----'", "signed_text", "=", "cms_text", "signed_text", "=", "signed_text", ".", "replace", "(", "'/'", ",", "'-'", ")", "signed_text", "=", "signed_text", ".", "replace", "(", "start_delim", ",", "''", ")", "signed_text", "=", "signed_text", ".", "replace", "(", "end_delim", ",", "''", ")", "signed_text", "=", "signed_text", ".", "replace", "(", "'\\n'", ",", "''", ")", "return", "signed_text" ]
convert a cms-signed token in pem format to a custom url-safe format .
train
false
48,639
@then(u'the command output should contain') def step_command_output_should_contain(context): assert (context.text is not None), 'REQUIRE: multi-line text' step_command_output_should_contain_text(context, context.text)
[ "@", "then", "(", "u'the command output should contain'", ")", "def", "step_command_output_should_contain", "(", "context", ")", ":", "assert", "(", "context", ".", "text", "is", "not", "None", ")", ",", "'REQUIRE: multi-line text'", "step_command_output_should_contain_text", "(", "context", ",", "context", ".", "text", ")" ]
example: when i run "behave .
train
false
48,640
@simple_decorator def check_local_site_access(view_func): def _check(request, local_site_name=None, *args, **kwargs): if local_site_name: if (not request.local_site): raise Http404 local_site = request.local_site if (not local_site.is_accessible_by(request.user)): if (local_site.public or request.user.is_authenticated()): response = render_to_response('permission_denied.html', RequestContext(request)) response.status_code = 403 return response else: return HttpResponseRedirect(('%s?next_page=%s' % (reverse('login'), request.get_full_path()))) else: local_site = None return view_func(request, local_site=local_site, *args, **kwargs) return _check
[ "@", "simple_decorator", "def", "check_local_site_access", "(", "view_func", ")", ":", "def", "_check", "(", "request", ",", "local_site_name", "=", "None", ",", "*", "args", ",", "**", "kwargs", ")", ":", "if", "local_site_name", ":", "if", "(", "not", "request", ".", "local_site", ")", ":", "raise", "Http404", "local_site", "=", "request", ".", "local_site", "if", "(", "not", "local_site", ".", "is_accessible_by", "(", "request", ".", "user", ")", ")", ":", "if", "(", "local_site", ".", "public", "or", "request", ".", "user", ".", "is_authenticated", "(", ")", ")", ":", "response", "=", "render_to_response", "(", "'permission_denied.html'", ",", "RequestContext", "(", "request", ")", ")", "response", ".", "status_code", "=", "403", "return", "response", "else", ":", "return", "HttpResponseRedirect", "(", "(", "'%s?next_page=%s'", "%", "(", "reverse", "(", "'login'", ")", ",", "request", ".", "get_full_path", "(", ")", ")", ")", ")", "else", ":", "local_site", "=", "None", "return", "view_func", "(", "request", ",", "local_site", "=", "local_site", ",", "*", "args", ",", "**", "kwargs", ")", "return", "_check" ]
checks if a user has access to a local site .
train
false
48,641
def digestAuth(realm, algorithm=MD5, nonce=None, qop=AUTH): global SUPPORTED_ALGORITHM, DIGEST_AUTH_ENCODERS, SUPPORTED_QOP assert (algorithm in SUPPORTED_ALGORITHM) assert (qop in SUPPORTED_QOP) if (nonce is None): nonce = calculateNonce(realm, algorithm) return ('Digest realm="%s", nonce="%s", algorithm="%s", qop="%s"' % (realm, nonce, algorithm, qop))
[ "def", "digestAuth", "(", "realm", ",", "algorithm", "=", "MD5", ",", "nonce", "=", "None", ",", "qop", "=", "AUTH", ")", ":", "global", "SUPPORTED_ALGORITHM", ",", "DIGEST_AUTH_ENCODERS", ",", "SUPPORTED_QOP", "assert", "(", "algorithm", "in", "SUPPORTED_ALGORITHM", ")", "assert", "(", "qop", "in", "SUPPORTED_QOP", ")", "if", "(", "nonce", "is", "None", ")", ":", "nonce", "=", "calculateNonce", "(", "realm", ",", "algorithm", ")", "return", "(", "'Digest realm=\"%s\", nonce=\"%s\", algorithm=\"%s\", qop=\"%s\"'", "%", "(", "realm", ",", "nonce", ",", "algorithm", ",", "qop", ")", ")" ]
challenges the client for a digest authentication .
train
false
48,642
def assert_sp_series_equal(left, right, check_dtype=True, exact_indices=True, check_series_type=True, check_names=True, obj='SparseSeries'): assertIsInstance(left, pd.SparseSeries, '[SparseSeries]') assertIsInstance(right, pd.SparseSeries, '[SparseSeries]') if check_series_type: assert_class_equal(left, right, obj=obj) assert_index_equal(left.index, right.index, obj='{0}.index'.format(obj)) assert_sp_array_equal(left.block.values, right.block.values) if check_names: assert_attr_equal('name', left, right) if check_dtype: assert_attr_equal('dtype', left, right) assert_numpy_array_equal(left.values, right.values)
[ "def", "assert_sp_series_equal", "(", "left", ",", "right", ",", "check_dtype", "=", "True", ",", "exact_indices", "=", "True", ",", "check_series_type", "=", "True", ",", "check_names", "=", "True", ",", "obj", "=", "'SparseSeries'", ")", ":", "assertIsInstance", "(", "left", ",", "pd", ".", "SparseSeries", ",", "'[SparseSeries]'", ")", "assertIsInstance", "(", "right", ",", "pd", ".", "SparseSeries", ",", "'[SparseSeries]'", ")", "if", "check_series_type", ":", "assert_class_equal", "(", "left", ",", "right", ",", "obj", "=", "obj", ")", "assert_index_equal", "(", "left", ".", "index", ",", "right", ".", "index", ",", "obj", "=", "'{0}.index'", ".", "format", "(", "obj", ")", ")", "assert_sp_array_equal", "(", "left", ".", "block", ".", "values", ",", "right", ".", "block", ".", "values", ")", "if", "check_names", ":", "assert_attr_equal", "(", "'name'", ",", "left", ",", "right", ")", "if", "check_dtype", ":", "assert_attr_equal", "(", "'dtype'", ",", "left", ",", "right", ")", "assert_numpy_array_equal", "(", "left", ".", "values", ",", "right", ".", "values", ")" ]
check that the left and right sparseseries are equal .
train
false
48,644
def respond_as_web_page(title, html, success=None, http_status_code=None, context=None, indicator_color=None, primary_action=u'/', primary_label=None, fullpage=False): local.message_title = title local.message = html local.response[u'type'] = u'page' local.response[u'route'] = u'message' if http_status_code: local.response[u'http_status_code'] = http_status_code if (not context): context = {} if (not indicator_color): if success: indicator_color = u'green' elif (http_status_code and (http_status_code > 300)): indicator_color = u'red' else: indicator_color = u'blue' context[u'indicator_color'] = indicator_color context[u'primary_label'] = primary_label context[u'primary_action'] = primary_action context[u'error_code'] = http_status_code context[u'fullpage'] = fullpage local.response[u'context'] = context
[ "def", "respond_as_web_page", "(", "title", ",", "html", ",", "success", "=", "None", ",", "http_status_code", "=", "None", ",", "context", "=", "None", ",", "indicator_color", "=", "None", ",", "primary_action", "=", "u'/'", ",", "primary_label", "=", "None", ",", "fullpage", "=", "False", ")", ":", "local", ".", "message_title", "=", "title", "local", ".", "message", "=", "html", "local", ".", "response", "[", "u'type'", "]", "=", "u'page'", "local", ".", "response", "[", "u'route'", "]", "=", "u'message'", "if", "http_status_code", ":", "local", ".", "response", "[", "u'http_status_code'", "]", "=", "http_status_code", "if", "(", "not", "context", ")", ":", "context", "=", "{", "}", "if", "(", "not", "indicator_color", ")", ":", "if", "success", ":", "indicator_color", "=", "u'green'", "elif", "(", "http_status_code", "and", "(", "http_status_code", ">", "300", ")", ")", ":", "indicator_color", "=", "u'red'", "else", ":", "indicator_color", "=", "u'blue'", "context", "[", "u'indicator_color'", "]", "=", "indicator_color", "context", "[", "u'primary_label'", "]", "=", "primary_label", "context", "[", "u'primary_action'", "]", "=", "primary_action", "context", "[", "u'error_code'", "]", "=", "http_status_code", "context", "[", "u'fullpage'", "]", "=", "fullpage", "local", ".", "response", "[", "u'context'", "]", "=", "context" ]
send response as a web page with a message rather than json .
train
false
48,645
def apphook_pre_title_checker(instance, **kwargs): if instance.publisher_is_draft: return try: instance._old_data = Title.objects.filter(pk=instance.pk).select_related('page')[0] except IndexError: instance._old_data = None
[ "def", "apphook_pre_title_checker", "(", "instance", ",", "**", "kwargs", ")", ":", "if", "instance", ".", "publisher_is_draft", ":", "return", "try", ":", "instance", ".", "_old_data", "=", "Title", ".", "objects", ".", "filter", "(", "pk", "=", "instance", ".", "pk", ")", ".", "select_related", "(", "'page'", ")", "[", "0", "]", "except", "IndexError", ":", "instance", ".", "_old_data", "=", "None" ]
store the old application_urls and path on the instance .
train
false
48,646
def gf_multi_eval(f, A, p, K): return [gf_eval(f, a, p, K) for a in A]
[ "def", "gf_multi_eval", "(", "f", ",", "A", ",", "p", ",", "K", ")", ":", "return", "[", "gf_eval", "(", "f", ",", "a", ",", "p", ",", "K", ")", "for", "a", "in", "A", "]" ]
evaluate f(a) for a in [a_1 .
train
false
48,647
@contextmanager def dummy(): (yield)
[ "@", "contextmanager", "def", "dummy", "(", ")", ":", "(", "yield", ")" ]
a context manager that does nothing special .
train
false
48,648
def set_edge_attributes(G, name, values): if (not isinstance(values, dict)): if G.is_multigraph(): edges = G.edges(keys=True) else: edges = G.edges() values = dict(zip_longest(edges, [], fillvalue=values)) if G.is_multigraph(): for ((u, v, key), value) in values.items(): G[u][v][key][name] = value else: for ((u, v), value) in values.items(): G[u][v][name] = value
[ "def", "set_edge_attributes", "(", "G", ",", "name", ",", "values", ")", ":", "if", "(", "not", "isinstance", "(", "values", ",", "dict", ")", ")", ":", "if", "G", ".", "is_multigraph", "(", ")", ":", "edges", "=", "G", ".", "edges", "(", "keys", "=", "True", ")", "else", ":", "edges", "=", "G", ".", "edges", "(", ")", "values", "=", "dict", "(", "zip_longest", "(", "edges", ",", "[", "]", ",", "fillvalue", "=", "values", ")", ")", "if", "G", ".", "is_multigraph", "(", ")", ":", "for", "(", "(", "u", ",", "v", ",", "key", ")", ",", "value", ")", "in", "values", ".", "items", "(", ")", ":", "G", "[", "u", "]", "[", "v", "]", "[", "key", "]", "[", "name", "]", "=", "value", "else", ":", "for", "(", "(", "u", ",", "v", ")", ",", "value", ")", "in", "values", ".", "items", "(", ")", ":", "G", "[", "u", "]", "[", "v", "]", "[", "name", "]", "=", "value" ]
sets edge attributes from a given value or dictionary of values .
train
false
48,649
def orderedSet(iterable): res = [] for el in iterable: if (el not in res): res.append(el) return res
[ "def", "orderedSet", "(", "iterable", ")", ":", "res", "=", "[", "]", "for", "el", "in", "iterable", ":", "if", "(", "el", "not", "in", "res", ")", ":", "res", ".", "append", "(", "el", ")", "return", "res" ]
remove all duplicates from the input iterable .
train
false
48,652
def equals_lf(line): return (line == ('\n' if isinstance(line, str) else '\n'))
[ "def", "equals_lf", "(", "line", ")", ":", "return", "(", "line", "==", "(", "'\\n'", "if", "isinstance", "(", "line", ",", "str", ")", "else", "'\\n'", ")", ")" ]
return true if line equals .
train
false
48,653
def meanexcess(thresh, shape, scale): warnif((shape > (-1)), 'shape > -1') warnif((thresh >= 0), 'thresh >= 0') warnif(((scale - (shape * thresh)) > 0), '(scale - shape*thresh) > 0') return ((scale - (shape * thresh)) / (1 + shape))
[ "def", "meanexcess", "(", "thresh", ",", "shape", ",", "scale", ")", ":", "warnif", "(", "(", "shape", ">", "(", "-", "1", ")", ")", ",", "'shape > -1'", ")", "warnif", "(", "(", "thresh", ">=", "0", ")", ",", "'thresh >= 0'", ")", "warnif", "(", "(", "(", "scale", "-", "(", "shape", "*", "thresh", ")", ")", ">", "0", ")", ",", "'(scale - shape*thresh) > 0'", ")", "return", "(", "(", "scale", "-", "(", "shape", "*", "thresh", ")", ")", "/", "(", "1", "+", "shape", ")", ")" ]
mean excess function of genpareto assert are inequality conditions in de zea bermudez .
train
false
48,655
def allow_access(course, user, level, send_email=True): _change_access(course, user, level, 'allow', send_email)
[ "def", "allow_access", "(", "course", ",", "user", ",", "level", ",", "send_email", "=", "True", ")", ":", "_change_access", "(", "course", ",", "user", ",", "level", ",", "'allow'", ",", "send_email", ")" ]
allow user access to course modification .
train
false
48,656
@login_required @ensure_csrf_cookie @require_http_methods(['GET']) def course_rerun_handler(request, course_key_string): if (not GlobalStaff().has_user(request.user)): raise PermissionDenied() course_key = CourseKey.from_string(course_key_string) with modulestore().bulk_operations(course_key): course_module = get_course_and_check_access(course_key, request.user, depth=3) if (request.method == 'GET'): return render_to_response('course-create-rerun.html', {'source_course_key': course_key, 'display_name': course_module.display_name, 'user': request.user, 'course_creator_status': _get_course_creator_status(request.user), 'allow_unicode_course_id': settings.FEATURES.get('ALLOW_UNICODE_COURSE_ID', False)})
[ "@", "login_required", "@", "ensure_csrf_cookie", "@", "require_http_methods", "(", "[", "'GET'", "]", ")", "def", "course_rerun_handler", "(", "request", ",", "course_key_string", ")", ":", "if", "(", "not", "GlobalStaff", "(", ")", ".", "has_user", "(", "request", ".", "user", ")", ")", ":", "raise", "PermissionDenied", "(", ")", "course_key", "=", "CourseKey", ".", "from_string", "(", "course_key_string", ")", "with", "modulestore", "(", ")", ".", "bulk_operations", "(", "course_key", ")", ":", "course_module", "=", "get_course_and_check_access", "(", "course_key", ",", "request", ".", "user", ",", "depth", "=", "3", ")", "if", "(", "request", ".", "method", "==", "'GET'", ")", ":", "return", "render_to_response", "(", "'course-create-rerun.html'", ",", "{", "'source_course_key'", ":", "course_key", ",", "'display_name'", ":", "course_module", ".", "display_name", ",", "'user'", ":", "request", ".", "user", ",", "'course_creator_status'", ":", "_get_course_creator_status", "(", "request", ".", "user", ")", ",", "'allow_unicode_course_id'", ":", "settings", ".", "FEATURES", ".", "get", "(", "'ALLOW_UNICODE_COURSE_ID'", ",", "False", ")", "}", ")" ]
the restful handler for course reruns .
train
false
48,657
def my_update_subtask_status(entry_id, current_task_id, new_subtask_status): entry = InstructorTask.objects.get(pk=entry_id) subtask_dict = json.loads(entry.subtasks) subtask_status_info = subtask_dict['status'] current_subtask_status = SubtaskStatus.from_dict(subtask_status_info[current_task_id]) current_retry_count = current_subtask_status.get_retry_count() new_retry_count = new_subtask_status.get_retry_count() if (current_retry_count <= new_retry_count): update_subtask_status(entry_id, current_task_id, new_subtask_status)
[ "def", "my_update_subtask_status", "(", "entry_id", ",", "current_task_id", ",", "new_subtask_status", ")", ":", "entry", "=", "InstructorTask", ".", "objects", ".", "get", "(", "pk", "=", "entry_id", ")", "subtask_dict", "=", "json", ".", "loads", "(", "entry", ".", "subtasks", ")", "subtask_status_info", "=", "subtask_dict", "[", "'status'", "]", "current_subtask_status", "=", "SubtaskStatus", ".", "from_dict", "(", "subtask_status_info", "[", "current_task_id", "]", ")", "current_retry_count", "=", "current_subtask_status", ".", "get_retry_count", "(", ")", "new_retry_count", "=", "new_subtask_status", ".", "get_retry_count", "(", ")", "if", "(", "current_retry_count", "<=", "new_retry_count", ")", ":", "update_subtask_status", "(", "entry_id", ",", "current_task_id", ",", "new_subtask_status", ")" ]
check whether a subtask has been updated before really updating .
train
false
48,658
def logmps(params, xsorted, dist): xcdf = np.r_[(0.0, dist.cdf(xsorted, *params), 1.0)] D = np.diff(xcdf) return (- np.log(D).mean())
[ "def", "logmps", "(", "params", ",", "xsorted", ",", "dist", ")", ":", "xcdf", "=", "np", ".", "r_", "[", "(", "0.0", ",", "dist", ".", "cdf", "(", "xsorted", ",", "*", "params", ")", ",", "1.0", ")", "]", "D", "=", "np", ".", "diff", "(", "xcdf", ")", "return", "(", "-", "np", ".", "log", "(", "D", ")", ".", "mean", "(", ")", ")" ]
calculate negative log of product-of-spacings parameters params : array_like .
train
false
48,659
def test_ipv6(): if (not cfg.selftest_host()): return True try: info = getipaddress.addresslookup6(cfg.selftest_host()) except: logging.debug("Test IPv6: Disabling IPv6, because it looks like it's not available. Reason: %s", sys.exc_info()[0]) return False try: (af, socktype, proto, canonname, sa) = info[0] sock = socket.socket(af, socktype, proto) sock.settimeout(2) sock.connect(sa[0:2]) sock.close() logging.debug('Test IPv6: IPv6 test successful. Enabling IPv6') return True except socket.error: logging.debug('Test IPv6: Cannot reach IPv6 test host. Disabling IPv6') return False except: logging.debug('Test IPv6: Problem during IPv6 connect. Disabling IPv6. Reason: %s', sys.exc_info()[0]) return False
[ "def", "test_ipv6", "(", ")", ":", "if", "(", "not", "cfg", ".", "selftest_host", "(", ")", ")", ":", "return", "True", "try", ":", "info", "=", "getipaddress", ".", "addresslookup6", "(", "cfg", ".", "selftest_host", "(", ")", ")", "except", ":", "logging", ".", "debug", "(", "\"Test IPv6: Disabling IPv6, because it looks like it's not available. Reason: %s\"", ",", "sys", ".", "exc_info", "(", ")", "[", "0", "]", ")", "return", "False", "try", ":", "(", "af", ",", "socktype", ",", "proto", ",", "canonname", ",", "sa", ")", "=", "info", "[", "0", "]", "sock", "=", "socket", ".", "socket", "(", "af", ",", "socktype", ",", "proto", ")", "sock", ".", "settimeout", "(", "2", ")", "sock", ".", "connect", "(", "sa", "[", "0", ":", "2", "]", ")", "sock", ".", "close", "(", ")", "logging", ".", "debug", "(", "'Test IPv6: IPv6 test successful. Enabling IPv6'", ")", "return", "True", "except", "socket", ".", "error", ":", "logging", ".", "debug", "(", "'Test IPv6: Cannot reach IPv6 test host. Disabling IPv6'", ")", "return", "False", "except", ":", "logging", ".", "debug", "(", "'Test IPv6: Problem during IPv6 connect. Disabling IPv6. Reason: %s'", ",", "sys", ".", "exc_info", "(", ")", "[", "0", "]", ")", "return", "False" ]
check if external ipv6 addresses are reachable .
train
false
48,660
def addCircleIntersectionLoop(circleIntersectionLoop, circleIntersections): firstCircleIntersection = circleIntersectionLoop[0] circleIntersectionAhead = firstCircleIntersection for circleIntersectionIndex in xrange((len(circleIntersections) + 1)): circleIntersectionAhead = circleIntersectionAhead.getCircleIntersectionAhead() if ((circleIntersectionAhead == firstCircleIntersection) or (circleIntersectionAhead == None)): firstCircleIntersection.steppedOn = True return circleIntersectionAhead.addToList(circleIntersectionLoop) firstCircleIntersection.steppedOn = True print 'Warning, addCircleIntersectionLoop would have gone into an endless loop.' print 'circleIntersectionLoop' for circleIntersection in circleIntersectionLoop: print circleIntersection print circleIntersection.circleNodeAhead print circleIntersection.circleNodeBehind print 'firstCircleIntersection' print firstCircleIntersection print 'circleIntersections' for circleIntersection in circleIntersections: print circleIntersection
[ "def", "addCircleIntersectionLoop", "(", "circleIntersectionLoop", ",", "circleIntersections", ")", ":", "firstCircleIntersection", "=", "circleIntersectionLoop", "[", "0", "]", "circleIntersectionAhead", "=", "firstCircleIntersection", "for", "circleIntersectionIndex", "in", "xrange", "(", "(", "len", "(", "circleIntersections", ")", "+", "1", ")", ")", ":", "circleIntersectionAhead", "=", "circleIntersectionAhead", ".", "getCircleIntersectionAhead", "(", ")", "if", "(", "(", "circleIntersectionAhead", "==", "firstCircleIntersection", ")", "or", "(", "circleIntersectionAhead", "==", "None", ")", ")", ":", "firstCircleIntersection", ".", "steppedOn", "=", "True", "return", "circleIntersectionAhead", ".", "addToList", "(", "circleIntersectionLoop", ")", "firstCircleIntersection", ".", "steppedOn", "=", "True", "print", "'Warning, addCircleIntersectionLoop would have gone into an endless loop.'", "print", "'circleIntersectionLoop'", "for", "circleIntersection", "in", "circleIntersectionLoop", ":", "print", "circleIntersection", "print", "circleIntersection", ".", "circleNodeAhead", "print", "circleIntersection", ".", "circleNodeBehind", "print", "'firstCircleIntersection'", "print", "firstCircleIntersection", "print", "'circleIntersections'", "for", "circleIntersection", "in", "circleIntersections", ":", "print", "circleIntersection" ]
add a circle intersection loop .
train
false
48,661
def sync_with_om(func): def om_wrapper(*args, **kwds): om.manager.process_all_messages() return func(*args, **kwds) return om_wrapper
[ "def", "sync_with_om", "(", "func", ")", ":", "def", "om_wrapper", "(", "*", "args", ",", "**", "kwds", ")", ":", "om", ".", "manager", ".", "process_all_messages", "(", ")", "return", "func", "(", "*", "args", ",", "**", "kwds", ")", "return", "om_wrapper" ]
given that the output manager has been migrated into a producer/consumer model .
train
false
48,663
def _append_comment(ret, comment): if len(ret['comment']): ret['comment'] = ((ret['comment'].rstrip() + '\n') + comment) else: ret['comment'] = comment return ret
[ "def", "_append_comment", "(", "ret", ",", "comment", ")", ":", "if", "len", "(", "ret", "[", "'comment'", "]", ")", ":", "ret", "[", "'comment'", "]", "=", "(", "(", "ret", "[", "'comment'", "]", ".", "rstrip", "(", ")", "+", "'\\n'", ")", "+", "comment", ")", "else", ":", "ret", "[", "'comment'", "]", "=", "comment", "return", "ret" ]
append comment to ret[comment] .
train
true
48,664
def construct_query_part(model_cls, prefixes, query_part): if (not query_part): return query.TrueQuery() query_classes = {} for (k, t) in itertools.chain(model_cls._fields.items(), model_cls._types.items()): query_classes[k] = t.query (key, pattern, query_class) = parse_query_part(query_part, query_classes, prefixes) if (key is None): if issubclass(query_class, query.FieldQuery): return query.AnyFieldQuery(pattern, model_cls._search_fields, query_class) else: return query_class(pattern) key = key.lower() return query_class(key.lower(), pattern, (key in model_cls._fields))
[ "def", "construct_query_part", "(", "model_cls", ",", "prefixes", ",", "query_part", ")", ":", "if", "(", "not", "query_part", ")", ":", "return", "query", ".", "TrueQuery", "(", ")", "query_classes", "=", "{", "}", "for", "(", "k", ",", "t", ")", "in", "itertools", ".", "chain", "(", "model_cls", ".", "_fields", ".", "items", "(", ")", ",", "model_cls", ".", "_types", ".", "items", "(", ")", ")", ":", "query_classes", "[", "k", "]", "=", "t", ".", "query", "(", "key", ",", "pattern", ",", "query_class", ")", "=", "parse_query_part", "(", "query_part", ",", "query_classes", ",", "prefixes", ")", "if", "(", "key", "is", "None", ")", ":", "if", "issubclass", "(", "query_class", ",", "query", ".", "FieldQuery", ")", ":", "return", "query", ".", "AnyFieldQuery", "(", "pattern", ",", "model_cls", ".", "_search_fields", ",", "query_class", ")", "else", ":", "return", "query_class", "(", "pattern", ")", "key", "=", "key", ".", "lower", "(", ")", "return", "query_class", "(", "key", ".", "lower", "(", ")", ",", "pattern", ",", "(", "key", "in", "model_cls", ".", "_fields", ")", ")" ]
create a query from a single query component .
train
false
48,666
def get_console_size(): display_width = get_option('display.width') display_height = get_option('display.height', silent=True) if com.in_interactive_session(): if com.in_ipython_frontend(): from pandas.core.config import get_default_val terminal_width = get_default_val('display.width') terminal_height = get_default_val('display.height') else: (terminal_width, terminal_height) = get_terminal_size() else: (terminal_width, terminal_height) = (None, None) return ((display_width or terminal_width), (display_height or terminal_height))
[ "def", "get_console_size", "(", ")", ":", "display_width", "=", "get_option", "(", "'display.width'", ")", "display_height", "=", "get_option", "(", "'display.height'", ",", "silent", "=", "True", ")", "if", "com", ".", "in_interactive_session", "(", ")", ":", "if", "com", ".", "in_ipython_frontend", "(", ")", ":", "from", "pandas", ".", "core", ".", "config", "import", "get_default_val", "terminal_width", "=", "get_default_val", "(", "'display.width'", ")", "terminal_height", "=", "get_default_val", "(", "'display.height'", ")", "else", ":", "(", "terminal_width", ",", "terminal_height", ")", "=", "get_terminal_size", "(", ")", "else", ":", "(", "terminal_width", ",", "terminal_height", ")", "=", "(", "None", ",", "None", ")", "return", "(", "(", "display_width", "or", "terminal_width", ")", ",", "(", "display_height", "or", "terminal_height", ")", ")" ]
return console size as tuple = .
train
false
48,667
def patch_environ(case, key, value): old_environ = os.environ.copy() def cleanup(): os.environ.clear() os.environ.update(old_environ) os.environ[key] = value case.addCleanup(cleanup)
[ "def", "patch_environ", "(", "case", ",", "key", ",", "value", ")", ":", "old_environ", "=", "os", ".", "environ", ".", "copy", "(", ")", "def", "cleanup", "(", ")", ":", "os", ".", "environ", ".", "clear", "(", ")", "os", ".", "environ", ".", "update", "(", "old_environ", ")", "os", ".", "environ", "[", "key", "]", "=", "value", "case", ".", "addCleanup", "(", "cleanup", ")" ]
add an environment variable for the duration of a test .
train
false
48,668
def get_course_info_section(request, user, course, section_key): info_module = get_course_info_section_module(request, user, course, section_key) html = '' if (info_module is not None): try: html = info_module.render(STUDENT_VIEW).content except Exception: html = render_to_string('courseware/error-message.html', None) log.exception(u'Error rendering course_id=%s, section_key=%s', unicode(course.id), section_key) return html
[ "def", "get_course_info_section", "(", "request", ",", "user", ",", "course", ",", "section_key", ")", ":", "info_module", "=", "get_course_info_section_module", "(", "request", ",", "user", ",", "course", ",", "section_key", ")", "html", "=", "''", "if", "(", "info_module", "is", "not", "None", ")", ":", "try", ":", "html", "=", "info_module", ".", "render", "(", "STUDENT_VIEW", ")", ".", "content", "except", "Exception", ":", "html", "=", "render_to_string", "(", "'courseware/error-message.html'", ",", "None", ")", "log", ".", "exception", "(", "u'Error rendering course_id=%s, section_key=%s'", ",", "unicode", "(", "course", ".", "id", ")", ",", "section_key", ")", "return", "html" ]
this returns the snippet of html to be rendered on the course info page .
train
false
48,669
def get_page(self, suffix): try: return int(self.REQUEST[('page%s' % suffix)]) except (KeyError, ValueError, TypeError): return 1
[ "def", "get_page", "(", "self", ",", "suffix", ")", ":", "try", ":", "return", "int", "(", "self", ".", "REQUEST", "[", "(", "'page%s'", "%", "suffix", ")", "]", ")", "except", "(", "KeyError", ",", "ValueError", ",", "TypeError", ")", ":", "return", "1" ]
a function which will be monkeypatched onto the request to get the current integer representing the current page .
train
false
48,670
def rsync_ip(ip): try: socket.inet_pton(socket.AF_INET6, ip) except socket.error: return ip else: return ('[%s]' % ip)
[ "def", "rsync_ip", "(", "ip", ")", ":", "try", ":", "socket", ".", "inet_pton", "(", "socket", ".", "AF_INET6", ",", "ip", ")", "except", "socket", ".", "error", ":", "return", "ip", "else", ":", "return", "(", "'[%s]'", "%", "ip", ")" ]
transform ip string to an rsync-compatible form will return ipv4 addresses unchanged .
train
false
48,671
@require_GET def user_lookup(request): userlist = [] if request.is_ajax(): user = request.GET.get('user', '') if user: matches = get_user_model().objects.filter(username__istartswith=user) for match in matches: userlist.append({'label': match.username}) data = json.dumps(userlist) return HttpResponse(data, content_type='application/json; charset=utf-8')
[ "@", "require_GET", "def", "user_lookup", "(", "request", ")", ":", "userlist", "=", "[", "]", "if", "request", ".", "is_ajax", "(", ")", ":", "user", "=", "request", ".", "GET", ".", "get", "(", "'user'", ",", "''", ")", "if", "user", ":", "matches", "=", "get_user_model", "(", ")", ".", "objects", ".", "filter", "(", "username__istartswith", "=", "user", ")", "for", "match", "in", "matches", ":", "userlist", ".", "append", "(", "{", "'label'", ":", "match", ".", "username", "}", ")", "data", "=", "json", ".", "dumps", "(", "userlist", ")", "return", "HttpResponse", "(", "data", ",", "content_type", "=", "'application/json; charset=utf-8'", ")" ]
returns partial username matches .
train
false
48,672
@not_implemented_for('undirected') def stochastic_graph(G, copy=True, weight='weight'): if copy: G = (MultiDiGraph(G) if G.is_multigraph() else DiGraph(G)) degree = dict(G.out_degree(weight=weight)) for (u, v, d) in G.edges(data=True): if (degree[u] == 0): d[weight] = 0 else: d[weight] = (d.get(weight, 1) / degree[u]) return G
[ "@", "not_implemented_for", "(", "'undirected'", ")", "def", "stochastic_graph", "(", "G", ",", "copy", "=", "True", ",", "weight", "=", "'weight'", ")", ":", "if", "copy", ":", "G", "=", "(", "MultiDiGraph", "(", "G", ")", "if", "G", ".", "is_multigraph", "(", ")", "else", "DiGraph", "(", "G", ")", ")", "degree", "=", "dict", "(", "G", ".", "out_degree", "(", "weight", "=", "weight", ")", ")", "for", "(", "u", ",", "v", ",", "d", ")", "in", "G", ".", "edges", "(", "data", "=", "True", ")", ":", "if", "(", "degree", "[", "u", "]", "==", "0", ")", ":", "d", "[", "weight", "]", "=", "0", "else", ":", "d", "[", "weight", "]", "=", "(", "d", ".", "get", "(", "weight", ",", "1", ")", "/", "degree", "[", "u", "]", ")", "return", "G" ]
returns a right-stochastic representation of directed graph g .
train
false
48,674
def relocate_vm_spec(client_factory, datastore=None, host=None, disk_move_type='moveAllDiskBackingsAndAllowSharing'): rel_spec = client_factory.create('ns0:VirtualMachineRelocateSpec') rel_spec.datastore = datastore rel_spec.diskMoveType = disk_move_type rel_spec.host = host return rel_spec
[ "def", "relocate_vm_spec", "(", "client_factory", ",", "datastore", "=", "None", ",", "host", "=", "None", ",", "disk_move_type", "=", "'moveAllDiskBackingsAndAllowSharing'", ")", ":", "rel_spec", "=", "client_factory", ".", "create", "(", "'ns0:VirtualMachineRelocateSpec'", ")", "rel_spec", ".", "datastore", "=", "datastore", "rel_spec", ".", "diskMoveType", "=", "disk_move_type", "rel_spec", ".", "host", "=", "host", "return", "rel_spec" ]
builds the vm relocation spec .
train
false
48,675
@cache_permission def can_accept_suggestion(user, translation): return can_edit(user, translation, 'trans.accept_suggestion')
[ "@", "cache_permission", "def", "can_accept_suggestion", "(", "user", ",", "translation", ")", ":", "return", "can_edit", "(", "user", ",", "translation", ",", "'trans.accept_suggestion'", ")" ]
checks whether user can accept suggestions to given translation .
train
false
48,676
def site_disabled(config): disable_site(config) reload_service('apache2')
[ "def", "site_disabled", "(", "config", ")", ":", "disable_site", "(", "config", ")", "reload_service", "(", "'apache2'", ")" ]
require an apache site to be disabled .
train
false
48,677
def test_for(item, min_version=None, callback=extract_version): try: check = import_item(item) except (ImportError, RuntimeError): return False else: if min_version: if callback: check = callback(check) return (check >= min_version) else: return True
[ "def", "test_for", "(", "item", ",", "min_version", "=", "None", ",", "callback", "=", "extract_version", ")", ":", "try", ":", "check", "=", "import_item", "(", "item", ")", "except", "(", "ImportError", ",", "RuntimeError", ")", ":", "return", "False", "else", ":", "if", "min_version", ":", "if", "callback", ":", "check", "=", "callback", "(", "check", ")", "return", "(", "check", ">=", "min_version", ")", "else", ":", "return", "True" ]
test to see if item is importable .
train
false
48,678
def _get_instance_changes(current, state): current_keys = set(current.keys()) state_keys = set(state.keys()) changed = salt.utils.compare_dicts(current, state) for change in changed.keys(): if ((change in changed) and (changed[change]['old'] == '')): del changed[change] if ((change in changed) and (changed[change]['new'] == '')): del changed[change] return changed
[ "def", "_get_instance_changes", "(", "current", ",", "state", ")", ":", "current_keys", "=", "set", "(", "current", ".", "keys", "(", ")", ")", "state_keys", "=", "set", "(", "state", ".", "keys", "(", ")", ")", "changed", "=", "salt", ".", "utils", ".", "compare_dicts", "(", "current", ",", "state", ")", "for", "change", "in", "changed", ".", "keys", "(", ")", ":", "if", "(", "(", "change", "in", "changed", ")", "and", "(", "changed", "[", "change", "]", "[", "'old'", "]", "==", "''", ")", ")", ":", "del", "changed", "[", "change", "]", "if", "(", "(", "change", "in", "changed", ")", "and", "(", "changed", "[", "change", "]", "[", "'new'", "]", "==", "''", ")", ")", ":", "del", "changed", "[", "change", "]", "return", "changed" ]
get modified properties .
train
true
48,680
def _ParseResponse(response): response.reset() parser = feedparser.FeedParser() parser._set_headersonly() while True: line = response.readline() if (not feedparser.headerRE.match(line)): if (not feedparser.NLCRE.match(line)): parser.feed(line) break parser.feed(line) parsed_response = parser.close() if ('Status' in parsed_response): status = int(parsed_response['Status'].split(' ', 1)[0]) del parsed_response['Status'] else: status = 200 return {'body': (parsed_response.get_payload() + response.read()), 'headers': parsed_response.items(), 'response_code': status}
[ "def", "_ParseResponse", "(", "response", ")", ":", "response", ".", "reset", "(", ")", "parser", "=", "feedparser", ".", "FeedParser", "(", ")", "parser", ".", "_set_headersonly", "(", ")", "while", "True", ":", "line", "=", "response", ".", "readline", "(", ")", "if", "(", "not", "feedparser", ".", "headerRE", ".", "match", "(", "line", ")", ")", ":", "if", "(", "not", "feedparser", ".", "NLCRE", ".", "match", "(", "line", ")", ")", ":", "parser", ".", "feed", "(", "line", ")", "break", "parser", ".", "feed", "(", "line", ")", "parsed_response", "=", "parser", ".", "close", "(", ")", "if", "(", "'Status'", "in", "parsed_response", ")", ":", "status", "=", "int", "(", "parsed_response", "[", "'Status'", "]", ".", "split", "(", "' '", ",", "1", ")", "[", "0", "]", ")", "del", "parsed_response", "[", "'Status'", "]", "else", ":", "status", "=", "200", "return", "{", "'body'", ":", "(", "parsed_response", ".", "get_payload", "(", ")", "+", "response", ".", "read", "(", ")", ")", ",", "'headers'", ":", "parsed_response", ".", "items", "(", ")", ",", "'response_code'", ":", "status", "}" ]
parses an http response into a dict .
train
false
48,681
def servers(): ntp_servers = __proxy__['napalm.call']('get_ntp_servers', **{}) if (not ntp_servers.get('result')): return ntp_servers ntp_servers_list = list(ntp_servers.get('out', {}).keys()) ntp_servers['out'] = ntp_servers_list return ntp_servers
[ "def", "servers", "(", ")", ":", "ntp_servers", "=", "__proxy__", "[", "'napalm.call'", "]", "(", "'get_ntp_servers'", ",", "**", "{", "}", ")", "if", "(", "not", "ntp_servers", ".", "get", "(", "'result'", ")", ")", ":", "return", "ntp_servers", "ntp_servers_list", "=", "list", "(", "ntp_servers", ".", "get", "(", "'out'", ",", "{", "}", ")", ".", "keys", "(", ")", ")", "ntp_servers", "[", "'out'", "]", "=", "ntp_servers_list", "return", "ntp_servers" ]
returns a list of the configured ntp servers on the device .
train
false
48,682
def get_secrets(namespace, name='', apiserver_url=None, decode=False, brief=False): apiserver_url = _guess_apiserver(apiserver_url) if (apiserver_url is None): return False if (not decode): ret = _get_secrets(namespace, name, apiserver_url) else: ret = _decode_secrets(_get_secrets(namespace, name, apiserver_url)) return ret
[ "def", "get_secrets", "(", "namespace", ",", "name", "=", "''", ",", "apiserver_url", "=", "None", ",", "decode", "=", "False", ",", "brief", "=", "False", ")", ":", "apiserver_url", "=", "_guess_apiserver", "(", "apiserver_url", ")", "if", "(", "apiserver_url", "is", "None", ")", ":", "return", "False", "if", "(", "not", "decode", ")", ":", "ret", "=", "_get_secrets", "(", "namespace", ",", "name", ",", "apiserver_url", ")", "else", ":", "ret", "=", "_decode_secrets", "(", "_get_secrets", "(", "namespace", ",", "name", ",", "apiserver_url", ")", ")", "return", "ret" ]
get k8s namespaces cli example: .
train
true
48,683
def Laplace(name, mu, b): return rv(name, LaplaceDistribution, (mu, b))
[ "def", "Laplace", "(", "name", ",", "mu", ",", "b", ")", ":", "return", "rv", "(", "name", ",", "LaplaceDistribution", ",", "(", "mu", ",", "b", ")", ")" ]
create a continuous random variable with a laplace distribution .
train
false
48,684
def direct_delete_object(node, part, account, container, obj, conn_timeout=5, response_timeout=15, headers=None): if (headers is None): headers = {} headers = gen_headers(headers, add_ts=('x-timestamp' not in (k.lower() for k in headers))) path = ('/%s/%s/%s' % (account, container, obj)) _make_req(node, part, 'DELETE', path, headers, 'Object', conn_timeout, response_timeout)
[ "def", "direct_delete_object", "(", "node", ",", "part", ",", "account", ",", "container", ",", "obj", ",", "conn_timeout", "=", "5", ",", "response_timeout", "=", "15", ",", "headers", "=", "None", ")", ":", "if", "(", "headers", "is", "None", ")", ":", "headers", "=", "{", "}", "headers", "=", "gen_headers", "(", "headers", ",", "add_ts", "=", "(", "'x-timestamp'", "not", "in", "(", "k", ".", "lower", "(", ")", "for", "k", "in", "headers", ")", ")", ")", "path", "=", "(", "'/%s/%s/%s'", "%", "(", "account", ",", "container", ",", "obj", ")", ")", "_make_req", "(", "node", ",", "part", ",", "'DELETE'", ",", "path", ",", "headers", ",", "'Object'", ",", "conn_timeout", ",", "response_timeout", ")" ]
delete object directly from the object server .
train
false
48,685
def p_boolean_sliding_window(texts, segmented_topics, dictionary, window_size): top_ids = _ret_top_ids(segmented_topics) window_id = 0 per_topic_postings = {} token2id_dict = dictionary.token2id def add_topic_posting(top_ids, window, per_topic_postings, window_id, token2id_dict): for word in window: word_id = token2id_dict[word] if (word_id in top_ids): if (word_id in per_topic_postings): per_topic_postings[word_id].add(window_id) else: per_topic_postings[word_id] = set([window_id]) window_id += 1 return (window_id, per_topic_postings) for document in texts: it = iter(document) window = tuple(islice(it, window_size)) (window_id, per_topic_postings) = add_topic_posting(top_ids, window, per_topic_postings, window_id, token2id_dict) for elem in it: window = (window[1:] + (elem,)) (window_id, per_topic_postings) = add_topic_posting(top_ids, window, per_topic_postings, window_id, token2id_dict) return (per_topic_postings, window_id)
[ "def", "p_boolean_sliding_window", "(", "texts", ",", "segmented_topics", ",", "dictionary", ",", "window_size", ")", ":", "top_ids", "=", "_ret_top_ids", "(", "segmented_topics", ")", "window_id", "=", "0", "per_topic_postings", "=", "{", "}", "token2id_dict", "=", "dictionary", ".", "token2id", "def", "add_topic_posting", "(", "top_ids", ",", "window", ",", "per_topic_postings", ",", "window_id", ",", "token2id_dict", ")", ":", "for", "word", "in", "window", ":", "word_id", "=", "token2id_dict", "[", "word", "]", "if", "(", "word_id", "in", "top_ids", ")", ":", "if", "(", "word_id", "in", "per_topic_postings", ")", ":", "per_topic_postings", "[", "word_id", "]", ".", "add", "(", "window_id", ")", "else", ":", "per_topic_postings", "[", "word_id", "]", "=", "set", "(", "[", "window_id", "]", ")", "window_id", "+=", "1", "return", "(", "window_id", ",", "per_topic_postings", ")", "for", "document", "in", "texts", ":", "it", "=", "iter", "(", "document", ")", "window", "=", "tuple", "(", "islice", "(", "it", ",", "window_size", ")", ")", "(", "window_id", ",", "per_topic_postings", ")", "=", "add_topic_posting", "(", "top_ids", ",", "window", ",", "per_topic_postings", ",", "window_id", ",", "token2id_dict", ")", "for", "elem", "in", "it", ":", "window", "=", "(", "window", "[", "1", ":", "]", "+", "(", "elem", ",", ")", ")", "(", "window_id", ",", "per_topic_postings", ")", "=", "add_topic_posting", "(", "top_ids", ",", "window", ",", "per_topic_postings", ",", "window_id", ",", "token2id_dict", ")", "return", "(", "per_topic_postings", ",", "window_id", ")" ]
this function performs the boolean sliding window probability estimation .
train
false
48,686
def check_fill_values(data): assert_true((data['a'].mask == [False, False]).all()) assert_true((data['a'] == ['1', 'a']).all()) assert_true((data['b'].mask == [False, True]).all()) assert_true((data['b'] == [2, (-999)]).all()) data['b'].mask = False assert_true((data['b'] == [2, 1]).all())
[ "def", "check_fill_values", "(", "data", ")", ":", "assert_true", "(", "(", "data", "[", "'a'", "]", ".", "mask", "==", "[", "False", ",", "False", "]", ")", ".", "all", "(", ")", ")", "assert_true", "(", "(", "data", "[", "'a'", "]", "==", "[", "'1'", ",", "'a'", "]", ")", ".", "all", "(", ")", ")", "assert_true", "(", "(", "data", "[", "'b'", "]", ".", "mask", "==", "[", "False", ",", "True", "]", ")", ".", "all", "(", ")", ")", "assert_true", "(", "(", "data", "[", "'b'", "]", "==", "[", "2", ",", "(", "-", "999", ")", "]", ")", ".", "all", "(", ")", ")", "data", "[", "'b'", "]", ".", "mask", "=", "False", "assert_true", "(", "(", "data", "[", "'b'", "]", "==", "[", "2", ",", "1", "]", ")", ".", "all", "(", ")", ")" ]
compare array column by column with expectation .
train
false