id_within_dataset
int64
1
55.5k
snippet
stringlengths
19
14.2k
tokens
listlengths
6
1.63k
nl
stringlengths
6
352
split_within_dataset
stringclasses
1 value
is_duplicated
bool
2 classes
51,300
def euclidean(x, y): return sqrt(sum((((a - b) ** 2) for (a, b) in zip(x, y))))
[ "def", "euclidean", "(", "x", ",", "y", ")", ":", "return", "sqrt", "(", "sum", "(", "(", "(", "(", "a", "-", "b", ")", "**", "2", ")", "for", "(", "a", ",", "b", ")", "in", "zip", "(", "x", ",", "y", ")", ")", ")", ")" ]
returns the euclidean distance between the vectors x and y .
train
false
51,302
def isValidHostname(hostname): if (len(hostname) > 255): return False if (hostname[(-1):] == '.'): hostname = hostname[:(-1)] allowed = re.compile('(?!-)[A-Z\\d-]{1,63}(?<!-)$', re.IGNORECASE) return all((allowed.match(x) for x in hostname.split('.')))
[ "def", "isValidHostname", "(", "hostname", ")", ":", "if", "(", "len", "(", "hostname", ")", ">", "255", ")", ":", "return", "False", "if", "(", "hostname", "[", "(", "-", "1", ")", ":", "]", "==", "'.'", ")", ":", "hostname", "=", "hostname", "[", ":", "(", "-", "1", ")", "]", "allowed", "=", "re", ".", "compile", "(", "'(?!-)[A-Z\\\\d-]{1,63}(?<!-)$'", ",", "re", ".", "IGNORECASE", ")", "return", "all", "(", "(", "allowed", ".", "match", "(", "x", ")", "for", "x", "in", "hostname", ".", "split", "(", "'.'", ")", ")", ")" ]
try to validate the passed host name .
train
true
51,303
def cxUniformPartialyMatched(ind1, ind2, indpb): size = min(len(ind1), len(ind2)) (p1, p2) = (([0] * size), ([0] * size)) for i in xrange(size): p1[ind1[i]] = i p2[ind2[i]] = i for i in xrange(size): if (random.random() < indpb): temp1 = ind1[i] temp2 = ind2[i] (ind1[i], ind1[p1[temp2]]) = (temp2, temp1) (ind2[i], ind2[p2[temp1]]) = (temp1, temp2) (p1[temp1], p1[temp2]) = (p1[temp2], p1[temp1]) (p2[temp1], p2[temp2]) = (p2[temp2], p2[temp1]) return (ind1, ind2)
[ "def", "cxUniformPartialyMatched", "(", "ind1", ",", "ind2", ",", "indpb", ")", ":", "size", "=", "min", "(", "len", "(", "ind1", ")", ",", "len", "(", "ind2", ")", ")", "(", "p1", ",", "p2", ")", "=", "(", "(", "[", "0", "]", "*", "size", ")", ",", "(", "[", "0", "]", "*", "size", ")", ")", "for", "i", "in", "xrange", "(", "size", ")", ":", "p1", "[", "ind1", "[", "i", "]", "]", "=", "i", "p2", "[", "ind2", "[", "i", "]", "]", "=", "i", "for", "i", "in", "xrange", "(", "size", ")", ":", "if", "(", "random", ".", "random", "(", ")", "<", "indpb", ")", ":", "temp1", "=", "ind1", "[", "i", "]", "temp2", "=", "ind2", "[", "i", "]", "(", "ind1", "[", "i", "]", ",", "ind1", "[", "p1", "[", "temp2", "]", "]", ")", "=", "(", "temp2", ",", "temp1", ")", "(", "ind2", "[", "i", "]", ",", "ind2", "[", "p2", "[", "temp1", "]", "]", ")", "=", "(", "temp1", ",", "temp2", ")", "(", "p1", "[", "temp1", "]", ",", "p1", "[", "temp2", "]", ")", "=", "(", "p1", "[", "temp2", "]", ",", "p1", "[", "temp1", "]", ")", "(", "p2", "[", "temp1", "]", ",", "p2", "[", "temp2", "]", ")", "=", "(", "p2", "[", "temp2", "]", ",", "p2", "[", "temp1", "]", ")", "return", "(", "ind1", ",", "ind2", ")" ]
executes a uniform partially matched crossover on the input individuals .
train
false
51,304
def IsCppString(line): line = line.replace('\\\\', 'XX') return ((((line.count('"') - line.count('\\"')) - line.count('\'"\'')) & 1) == 1)
[ "def", "IsCppString", "(", "line", ")", ":", "line", "=", "line", ".", "replace", "(", "'\\\\\\\\'", ",", "'XX'", ")", "return", "(", "(", "(", "(", "line", ".", "count", "(", "'\"'", ")", "-", "line", ".", "count", "(", "'\\\\\"'", ")", ")", "-", "line", ".", "count", "(", "'\\'\"\\''", ")", ")", "&", "1", ")", "==", "1", ")" ]
does line terminate so .
train
false
51,305
def sublist_reverse(lst, a, b): while (b > a): (lst[a], lst[b]) = (lst[b], lst[a]) b -= 1 a += 1
[ "def", "sublist_reverse", "(", "lst", ",", "a", ",", "b", ")", ":", "while", "(", "b", ">", "a", ")", ":", "(", "lst", "[", "a", "]", ",", "lst", "[", "b", "]", ")", "=", "(", "lst", "[", "b", "]", ",", "lst", "[", "a", "]", ")", "b", "-=", "1", "a", "+=", "1" ]
reverses the elements lst[a:b] .
train
false
51,307
def build_network_settings(**settings): current_network_settings = _parse_rh_config(_RH_NETWORK_FILE) opts = _parse_network_settings(settings, current_network_settings) try: template = JINJA.get_template('network.jinja') except jinja2.exceptions.TemplateNotFound: log.error('Could not load template network.jinja') return '' network = template.render(opts) if settings['test']: return _read_temp(network) _write_file_network(network, _RH_NETWORK_FILE) return _read_file(_RH_NETWORK_FILE)
[ "def", "build_network_settings", "(", "**", "settings", ")", ":", "current_network_settings", "=", "_parse_rh_config", "(", "_RH_NETWORK_FILE", ")", "opts", "=", "_parse_network_settings", "(", "settings", ",", "current_network_settings", ")", "try", ":", "template", "=", "JINJA", ".", "get_template", "(", "'network.jinja'", ")", "except", "jinja2", ".", "exceptions", ".", "TemplateNotFound", ":", "log", ".", "error", "(", "'Could not load template network.jinja'", ")", "return", "''", "network", "=", "template", ".", "render", "(", "opts", ")", "if", "settings", "[", "'test'", "]", ":", "return", "_read_temp", "(", "network", ")", "_write_file_network", "(", "network", ",", "_RH_NETWORK_FILE", ")", "return", "_read_file", "(", "_RH_NETWORK_FILE", ")" ]
build the global network script .
train
true
51,308
def cache_model(model, timeout=None): if hasattr(model, 'get_cached'): return def clear_cache(sender, instance, *args, **kwargs): '\n Clears the cache for the given instance.\n ' delete_instance(sender, instance) post_save.connect(clear_cache, sender=model, weak=False) post_delete.connect(clear_cache, sender=model, weak=False) @classmethod def get(cls, pk, using=None): '\n Returns the model for the given primary key (pk).\n ' if (pk is None): return None return get_instance(cls, pk, timeout, using) model.get_cached = get
[ "def", "cache_model", "(", "model", ",", "timeout", "=", "None", ")", ":", "if", "hasattr", "(", "model", ",", "'get_cached'", ")", ":", "return", "def", "clear_cache", "(", "sender", ",", "instance", ",", "*", "args", ",", "**", "kwargs", ")", ":", "delete_instance", "(", "sender", ",", "instance", ")", "post_save", ".", "connect", "(", "clear_cache", ",", "sender", "=", "model", ",", "weak", "=", "False", ")", "post_delete", ".", "connect", "(", "clear_cache", ",", "sender", "=", "model", ",", "weak", "=", "False", ")", "@", "classmethod", "def", "get", "(", "cls", ",", "pk", ",", "using", "=", "None", ")", ":", "if", "(", "pk", "is", "None", ")", ":", "return", "None", "return", "get_instance", "(", "cls", ",", "pk", ",", "timeout", ",", "using", ")", "model", ".", "get_cached", "=", "get" ]
adds utility methods to the given model to obtain foreignkey instances via the cache .
train
false
51,309
@login_required def upload_manifest(*args, **kwargs): return _upload_manifest(*args, **kwargs)
[ "@", "login_required", "def", "upload_manifest", "(", "*", "args", ",", "**", "kwargs", ")", ":", "return", "_upload_manifest", "(", "*", "args", ",", "**", "kwargs", ")" ]
wrapper function for _upload_manifest so we can keep the standalone validator separate from the manifest upload stuff .
train
false
51,310
@given(u'we have pgcli installed') def step_install_cli(_): dists = set([di.key for di in pip.get_installed_distributions()]) assert (u'pgcli' in dists)
[ "@", "given", "(", "u'we have pgcli installed'", ")", "def", "step_install_cli", "(", "_", ")", ":", "dists", "=", "set", "(", "[", "di", ".", "key", "for", "di", "in", "pip", ".", "get_installed_distributions", "(", ")", "]", ")", "assert", "(", "u'pgcli'", "in", "dists", ")" ]
check that pgcli is in installed modules .
train
false
51,311
def p_command_def_bad_rhs(p): p[0] = 'BAD EXPRESSION IN DEF STATEMENT'
[ "def", "p_command_def_bad_rhs", "(", "p", ")", ":", "p", "[", "0", "]", "=", "'BAD EXPRESSION IN DEF STATEMENT'" ]
command : def id lparen id rparen equals error .
train
false
51,312
def track_from_id(identifier, timeout=DEFAULT_ASYNC_TIMEOUT): param_dict = dict(id=identifier) return _profile(param_dict, timeout)
[ "def", "track_from_id", "(", "identifier", ",", "timeout", "=", "DEFAULT_ASYNC_TIMEOUT", ")", ":", "param_dict", "=", "dict", "(", "id", "=", "identifier", ")", "return", "_profile", "(", "param_dict", ",", "timeout", ")" ]
create a track object from an echo nest track id .
train
true
51,313
def get_taskqueue_nodes(): nodes = file_io.read(constants.TASKQUEUE_NODE_FILE) nodes = nodes.split('\n') if (nodes[(-1)] == ''): nodes = nodes[:(-1)] return nodes
[ "def", "get_taskqueue_nodes", "(", ")", ":", "nodes", "=", "file_io", ".", "read", "(", "constants", ".", "TASKQUEUE_NODE_FILE", ")", "nodes", "=", "nodes", ".", "split", "(", "'\\n'", ")", "if", "(", "nodes", "[", "(", "-", "1", ")", "]", "==", "''", ")", ":", "nodes", "=", "nodes", "[", ":", "(", "-", "1", ")", "]", "return", "nodes" ]
returns a list of all the taskqueue nodes .
train
false
51,314
def inner(a, b): a_ndim = a.ndim b_ndim = b.ndim if ((a_ndim == 0) or (b_ndim == 0)): return cupy.multiply(a, b) a_axis = (a_ndim - 1) b_axis = (b_ndim - 1) if (a.shape[(-1)] != b.shape[(-1)]): raise ValueError('Axis dimension mismatch') if a_axis: a = cupy.rollaxis(a, a_axis, 0) if b_axis: b = cupy.rollaxis(b, b_axis, 0) ret_shape = (a.shape[1:] + b.shape[1:]) k = a.shape[0] n = (a.size // k) m = (b.size // k) return core.tensordot_core(a, b, None, n, m, k, ret_shape)
[ "def", "inner", "(", "a", ",", "b", ")", ":", "a_ndim", "=", "a", ".", "ndim", "b_ndim", "=", "b", ".", "ndim", "if", "(", "(", "a_ndim", "==", "0", ")", "or", "(", "b_ndim", "==", "0", ")", ")", ":", "return", "cupy", ".", "multiply", "(", "a", ",", "b", ")", "a_axis", "=", "(", "a_ndim", "-", "1", ")", "b_axis", "=", "(", "b_ndim", "-", "1", ")", "if", "(", "a", ".", "shape", "[", "(", "-", "1", ")", "]", "!=", "b", ".", "shape", "[", "(", "-", "1", ")", "]", ")", ":", "raise", "ValueError", "(", "'Axis dimension mismatch'", ")", "if", "a_axis", ":", "a", "=", "cupy", ".", "rollaxis", "(", "a", ",", "a_axis", ",", "0", ")", "if", "b_axis", ":", "b", "=", "cupy", ".", "rollaxis", "(", "b", ",", "b_axis", ",", "0", ")", "ret_shape", "=", "(", "a", ".", "shape", "[", "1", ":", "]", "+", "b", ".", "shape", "[", "1", ":", "]", ")", "k", "=", "a", ".", "shape", "[", "0", "]", "n", "=", "(", "a", ".", "size", "//", "k", ")", "m", "=", "(", "b", ".", "size", "//", "k", ")", "return", "core", ".", "tensordot_core", "(", "a", ",", "b", ",", "None", ",", "n", ",", "m", ",", "k", ",", "ret_shape", ")" ]
returns the inner product of two arrays .
train
false
51,315
def complete_course_mode_info(course_id, enrollment, modes=None): if (modes is None): modes = CourseMode.modes_for_course_dict(course_id) mode_info = {'show_upsell': False, 'days_for_upsell': None} if ((CourseMode.VERIFIED in modes) and (enrollment.mode in CourseMode.UPSELL_TO_VERIFIED_MODES)): mode_info['show_upsell'] = True mode_info['verified_sku'] = modes['verified'].sku mode_info['verified_bulk_sku'] = modes['verified'].bulk_sku if modes['verified'].expiration_datetime: today = datetime.datetime.now(UTC).date() mode_info['days_for_upsell'] = (modes['verified'].expiration_datetime.date() - today).days return mode_info
[ "def", "complete_course_mode_info", "(", "course_id", ",", "enrollment", ",", "modes", "=", "None", ")", ":", "if", "(", "modes", "is", "None", ")", ":", "modes", "=", "CourseMode", ".", "modes_for_course_dict", "(", "course_id", ")", "mode_info", "=", "{", "'show_upsell'", ":", "False", ",", "'days_for_upsell'", ":", "None", "}", "if", "(", "(", "CourseMode", ".", "VERIFIED", "in", "modes", ")", "and", "(", "enrollment", ".", "mode", "in", "CourseMode", ".", "UPSELL_TO_VERIFIED_MODES", ")", ")", ":", "mode_info", "[", "'show_upsell'", "]", "=", "True", "mode_info", "[", "'verified_sku'", "]", "=", "modes", "[", "'verified'", "]", ".", "sku", "mode_info", "[", "'verified_bulk_sku'", "]", "=", "modes", "[", "'verified'", "]", ".", "bulk_sku", "if", "modes", "[", "'verified'", "]", ".", "expiration_datetime", ":", "today", "=", "datetime", ".", "datetime", ".", "now", "(", "UTC", ")", ".", "date", "(", ")", "mode_info", "[", "'days_for_upsell'", "]", "=", "(", "modes", "[", "'verified'", "]", ".", "expiration_datetime", ".", "date", "(", ")", "-", "today", ")", ".", "days", "return", "mode_info" ]
we would like to compute some more information from the given course modes and the users current enrollment returns the given information: - whether to show the course upsell information - numbers of days until they cant upsell anymore .
train
false
51,316
def parse_qs(qs, keep_blank_values=0, strict_parsing=0, unquote=unquote): d = {} items = [s2 for s1 in qs.split('&') for s2 in s1.split(';')] for item in items: try: (k, v) = item.split('=', 1) except ValueError: if strict_parsing: raise continue if (v or keep_blank_values): k = unquote(k.replace('+', ' ')) v = unquote(v.replace('+', ' ')) if (k in d): d[k].append(v) else: d[k] = [v] return d
[ "def", "parse_qs", "(", "qs", ",", "keep_blank_values", "=", "0", ",", "strict_parsing", "=", "0", ",", "unquote", "=", "unquote", ")", ":", "d", "=", "{", "}", "items", "=", "[", "s2", "for", "s1", "in", "qs", ".", "split", "(", "'&'", ")", "for", "s2", "in", "s1", ".", "split", "(", "';'", ")", "]", "for", "item", "in", "items", ":", "try", ":", "(", "k", ",", "v", ")", "=", "item", ".", "split", "(", "'='", ",", "1", ")", "except", "ValueError", ":", "if", "strict_parsing", ":", "raise", "continue", "if", "(", "v", "or", "keep_blank_values", ")", ":", "k", "=", "unquote", "(", "k", ".", "replace", "(", "'+'", ",", "' '", ")", ")", "v", "=", "unquote", "(", "v", ".", "replace", "(", "'+'", ",", "' '", ")", ")", "if", "(", "k", "in", "d", ")", ":", "d", "[", "k", "]", ".", "append", "(", "v", ")", "else", ":", "d", "[", "k", "]", "=", "[", "v", "]", "return", "d" ]
parse a query given as a string argument .
train
false
51,317
def _absolute_url_staticfile(is_secure, name): url_path = staticfiles_storage.url(name) if urlparse.urlparse(url_path).netloc: return url_path return _absolute_url(is_secure, url_path)
[ "def", "_absolute_url_staticfile", "(", "is_secure", ",", "name", ")", ":", "url_path", "=", "staticfiles_storage", ".", "url", "(", "name", ")", "if", "urlparse", ".", "urlparse", "(", "url_path", ")", ".", "netloc", ":", "return", "url_path", "return", "_absolute_url", "(", "is_secure", ",", "url_path", ")" ]
construct an absolute url to a static resource on the site .
train
false
51,318
def ordered_set(iterable): mmap = {} ord_set = [] for item in iterable: if (item not in mmap): mmap[item] = 1 ord_set.append(item) return ord_set
[ "def", "ordered_set", "(", "iterable", ")", ":", "mmap", "=", "{", "}", "ord_set", "=", "[", "]", "for", "item", "in", "iterable", ":", "if", "(", "item", "not", "in", "mmap", ")", ":", "mmap", "[", "item", "]", "=", "1", "ord_set", ".", "append", "(", "item", ")", "return", "ord_set" ]
creates an ordered list from strings .
train
false
51,319
def has_nibabel(vox2ras_tkr=False): try: import nibabel out = True if vox2ras_tkr: out = (getattr(getattr(getattr(nibabel, 'MGHImage', 0), 'header_class', 0), 'get_vox2ras_tkr', None) is not None) return out except ImportError: return False
[ "def", "has_nibabel", "(", "vox2ras_tkr", "=", "False", ")", ":", "try", ":", "import", "nibabel", "out", "=", "True", "if", "vox2ras_tkr", ":", "out", "=", "(", "getattr", "(", "getattr", "(", "getattr", "(", "nibabel", ",", "'MGHImage'", ",", "0", ")", ",", "'header_class'", ",", "0", ")", ",", "'get_vox2ras_tkr'", ",", "None", ")", "is", "not", "None", ")", "return", "out", "except", "ImportError", ":", "return", "False" ]
determine if nibabel is installed .
train
false
51,320
def default_json_error(ex): logger = get_logger() logger.error('Uncaught error thrown by Flask/Werkzeug', exc_info=ex) response = jsonify(message=str(ex), type='api_error') response.status_code = (ex.code if isinstance(ex, HTTPException) else 500) return response
[ "def", "default_json_error", "(", "ex", ")", ":", "logger", "=", "get_logger", "(", ")", "logger", ".", "error", "(", "'Uncaught error thrown by Flask/Werkzeug'", ",", "exc_info", "=", "ex", ")", "response", "=", "jsonify", "(", "message", "=", "str", "(", "ex", ")", ",", "type", "=", "'api_error'", ")", "response", ".", "status_code", "=", "(", "ex", ".", "code", "if", "isinstance", "(", "ex", ",", "HTTPException", ")", "else", "500", ")", "return", "response" ]
exception -> flask json responder .
train
false
51,321
def get_user_objects(current_user, user): objects = dict(USER_OBJECTS) for key in objects: if hasattr(user, key): manager = getattr(user, key) try: manager = manager.filter(status__hidden=False) except: pass objects[key]['objects'] = Object.filter_permitted(current_user, manager) return objects
[ "def", "get_user_objects", "(", "current_user", ",", "user", ")", ":", "objects", "=", "dict", "(", "USER_OBJECTS", ")", "for", "key", "in", "objects", ":", "if", "hasattr", "(", "user", ",", "key", ")", ":", "manager", "=", "getattr", "(", "user", ",", "key", ")", "try", ":", "manager", "=", "manager", ".", "filter", "(", "status__hidden", "=", "False", ")", "except", ":", "pass", "objects", "[", "key", "]", "[", "'objects'", "]", "=", "Object", ".", "filter_permitted", "(", "current_user", ",", "manager", ")", "return", "objects" ]
returns a dictionary with keys specified as contact attributes and values as dictionaries with labels and set of relevant objects .
train
false
51,322
@pytest.fixture(params=DISABLED_PROJECT_URL_PARAMS.keys()) def dp_view_urls(request, view_types): kwargs = DISABLED_PROJECT_URL_PARAMS[request.param].copy() view_name = kwargs.pop('view_name') view_name = ('%s-%s' % (view_name, view_types)) return reverse(view_name, kwargs=kwargs)
[ "@", "pytest", ".", "fixture", "(", "params", "=", "DISABLED_PROJECT_URL_PARAMS", ".", "keys", "(", ")", ")", "def", "dp_view_urls", "(", "request", ",", "view_types", ")", ":", "kwargs", "=", "DISABLED_PROJECT_URL_PARAMS", "[", "request", ".", "param", "]", ".", "copy", "(", ")", "view_name", "=", "kwargs", ".", "pop", "(", "'view_name'", ")", "view_name", "=", "(", "'%s-%s'", "%", "(", "view_name", ",", "view_types", ")", ")", "return", "reverse", "(", "view_name", ",", "kwargs", "=", "kwargs", ")" ]
list of url params required for disabled project tests .
train
false
51,323
def length_gt(value, arg): return (len(value) > int(arg))
[ "def", "length_gt", "(", "value", ",", "arg", ")", ":", "return", "(", "len", "(", "value", ")", ">", "int", "(", "arg", ")", ")" ]
returns a boolean of whether the values length is greater than the argument .
train
false
51,324
def make_unifrac_metric(weighted, metric, is_symmetric): def result(data, taxon_names, tree, sample_names, **kwargs): ' wraps the fast_unifrac fn to return just a matrix, in correct order\n\n sample_names: list of unique strings\n ' envs = make_envs_dict(data, sample_names, taxon_names) unifrac_res = fast_unifrac(tree, envs, weighted=weighted, metric=metric, is_symmetric=is_symmetric, modes=['distance_matrix'], **kwargs) dist_mtx = _reorder_unifrac_res(unifrac_res['distance_matrix'], sample_names) return dist_mtx return result
[ "def", "make_unifrac_metric", "(", "weighted", ",", "metric", ",", "is_symmetric", ")", ":", "def", "result", "(", "data", ",", "taxon_names", ",", "tree", ",", "sample_names", ",", "**", "kwargs", ")", ":", "envs", "=", "make_envs_dict", "(", "data", ",", "sample_names", ",", "taxon_names", ")", "unifrac_res", "=", "fast_unifrac", "(", "tree", ",", "envs", ",", "weighted", "=", "weighted", ",", "metric", "=", "metric", ",", "is_symmetric", "=", "is_symmetric", ",", "modes", "=", "[", "'distance_matrix'", "]", ",", "**", "kwargs", ")", "dist_mtx", "=", "_reorder_unifrac_res", "(", "unifrac_res", "[", "'distance_matrix'", "]", ",", "sample_names", ")", "return", "dist_mtx", "return", "result" ]
make a unifrac-like metric .
train
false
51,325
def set_default_color_scheme(name, replace=True): assert (name in sh.COLOR_SCHEME_NAMES) set_color_scheme(name, sh.get_color_scheme(name), replace=replace)
[ "def", "set_default_color_scheme", "(", "name", ",", "replace", "=", "True", ")", ":", "assert", "(", "name", "in", "sh", ".", "COLOR_SCHEME_NAMES", ")", "set_color_scheme", "(", "name", ",", "sh", ".", "get_color_scheme", "(", "name", ")", ",", "replace", "=", "replace", ")" ]
reset color scheme to default values .
train
true
51,326
def idd_findrank(eps, m, n, matvect): (k, ra, ier) = _id.idd_findrank(eps, m, n, matvect) if ier: raise _RETCODE_ERROR return k
[ "def", "idd_findrank", "(", "eps", ",", "m", ",", "n", ",", "matvect", ")", ":", "(", "k", ",", "ra", ",", "ier", ")", "=", "_id", ".", "idd_findrank", "(", "eps", ",", "m", ",", "n", ",", "matvect", ")", "if", "ier", ":", "raise", "_RETCODE_ERROR", "return", "k" ]
estimate rank of a real matrix to a specified relative precision using random matrix-vector multiplication .
train
false
51,327
def print_formatters(title=None, stream=None): from behave.formatter._registry import format_items from operator import itemgetter if (stream is None): stream = sys.stdout if title: stream.write((u'%s\n' % title)) format_items = sorted(format_items(resolved=True), key=itemgetter(0)) format_names = [item[0] for item in format_items] column_size = compute_words_maxsize(format_names) schema = ((u' %-' + _text(column_size)) + 's %s\n') for (name, formatter_class) in format_items: formatter_description = getattr(formatter_class, 'description', '') stream.write((schema % (name, formatter_description)))
[ "def", "print_formatters", "(", "title", "=", "None", ",", "stream", "=", "None", ")", ":", "from", "behave", ".", "formatter", ".", "_registry", "import", "format_items", "from", "operator", "import", "itemgetter", "if", "(", "stream", "is", "None", ")", ":", "stream", "=", "sys", ".", "stdout", "if", "title", ":", "stream", ".", "write", "(", "(", "u'%s\\n'", "%", "title", ")", ")", "format_items", "=", "sorted", "(", "format_items", "(", "resolved", "=", "True", ")", ",", "key", "=", "itemgetter", "(", "0", ")", ")", "format_names", "=", "[", "item", "[", "0", "]", "for", "item", "in", "format_items", "]", "column_size", "=", "compute_words_maxsize", "(", "format_names", ")", "schema", "=", "(", "(", "u' %-'", "+", "_text", "(", "column_size", ")", ")", "+", "'s %s\\n'", ")", "for", "(", "name", ",", "formatter_class", ")", "in", "format_items", ":", "formatter_description", "=", "getattr", "(", "formatter_class", ",", "'description'", ",", "''", ")", "stream", ".", "write", "(", "(", "schema", "%", "(", "name", ",", "formatter_description", ")", ")", ")" ]
prints the list of available formatters and their description .
train
false
51,328
def inertia_of_point_mass(mass, pos_vec, frame): return (mass * (((((frame.x | frame.x) + (frame.y | frame.y)) + (frame.z | frame.z)) * (pos_vec & pos_vec)) - (pos_vec | pos_vec)))
[ "def", "inertia_of_point_mass", "(", "mass", ",", "pos_vec", ",", "frame", ")", ":", "return", "(", "mass", "*", "(", "(", "(", "(", "(", "frame", ".", "x", "|", "frame", ".", "x", ")", "+", "(", "frame", ".", "y", "|", "frame", ".", "y", ")", ")", "+", "(", "frame", ".", "z", "|", "frame", ".", "z", ")", ")", "*", "(", "pos_vec", "&", "pos_vec", ")", ")", "-", "(", "pos_vec", "|", "pos_vec", ")", ")", ")" ]
inertia dyadic of a point mass relative to point o .
train
false
51,329
def _normalize_interval(start, end, value): if (not isinstance(start, datetime)): start = datetime.combine(start, START_OF_DAY) end = datetime.combine(end, START_OF_DAY) if (start.tzinfo is None): start = pytz.UTC.localize(start) end = pytz.UTC.localize(end) else: start = start.astimezone(pytz.UTC) end = end.astimezone(pytz.UTC) return (start, end)
[ "def", "_normalize_interval", "(", "start", ",", "end", ",", "value", ")", ":", "if", "(", "not", "isinstance", "(", "start", ",", "datetime", ")", ")", ":", "start", "=", "datetime", ".", "combine", "(", "start", ",", "START_OF_DAY", ")", "end", "=", "datetime", ".", "combine", "(", "end", ",", "START_OF_DAY", ")", "if", "(", "start", ".", "tzinfo", "is", "None", ")", ":", "start", "=", "pytz", ".", "UTC", ".", "localize", "(", "start", ")", "end", "=", "pytz", ".", "UTC", ".", "localize", "(", "end", ")", "else", ":", "start", "=", "start", ".", "astimezone", "(", "pytz", ".", "UTC", ")", "end", "=", "end", ".", "astimezone", "(", "pytz", ".", "UTC", ")", "return", "(", "start", ",", "end", ")" ]
normalize datetime intervals .
train
true
51,330
def _check_versions(current_release): feconf_changed_version = [] git_show_cmd = (GIT_CMD_SHOW_FORMAT_STRING % current_release) old_feconf = _run_cmd(git_show_cmd) with open('feconf.py', 'r') as feconf: new_feconf = feconf.read() for variable in FECONF_VAR_NAMES: old_version = re.findall((VERSION_RE_FORMAT_STRING % variable), old_feconf)[0] new_version = re.findall((VERSION_RE_FORMAT_STRING % variable), new_feconf)[0] if (old_version != new_version): feconf_changed_version.append(variable) return feconf_changed_version
[ "def", "_check_versions", "(", "current_release", ")", ":", "feconf_changed_version", "=", "[", "]", "git_show_cmd", "=", "(", "GIT_CMD_SHOW_FORMAT_STRING", "%", "current_release", ")", "old_feconf", "=", "_run_cmd", "(", "git_show_cmd", ")", "with", "open", "(", "'feconf.py'", ",", "'r'", ")", "as", "feconf", ":", "new_feconf", "=", "feconf", ".", "read", "(", ")", "for", "variable", "in", "FECONF_VAR_NAMES", ":", "old_version", "=", "re", ".", "findall", "(", "(", "VERSION_RE_FORMAT_STRING", "%", "variable", ")", ",", "old_feconf", ")", "[", "0", "]", "new_version", "=", "re", ".", "findall", "(", "(", "VERSION_RE_FORMAT_STRING", "%", "variable", ")", ",", "new_feconf", ")", "[", "0", "]", "if", "(", "old_version", "!=", "new_version", ")", ":", "feconf_changed_version", ".", "append", "(", "variable", ")", "return", "feconf_changed_version" ]
checks if the versions for the exploration or collection schemas have changed .
train
false
51,331
def New_Page_Info(new_page): dom = etree.HTML(new_page) new_items = dom.xpath('//tr/td/a/text()') new_urls = dom.xpath('//tr/td/a/@href') assert (len(new_items) == len(new_urls)) return zip(new_items, new_urls)
[ "def", "New_Page_Info", "(", "new_page", ")", ":", "dom", "=", "etree", ".", "HTML", "(", "new_page", ")", "new_items", "=", "dom", ".", "xpath", "(", "'//tr/td/a/text()'", ")", "new_urls", "=", "dom", ".", "xpath", "(", "'//tr/td/a/@href'", ")", "assert", "(", "len", "(", "new_items", ")", "==", "len", "(", "new_urls", ")", ")", "return", "zip", "(", "new_items", ",", "new_urls", ")" ]
regex or xpath .
train
false
51,332
def _write_credentials_file(credentials_file, credentials): data = {'file_version': 2, 'credentials': {}} for (key, credential) in iteritems(credentials): credential_json = credential.to_json() encoded_credential = _helpers._from_bytes(base64.b64encode(_helpers._to_bytes(credential_json))) data['credentials'][key] = encoded_credential credentials_file.seek(0) json.dump(data, credentials_file) credentials_file.truncate()
[ "def", "_write_credentials_file", "(", "credentials_file", ",", "credentials", ")", ":", "data", "=", "{", "'file_version'", ":", "2", ",", "'credentials'", ":", "{", "}", "}", "for", "(", "key", ",", "credential", ")", "in", "iteritems", "(", "credentials", ")", ":", "credential_json", "=", "credential", ".", "to_json", "(", ")", "encoded_credential", "=", "_helpers", ".", "_from_bytes", "(", "base64", ".", "b64encode", "(", "_helpers", ".", "_to_bytes", "(", "credential_json", ")", ")", ")", "data", "[", "'credentials'", "]", "[", "key", "]", "=", "encoded_credential", "credentials_file", ".", "seek", "(", "0", ")", "json", ".", "dump", "(", "data", ",", "credentials_file", ")", "credentials_file", ".", "truncate", "(", ")" ]
writes credentials to a file .
train
true
51,333
def make_parser(parser_list=[]): for parser_name in (parser_list + default_parser_list): try: return _create_parser(parser_name) except ImportError as e: import sys if (parser_name in sys.modules): raise except SAXReaderNotAvailable: pass raise SAXReaderNotAvailable('No parsers found', None)
[ "def", "make_parser", "(", "parser_list", "=", "[", "]", ")", ":", "for", "parser_name", "in", "(", "parser_list", "+", "default_parser_list", ")", ":", "try", ":", "return", "_create_parser", "(", "parser_name", ")", "except", "ImportError", "as", "e", ":", "import", "sys", "if", "(", "parser_name", "in", "sys", ".", "modules", ")", ":", "raise", "except", "SAXReaderNotAvailable", ":", "pass", "raise", "SAXReaderNotAvailable", "(", "'No parsers found'", ",", "None", ")" ]
creates and returns a sax parser .
train
false
51,334
def raise_if_deadlock_error(operational_error, engine_name): re = _DEADLOCK_RE_DB.get(engine_name) if (re is None): return m = re.match(operational_error.message) if (not m): return raise exception.DBDeadlock(operational_error)
[ "def", "raise_if_deadlock_error", "(", "operational_error", ",", "engine_name", ")", ":", "re", "=", "_DEADLOCK_RE_DB", ".", "get", "(", "engine_name", ")", "if", "(", "re", "is", "None", ")", ":", "return", "m", "=", "re", ".", "match", "(", "operational_error", ".", "message", ")", "if", "(", "not", "m", ")", ":", "return", "raise", "exception", ".", "DBDeadlock", "(", "operational_error", ")" ]
raise dbdeadlock exception if operationalerror contains a deadlock condition .
train
false
51,335
def _found_target_class(module, name): members = inspect.getmembers(module, inspect.isclass) return [x[1] for x in members if (x[0] == name.capitalize())][0]
[ "def", "_found_target_class", "(", "module", ",", "name", ")", ":", "members", "=", "inspect", ".", "getmembers", "(", "module", ",", "inspect", ".", "isclass", ")", "return", "[", "x", "[", "1", "]", "for", "x", "in", "members", "if", "(", "x", "[", "0", "]", "==", "name", ".", "capitalize", "(", ")", ")", "]", "[", "0", "]" ]
searches for a class with the specific name: it should be equal to capitalized $name .
train
false
51,340
def lqmn(m, n, z): if ((not isscalar(m)) or (m < 0)): raise ValueError('m must be a non-negative integer.') if ((not isscalar(n)) or (n < 0)): raise ValueError('n must be a non-negative integer.') if (not isscalar(z)): raise ValueError('z must be scalar.') m = int(m) n = int(n) mm = max(1, m) nn = max(1, n) if iscomplex(z): (q, qd) = specfun.clqmn(mm, nn, z) else: (q, qd) = specfun.lqmn(mm, nn, z) return (q[:(m + 1), :(n + 1)], qd[:(m + 1), :(n + 1)])
[ "def", "lqmn", "(", "m", ",", "n", ",", "z", ")", ":", "if", "(", "(", "not", "isscalar", "(", "m", ")", ")", "or", "(", "m", "<", "0", ")", ")", ":", "raise", "ValueError", "(", "'m must be a non-negative integer.'", ")", "if", "(", "(", "not", "isscalar", "(", "n", ")", ")", "or", "(", "n", "<", "0", ")", ")", ":", "raise", "ValueError", "(", "'n must be a non-negative integer.'", ")", "if", "(", "not", "isscalar", "(", "z", ")", ")", ":", "raise", "ValueError", "(", "'z must be scalar.'", ")", "m", "=", "int", "(", "m", ")", "n", "=", "int", "(", "n", ")", "mm", "=", "max", "(", "1", ",", "m", ")", "nn", "=", "max", "(", "1", ",", "n", ")", "if", "iscomplex", "(", "z", ")", ":", "(", "q", ",", "qd", ")", "=", "specfun", ".", "clqmn", "(", "mm", ",", "nn", ",", "z", ")", "else", ":", "(", "q", ",", "qd", ")", "=", "specfun", ".", "lqmn", "(", "mm", ",", "nn", ",", "z", ")", "return", "(", "q", "[", ":", "(", "m", "+", "1", ")", ",", ":", "(", "n", "+", "1", ")", "]", ",", "qd", "[", ":", "(", "m", "+", "1", ")", ",", ":", "(", "n", "+", "1", ")", "]", ")" ]
sequence of associated legendre functions of the second kind .
train
false
51,341
def pspace(expr): expr = sympify(expr) rvs = random_symbols(expr) if (not rvs): raise ValueError(('Expression containing Random Variable expected, not %s' % expr)) if all(((rv.pspace == rvs[0].pspace) for rv in rvs)): return rvs[0].pspace return ProductPSpace(*[rv.pspace for rv in rvs])
[ "def", "pspace", "(", "expr", ")", ":", "expr", "=", "sympify", "(", "expr", ")", "rvs", "=", "random_symbols", "(", "expr", ")", "if", "(", "not", "rvs", ")", ":", "raise", "ValueError", "(", "(", "'Expression containing Random Variable expected, not %s'", "%", "expr", ")", ")", "if", "all", "(", "(", "(", "rv", ".", "pspace", "==", "rvs", "[", "0", "]", ".", "pspace", ")", "for", "rv", "in", "rvs", ")", ")", ":", "return", "rvs", "[", "0", "]", ".", "pspace", "return", "ProductPSpace", "(", "*", "[", "rv", ".", "pspace", "for", "rv", "in", "rvs", "]", ")" ]
returns the underlying probability space of a random expression .
train
false
51,342
def get_object_as_xml(inst, cls=None, root_tag_name=None, no_namespace=False): if (cls is None): cls = inst.__class__ if ((cls.get_namespace() is None) and (no_namespace is None)): no_namespace = True if (no_namespace is None): no_namespace = False parent = etree.Element('parent') xml_object.to_parent(None, cls, inst, parent, cls.get_namespace(), root_tag_name) if no_namespace: _dig(parent) etree.cleanup_namespaces(parent) return parent[0]
[ "def", "get_object_as_xml", "(", "inst", ",", "cls", "=", "None", ",", "root_tag_name", "=", "None", ",", "no_namespace", "=", "False", ")", ":", "if", "(", "cls", "is", "None", ")", ":", "cls", "=", "inst", ".", "__class__", "if", "(", "(", "cls", ".", "get_namespace", "(", ")", "is", "None", ")", "and", "(", "no_namespace", "is", "None", ")", ")", ":", "no_namespace", "=", "True", "if", "(", "no_namespace", "is", "None", ")", ":", "no_namespace", "=", "False", "parent", "=", "etree", ".", "Element", "(", "'parent'", ")", "xml_object", ".", "to_parent", "(", "None", ",", "cls", ",", "inst", ",", "parent", ",", "cls", ".", "get_namespace", "(", ")", ",", "root_tag_name", ")", "if", "no_namespace", ":", "_dig", "(", "parent", ")", "etree", ".", "cleanup_namespaces", "(", "parent", ")", "return", "parent", "[", "0", "]" ]
returns an elementtree representation of a :class:spyne .
train
false
51,344
def convert_genepop_to_fdist(gp_rec, report_pops=None): if hasattr(gp_rec, 'populations'): return _convert_genepop_to_fdist(gp_rec) else: return _convert_genepop_to_fdist_big(gp_rec, report_pops)
[ "def", "convert_genepop_to_fdist", "(", "gp_rec", ",", "report_pops", "=", "None", ")", ":", "if", "hasattr", "(", "gp_rec", ",", "'populations'", ")", ":", "return", "_convert_genepop_to_fdist", "(", "gp_rec", ")", "else", ":", "return", "_convert_genepop_to_fdist_big", "(", "gp_rec", ",", "report_pops", ")" ]
converts a genepop record to a fdist one .
train
false
51,345
def test_preserve_refs(): text = u'la philologie m&#x00e8;ne au pire' assert (strip_tags(text) == u'la philologie m\xe8ne au pire') text = u'la philologie m&#232;ne au pire' assert (strip_tags(text) == u'la philologie m\xe8ne au pire') text = u'veer &amp; wander' assert (strip_tags(text) == 'veer & wander')
[ "def", "test_preserve_refs", "(", ")", ":", "text", "=", "u'la philologie m&#x00e8;ne au pire'", "assert", "(", "strip_tags", "(", "text", ")", "==", "u'la philologie m\\xe8ne au pire'", ")", "text", "=", "u'la philologie m&#232;ne au pire'", "assert", "(", "strip_tags", "(", "text", ")", "==", "u'la philologie m\\xe8ne au pire'", ")", "text", "=", "u'veer &amp; wander'", "assert", "(", "strip_tags", "(", "text", ")", "==", "'veer & wander'", ")" ]
test that html character/entity references are preserved when we strip tags .
train
false
51,347
def set_output_volume(volume): cmd = 'osascript -e "set volume output volume {0}"'.format(volume) call = __salt__['cmd.run_all'](cmd, output_loglevel='debug', python_shell=False) _check_cmd(call) return get_output_volume()
[ "def", "set_output_volume", "(", "volume", ")", ":", "cmd", "=", "'osascript -e \"set volume output volume {0}\"'", ".", "format", "(", "volume", ")", "call", "=", "__salt__", "[", "'cmd.run_all'", "]", "(", "cmd", ",", "output_loglevel", "=", "'debug'", ",", "python_shell", "=", "False", ")", "_check_cmd", "(", "call", ")", "return", "get_output_volume", "(", ")" ]
set the volume of sound .
train
true
51,348
def components(f, x): result = set() if (x in f.free_symbols): if f.is_Symbol: result.add(f) elif (f.is_Function or f.is_Derivative): for g in f.args: result |= components(g, x) result.add(f) elif f.is_Pow: result |= components(f.base, x) if (not f.exp.is_Integer): if f.exp.is_Rational: result.add((f.base ** Rational(1, f.exp.q))) else: result |= (components(f.exp, x) | {f}) else: for g in f.args: result |= components(g, x) return result
[ "def", "components", "(", "f", ",", "x", ")", ":", "result", "=", "set", "(", ")", "if", "(", "x", "in", "f", ".", "free_symbols", ")", ":", "if", "f", ".", "is_Symbol", ":", "result", ".", "add", "(", "f", ")", "elif", "(", "f", ".", "is_Function", "or", "f", ".", "is_Derivative", ")", ":", "for", "g", "in", "f", ".", "args", ":", "result", "|=", "components", "(", "g", ",", "x", ")", "result", ".", "add", "(", "f", ")", "elif", "f", ".", "is_Pow", ":", "result", "|=", "components", "(", "f", ".", "base", ",", "x", ")", "if", "(", "not", "f", ".", "exp", ".", "is_Integer", ")", ":", "if", "f", ".", "exp", ".", "is_Rational", ":", "result", ".", "add", "(", "(", "f", ".", "base", "**", "Rational", "(", "1", ",", "f", ".", "exp", ".", "q", ")", ")", ")", "else", ":", "result", "|=", "(", "components", "(", "f", ".", "exp", ",", "x", ")", "|", "{", "f", "}", ")", "else", ":", "for", "g", "in", "f", ".", "args", ":", "result", "|=", "components", "(", "g", ",", "x", ")", "return", "result" ]
returns a set of all functional components of the given expression which includes symbols .
train
false
51,349
def knn_classify(k, labeled_points, new_point): by_distance = sorted(labeled_points, key=(lambda (point, _): distance(point, new_point))) k_nearest_labels = [label for (_, label) in by_distance[:k]] return majority_vote(k_nearest_labels)
[ "def", "knn_classify", "(", "k", ",", "labeled_points", ",", "new_point", ")", ":", "by_distance", "=", "sorted", "(", "labeled_points", ",", "key", "=", "(", "lambda", "(", "point", ",", "_", ")", ":", "distance", "(", "point", ",", "new_point", ")", ")", ")", "k_nearest_labels", "=", "[", "label", "for", "(", "_", ",", "label", ")", "in", "by_distance", "[", ":", "k", "]", "]", "return", "majority_vote", "(", "k_nearest_labels", ")" ]
each labeled point should be a pair .
train
false
51,351
def uptime(format=False): uptime = (time() - SERVER_START_TIME) if format: return _format(uptime, 31536000, 2628000, 604800, 86400, 3600, 60) return uptime
[ "def", "uptime", "(", "format", "=", "False", ")", ":", "uptime", "=", "(", "time", "(", ")", "-", "SERVER_START_TIME", ")", "if", "format", ":", "return", "_format", "(", "uptime", ",", "31536000", ",", "2628000", ",", "604800", ",", "86400", ",", "3600", ",", "60", ")", "return", "uptime" ]
get system load average from startup .
train
false
51,352
def getInteriorOverhangRadians(elementNode): return math.radians(getInteriorOverhangAngle(elementNode))
[ "def", "getInteriorOverhangRadians", "(", "elementNode", ")", ":", "return", "math", ".", "radians", "(", "getInteriorOverhangAngle", "(", "elementNode", ")", ")" ]
get the interior overhang support angle in radians .
train
false
51,354
def revert_snapshot(name, snap_name, runas=None): name = _sdecode(name) snap_name = _validate_snap_name(name, snap_name, runas=runas) args = [name, '--id', snap_name] return prlctl('snapshot-switch', args, runas=runas)
[ "def", "revert_snapshot", "(", "name", ",", "snap_name", ",", "runas", "=", "None", ")", ":", "name", "=", "_sdecode", "(", "name", ")", "snap_name", "=", "_validate_snap_name", "(", "name", ",", "snap_name", ",", "runas", "=", "runas", ")", "args", "=", "[", "name", ",", "'--id'", ",", "snap_name", "]", "return", "prlctl", "(", "'snapshot-switch'", ",", "args", ",", "runas", "=", "runas", ")" ]
revert a vm to a snapshot .
train
false
51,355
def foreign(expr): return _annotate_columns(expression._clause_element_as_expr(expr), {'foreign': True})
[ "def", "foreign", "(", "expr", ")", ":", "return", "_annotate_columns", "(", "expression", ".", "_clause_element_as_expr", "(", "expr", ")", ",", "{", "'foreign'", ":", "True", "}", ")" ]
annotate a portion of a primaryjoin expression with a foreign annotation .
train
false
51,356
def recv(shape, dtype, source, tag): return MPIRecvWait(tag)(*irecv(shape, dtype, source, tag))
[ "def", "recv", "(", "shape", ",", "dtype", ",", "source", ",", "tag", ")", ":", "return", "MPIRecvWait", "(", "tag", ")", "(", "*", "irecv", "(", "shape", ",", "dtype", ",", "source", ",", "tag", ")", ")" ]
blocking receive .
train
false
51,358
def test_stratified_validation_shuffle_split(): skip_if_no_sklearn() from pylearn2.cross_validation.subset_iterators import StratifiedValidationShuffleSplit n = 60 y = np.concatenate((np.zeros((n / 2), dtype=int), np.ones((n / 2), dtype=int))) cv = StratifiedValidationShuffleSplit(y) for (train, valid, test) in cv: assert (np.unique(np.concatenate((train, valid, test))).size == n) assert (valid.size == (n * cv.test_size)) assert (test.size == (n * cv.test_size)) assert (np.count_nonzero(y[valid]) == ((n / 2) * cv.test_size)) assert (np.count_nonzero(y[test]) == ((n / 2) * cv.test_size))
[ "def", "test_stratified_validation_shuffle_split", "(", ")", ":", "skip_if_no_sklearn", "(", ")", "from", "pylearn2", ".", "cross_validation", ".", "subset_iterators", "import", "StratifiedValidationShuffleSplit", "n", "=", "60", "y", "=", "np", ".", "concatenate", "(", "(", "np", ".", "zeros", "(", "(", "n", "/", "2", ")", ",", "dtype", "=", "int", ")", ",", "np", ".", "ones", "(", "(", "n", "/", "2", ")", ",", "dtype", "=", "int", ")", ")", ")", "cv", "=", "StratifiedValidationShuffleSplit", "(", "y", ")", "for", "(", "train", ",", "valid", ",", "test", ")", "in", "cv", ":", "assert", "(", "np", ".", "unique", "(", "np", ".", "concatenate", "(", "(", "train", ",", "valid", ",", "test", ")", ")", ")", ".", "size", "==", "n", ")", "assert", "(", "valid", ".", "size", "==", "(", "n", "*", "cv", ".", "test_size", ")", ")", "assert", "(", "test", ".", "size", "==", "(", "n", "*", "cv", ".", "test_size", ")", ")", "assert", "(", "np", ".", "count_nonzero", "(", "y", "[", "valid", "]", ")", "==", "(", "(", "n", "/", "2", ")", "*", "cv", ".", "test_size", ")", ")", "assert", "(", "np", ".", "count_nonzero", "(", "y", "[", "test", "]", ")", "==", "(", "(", "n", "/", "2", ")", "*", "cv", ".", "test_size", ")", ")" ]
test stratifiedvalidationshufflesplit .
train
false
51,360
def feq(a, b): t_float = Float('1.0E-10') return ((- t_float) < (a - b) < t_float)
[ "def", "feq", "(", "a", ",", "b", ")", ":", "t_float", "=", "Float", "(", "'1.0E-10'", ")", "return", "(", "(", "-", "t_float", ")", "<", "(", "a", "-", "b", ")", "<", "t_float", ")" ]
test if two floating point values are equal .
train
false
51,363
def _stringify_na_values(na_values): result = [] for x in na_values: result.append(str(x)) result.append(x) try: v = float(x) if (v == int(v)): v = int(v) result.append(('%s.0' % v)) result.append(str(v)) result.append(v) except: pass try: result.append(int(x)) except: pass return set(result)
[ "def", "_stringify_na_values", "(", "na_values", ")", ":", "result", "=", "[", "]", "for", "x", "in", "na_values", ":", "result", ".", "append", "(", "str", "(", "x", ")", ")", "result", ".", "append", "(", "x", ")", "try", ":", "v", "=", "float", "(", "x", ")", "if", "(", "v", "==", "int", "(", "v", ")", ")", ":", "v", "=", "int", "(", "v", ")", "result", ".", "append", "(", "(", "'%s.0'", "%", "v", ")", ")", "result", ".", "append", "(", "str", "(", "v", ")", ")", "result", ".", "append", "(", "v", ")", "except", ":", "pass", "try", ":", "result", ".", "append", "(", "int", "(", "x", ")", ")", "except", ":", "pass", "return", "set", "(", "result", ")" ]
return a stringified and numeric for these values .
train
false
51,364
def checkDynParam(place, parameter, value): if kb.redirectChoice: return None kb.matchRatio = None dynResult = None randInt = randomInt() paramType = (conf.method if (conf.method not in (None, HTTPMETHOD.GET, HTTPMETHOD.POST)) else place) infoMsg = ("testing if %s parameter '%s' is dynamic" % (paramType, parameter)) logger.info(infoMsg) try: payload = agent.payload(place, parameter, value, getUnicode(randInt)) dynResult = Request.queryPage(payload, place, raise404=False) if (not dynResult): infoMsg = ("confirming that %s parameter '%s' is dynamic" % (paramType, parameter)) logger.info(infoMsg) randInt = randomInt() payload = agent.payload(place, parameter, value, getUnicode(randInt)) dynResult = Request.queryPage(payload, place, raise404=False) except SqlmapConnectionException: pass result = (None if (dynResult is None) else (not dynResult)) kb.dynamicParameter = result return result
[ "def", "checkDynParam", "(", "place", ",", "parameter", ",", "value", ")", ":", "if", "kb", ".", "redirectChoice", ":", "return", "None", "kb", ".", "matchRatio", "=", "None", "dynResult", "=", "None", "randInt", "=", "randomInt", "(", ")", "paramType", "=", "(", "conf", ".", "method", "if", "(", "conf", ".", "method", "not", "in", "(", "None", ",", "HTTPMETHOD", ".", "GET", ",", "HTTPMETHOD", ".", "POST", ")", ")", "else", "place", ")", "infoMsg", "=", "(", "\"testing if %s parameter '%s' is dynamic\"", "%", "(", "paramType", ",", "parameter", ")", ")", "logger", ".", "info", "(", "infoMsg", ")", "try", ":", "payload", "=", "agent", ".", "payload", "(", "place", ",", "parameter", ",", "value", ",", "getUnicode", "(", "randInt", ")", ")", "dynResult", "=", "Request", ".", "queryPage", "(", "payload", ",", "place", ",", "raise404", "=", "False", ")", "if", "(", "not", "dynResult", ")", ":", "infoMsg", "=", "(", "\"confirming that %s parameter '%s' is dynamic\"", "%", "(", "paramType", ",", "parameter", ")", ")", "logger", ".", "info", "(", "infoMsg", ")", "randInt", "=", "randomInt", "(", ")", "payload", "=", "agent", ".", "payload", "(", "place", ",", "parameter", ",", "value", ",", "getUnicode", "(", "randInt", ")", ")", "dynResult", "=", "Request", ".", "queryPage", "(", "payload", ",", "place", ",", "raise404", "=", "False", ")", "except", "SqlmapConnectionException", ":", "pass", "result", "=", "(", "None", "if", "(", "dynResult", "is", "None", ")", "else", "(", "not", "dynResult", ")", ")", "kb", ".", "dynamicParameter", "=", "result", "return", "result" ]
this function checks if the url parameter is dynamic .
train
false
51,365
def get_ip_addresses(include_loopback=True): system = platform.system() if (system.lower() in ['linux', 'darwin', 'macosx']): ips = [iface.get('inet') for iface in ifcfg.interfaces().values()] elif (system.lower() == 'windows'): ipconfig = os.popen('ipconfig /all').read() ips = [match[1] for match in re.findall('IP(v4)? Address[\\.\\: ]+([\\d\\.]+)', ipconfig)] else: ips = [] ips = (set(ips) - set([None, ''])) if include_loopback: ips = ips.union(['127.0.0.1']) else: ips = (ips - set(['127.0.0.1'])) return list(ips)
[ "def", "get_ip_addresses", "(", "include_loopback", "=", "True", ")", ":", "system", "=", "platform", ".", "system", "(", ")", "if", "(", "system", ".", "lower", "(", ")", "in", "[", "'linux'", ",", "'darwin'", ",", "'macosx'", "]", ")", ":", "ips", "=", "[", "iface", ".", "get", "(", "'inet'", ")", "for", "iface", "in", "ifcfg", ".", "interfaces", "(", ")", ".", "values", "(", ")", "]", "elif", "(", "system", ".", "lower", "(", ")", "==", "'windows'", ")", ":", "ipconfig", "=", "os", ".", "popen", "(", "'ipconfig /all'", ")", ".", "read", "(", ")", "ips", "=", "[", "match", "[", "1", "]", "for", "match", "in", "re", ".", "findall", "(", "'IP(v4)? Address[\\\\.\\\\: ]+([\\\\d\\\\.]+)'", ",", "ipconfig", ")", "]", "else", ":", "ips", "=", "[", "]", "ips", "=", "(", "set", "(", "ips", ")", "-", "set", "(", "[", "None", ",", "''", "]", ")", ")", "if", "include_loopback", ":", "ips", "=", "ips", ".", "union", "(", "[", "'127.0.0.1'", "]", ")", "else", ":", "ips", "=", "(", "ips", "-", "set", "(", "[", "'127.0.0.1'", "]", ")", ")", "return", "list", "(", "ips", ")" ]
get a list of all the ip addresses for adapters on the local system .
train
false
51,366
def mock_render_to_response(*args, **kwargs): return render_to_response(*args, **kwargs)
[ "def", "mock_render_to_response", "(", "*", "args", ",", "**", "kwargs", ")", ":", "return", "render_to_response", "(", "*", "args", ",", "**", "kwargs", ")" ]
mock the render_to_response function .
train
false
51,367
@cronjobs.register def update_google_analytics(date=None): if date: date = datetime.datetime.strptime(date, '%Y-%m-%d').date() else: date = (datetime.date.today() - datetime.timedelta(days=1)) tasks.update_google_analytics.delay(date=date)
[ "@", "cronjobs", ".", "register", "def", "update_google_analytics", "(", "date", "=", "None", ")", ":", "if", "date", ":", "date", "=", "datetime", ".", "datetime", ".", "strptime", "(", "date", ",", "'%Y-%m-%d'", ")", ".", "date", "(", ")", "else", ":", "date", "=", "(", "datetime", ".", "date", ".", "today", "(", ")", "-", "datetime", ".", "timedelta", "(", "days", "=", "1", ")", ")", "tasks", ".", "update_google_analytics", ".", "delay", "(", "date", "=", "date", ")" ]
update stats from google analytics .
train
false
51,368
def get_domain_info(libvirt, host, virt_dom): def is_race(e): code = e.get_error_code() message = e.get_error_message() return ((code == libvirt.VIR_ERR_OPERATION_FAILED) and ('cannot read cputime for domain' in message)) try: return virt_dom.info() except libvirt.libvirtError as e: if ((not host.has_min_version((1, 2, 11))) and is_race(e)): LOG.warning(_LW('Race detected in libvirt.virDomain.info, trying one more time')) return virt_dom.info() raise
[ "def", "get_domain_info", "(", "libvirt", ",", "host", ",", "virt_dom", ")", ":", "def", "is_race", "(", "e", ")", ":", "code", "=", "e", ".", "get_error_code", "(", ")", "message", "=", "e", ".", "get_error_message", "(", ")", "return", "(", "(", "code", "==", "libvirt", ".", "VIR_ERR_OPERATION_FAILED", ")", "and", "(", "'cannot read cputime for domain'", "in", "message", ")", ")", "try", ":", "return", "virt_dom", ".", "info", "(", ")", "except", "libvirt", ".", "libvirtError", "as", "e", ":", "if", "(", "(", "not", "host", ".", "has_min_version", "(", "(", "1", ",", "2", ",", "11", ")", ")", ")", "and", "is_race", "(", "e", ")", ")", ":", "LOG", ".", "warning", "(", "_LW", "(", "'Race detected in libvirt.virDomain.info, trying one more time'", ")", ")", "return", "virt_dom", ".", "info", "(", ")", "raise" ]
method virdomain .
train
false
51,372
def task_id_str(task_family, params): param_str = json.dumps(params, separators=(',', ':'), sort_keys=True) param_hash = hashlib.md5(param_str.encode('utf-8')).hexdigest() param_summary = '_'.join((p[:TASK_ID_TRUNCATE_PARAMS] for p in (params[p] for p in sorted(params)[:TASK_ID_INCLUDE_PARAMS]))) param_summary = TASK_ID_INVALID_CHAR_REGEX.sub('_', param_summary) return '{}_{}_{}'.format(task_family, param_summary, param_hash[:TASK_ID_TRUNCATE_HASH])
[ "def", "task_id_str", "(", "task_family", ",", "params", ")", ":", "param_str", "=", "json", ".", "dumps", "(", "params", ",", "separators", "=", "(", "','", ",", "':'", ")", ",", "sort_keys", "=", "True", ")", "param_hash", "=", "hashlib", ".", "md5", "(", "param_str", ".", "encode", "(", "'utf-8'", ")", ")", ".", "hexdigest", "(", ")", "param_summary", "=", "'_'", ".", "join", "(", "(", "p", "[", ":", "TASK_ID_TRUNCATE_PARAMS", "]", "for", "p", "in", "(", "params", "[", "p", "]", "for", "p", "in", "sorted", "(", "params", ")", "[", ":", "TASK_ID_INCLUDE_PARAMS", "]", ")", ")", ")", "param_summary", "=", "TASK_ID_INVALID_CHAR_REGEX", ".", "sub", "(", "'_'", ",", "param_summary", ")", "return", "'{}_{}_{}'", ".", "format", "(", "task_family", ",", "param_summary", ",", "param_hash", "[", ":", "TASK_ID_TRUNCATE_HASH", "]", ")" ]
returns a canonical string used to identify a particular task .
train
true
51,373
@utils.expects_func_args('migration') def errors_out_migration(function): @functools.wraps(function) def decorated_function(self, context, *args, **kwargs): try: return function(self, context, *args, **kwargs) except Exception as ex: with excutils.save_and_reraise_exception(): wrapped_func = safe_utils.get_wrapped_function(function) keyed_args = inspect.getcallargs(wrapped_func, self, context, *args, **kwargs) migration = keyed_args['migration'] if (not isinstance(ex, exception.InstanceNotFound)): status = migration.status if (status not in ['migrating', 'post-migrating']): return migration.status = 'error' try: with migration.obj_as_admin(): migration.save() except Exception: LOG.debug('Error setting migration status for instance %s.', migration.instance_uuid, exc_info=True) return decorated_function
[ "@", "utils", ".", "expects_func_args", "(", "'migration'", ")", "def", "errors_out_migration", "(", "function", ")", ":", "@", "functools", ".", "wraps", "(", "function", ")", "def", "decorated_function", "(", "self", ",", "context", ",", "*", "args", ",", "**", "kwargs", ")", ":", "try", ":", "return", "function", "(", "self", ",", "context", ",", "*", "args", ",", "**", "kwargs", ")", "except", "Exception", "as", "ex", ":", "with", "excutils", ".", "save_and_reraise_exception", "(", ")", ":", "wrapped_func", "=", "safe_utils", ".", "get_wrapped_function", "(", "function", ")", "keyed_args", "=", "inspect", ".", "getcallargs", "(", "wrapped_func", ",", "self", ",", "context", ",", "*", "args", ",", "**", "kwargs", ")", "migration", "=", "keyed_args", "[", "'migration'", "]", "if", "(", "not", "isinstance", "(", "ex", ",", "exception", ".", "InstanceNotFound", ")", ")", ":", "status", "=", "migration", ".", "status", "if", "(", "status", "not", "in", "[", "'migrating'", ",", "'post-migrating'", "]", ")", ":", "return", "migration", ".", "status", "=", "'error'", "try", ":", "with", "migration", ".", "obj_as_admin", "(", ")", ":", "migration", ".", "save", "(", ")", "except", "Exception", ":", "LOG", ".", "debug", "(", "'Error setting migration status for instance %s.'", ",", "migration", ".", "instance_uuid", ",", "exc_info", "=", "True", ")", "return", "decorated_function" ]
decorator to error out migration on failure .
train
false
51,374
def test_finds_local_repo(tmpdir): project_dir = repository.determine_repo_dir('tests/fake-repo', abbreviations={}, clone_to_dir=str(tmpdir), checkout=None, no_input=True) assert ('tests/fake-repo' == project_dir)
[ "def", "test_finds_local_repo", "(", "tmpdir", ")", ":", "project_dir", "=", "repository", ".", "determine_repo_dir", "(", "'tests/fake-repo'", ",", "abbreviations", "=", "{", "}", ",", "clone_to_dir", "=", "str", "(", "tmpdir", ")", ",", "checkout", "=", "None", ",", "no_input", "=", "True", ")", "assert", "(", "'tests/fake-repo'", "==", "project_dir", ")" ]
a valid local repository should be returned .
train
false
51,375
def color_func(func_name): if str(func_name).isdigit(): return term_color(int(func_name)) return globals()[func_name]
[ "def", "color_func", "(", "func_name", ")", ":", "if", "str", "(", "func_name", ")", ".", "isdigit", "(", ")", ":", "return", "term_color", "(", "int", "(", "func_name", ")", ")", "return", "globals", "(", ")", "[", "func_name", "]" ]
call color function base on name .
train
true
51,376
def validate_config_section(filename, config, section): if (not isinstance(config, dict)): raise ConfigurationError(u"In file '{filename}', {section} must be a mapping, not {type}.".format(filename=filename, section=section, type=anglicize_json_type(python_type_to_yaml_type(config)))) for (key, value) in config.items(): if (not isinstance(key, six.string_types)): raise ConfigurationError(u"In file '{filename}', the {section} name {name} must be a quoted string, i.e. '{name}'.".format(filename=filename, section=section, name=key)) if (not isinstance(value, (dict, type(None)))): raise ConfigurationError(u"In file '{filename}', {section} '{name}' must be a mapping not {type}.".format(filename=filename, section=section, name=key, type=anglicize_json_type(python_type_to_yaml_type(value))))
[ "def", "validate_config_section", "(", "filename", ",", "config", ",", "section", ")", ":", "if", "(", "not", "isinstance", "(", "config", ",", "dict", ")", ")", ":", "raise", "ConfigurationError", "(", "u\"In file '{filename}', {section} must be a mapping, not {type}.\"", ".", "format", "(", "filename", "=", "filename", ",", "section", "=", "section", ",", "type", "=", "anglicize_json_type", "(", "python_type_to_yaml_type", "(", "config", ")", ")", ")", ")", "for", "(", "key", ",", "value", ")", "in", "config", ".", "items", "(", ")", ":", "if", "(", "not", "isinstance", "(", "key", ",", "six", ".", "string_types", ")", ")", ":", "raise", "ConfigurationError", "(", "u\"In file '{filename}', the {section} name {name} must be a quoted string, i.e. '{name}'.\"", ".", "format", "(", "filename", "=", "filename", ",", "section", "=", "section", ",", "name", "=", "key", ")", ")", "if", "(", "not", "isinstance", "(", "value", ",", "(", "dict", ",", "type", "(", "None", ")", ")", ")", ")", ":", "raise", "ConfigurationError", "(", "u\"In file '{filename}', {section} '{name}' must be a mapping not {type}.\"", ".", "format", "(", "filename", "=", "filename", ",", "section", "=", "section", ",", "name", "=", "key", ",", "type", "=", "anglicize_json_type", "(", "python_type_to_yaml_type", "(", "value", ")", ")", ")", ")" ]
validate the structure of a configuration section .
train
false
51,377
def nn(model, text, vectors, query, k=5): qf = encode(model, [query]) qf /= norm(qf) scores = numpy.dot(qf, vectors.T).flatten() sorted_args = numpy.argsort(scores)[::(-1)] sentences = [text[a] for a in sorted_args[:k]] print ('QUERY: ' + query) print 'NEAREST: ' for (i, s) in enumerate(sentences): print s, sorted_args[i]
[ "def", "nn", "(", "model", ",", "text", ",", "vectors", ",", "query", ",", "k", "=", "5", ")", ":", "qf", "=", "encode", "(", "model", ",", "[", "query", "]", ")", "qf", "/=", "norm", "(", "qf", ")", "scores", "=", "numpy", ".", "dot", "(", "qf", ",", "vectors", ".", "T", ")", ".", "flatten", "(", ")", "sorted_args", "=", "numpy", ".", "argsort", "(", "scores", ")", "[", ":", ":", "(", "-", "1", ")", "]", "sentences", "=", "[", "text", "[", "a", "]", "for", "a", "in", "sorted_args", "[", ":", "k", "]", "]", "print", "(", "'QUERY: '", "+", "query", ")", "print", "'NEAREST: '", "for", "(", "i", ",", "s", ")", "in", "enumerate", "(", "sentences", ")", ":", "print", "s", ",", "sorted_args", "[", "i", "]" ]
return the nearest neighbour sentences to query text: list of sentences vectors: the corresponding representations for text query: a string to search .
train
false
51,378
def isLDAPUrl(s): s_lower = s.lower() return (s_lower.startswith('ldap://') or s_lower.startswith('ldaps://') or s_lower.startswith('ldapi://'))
[ "def", "isLDAPUrl", "(", "s", ")", ":", "s_lower", "=", "s", ".", "lower", "(", ")", "return", "(", "s_lower", ".", "startswith", "(", "'ldap://'", ")", "or", "s_lower", ".", "startswith", "(", "'ldaps://'", ")", "or", "s_lower", ".", "startswith", "(", "'ldapi://'", ")", ")" ]
returns 1 if s is a ldap url .
train
false
51,380
def register_config_key(key, schema, required=False): _root_config_schema[u'properties'][key] = schema if required: _root_config_schema.setdefault(u'required', []).append(key) register_schema((u'/schema/config/%s' % key), schema)
[ "def", "register_config_key", "(", "key", ",", "schema", ",", "required", "=", "False", ")", ":", "_root_config_schema", "[", "u'properties'", "]", "[", "key", "]", "=", "schema", "if", "required", ":", "_root_config_schema", ".", "setdefault", "(", "u'required'", ",", "[", "]", ")", ".", "append", "(", "key", ")", "register_schema", "(", "(", "u'/schema/config/%s'", "%", "key", ")", ",", "schema", ")" ]
registers a valid root level key for the config .
train
false
51,381
def _describe_volume(volume): return {'id': volume.id, 'creation_time': _format_time(_get_volume_creation_time(volume)), 'provider': volume.driver.name, 'region': _get_volume_region(volume), 'extra': repr(volume.extra)}
[ "def", "_describe_volume", "(", "volume", ")", ":", "return", "{", "'id'", ":", "volume", ".", "id", ",", "'creation_time'", ":", "_format_time", "(", "_get_volume_creation_time", "(", "volume", ")", ")", ",", "'provider'", ":", "volume", ".", "driver", ".", "name", ",", "'region'", ":", "_get_volume_region", "(", "volume", ")", ",", "'extra'", ":", "repr", "(", "volume", ".", "extra", ")", "}" ]
create a dictionary giving lots of interesting details about a cloud volume .
train
false
51,383
def format_params(paramlist, otherlist=None): hdr = u'Parameters' delim = u'----------' paramlist.insert(0, delim) paramlist.insert(0, hdr) params = u'\n'.join(paramlist) otherparams = [] doc = u''.join(params) if otherlist: hdr = u'Others Parameters' delim = u'-----------------' otherlist.insert(0, delim) otherlist.insert(0, hdr) otherlist.insert(0, u'\n') otherparams = u'\n'.join(otherlist) doc = u''.join([doc, otherparams]) return doc
[ "def", "format_params", "(", "paramlist", ",", "otherlist", "=", "None", ")", ":", "hdr", "=", "u'Parameters'", "delim", "=", "u'----------'", "paramlist", ".", "insert", "(", "0", ",", "delim", ")", "paramlist", ".", "insert", "(", "0", ",", "hdr", ")", "params", "=", "u'\\n'", ".", "join", "(", "paramlist", ")", "otherparams", "=", "[", "]", "doc", "=", "u''", ".", "join", "(", "params", ")", "if", "otherlist", ":", "hdr", "=", "u'Others Parameters'", "delim", "=", "u'-----------------'", "otherlist", ".", "insert", "(", "0", ",", "delim", ")", "otherlist", ".", "insert", "(", "0", ",", "hdr", ")", "otherlist", ".", "insert", "(", "0", ",", "u'\\n'", ")", "otherparams", "=", "u'\\n'", ".", "join", "(", "otherlist", ")", "doc", "=", "u''", ".", "join", "(", "[", "doc", ",", "otherparams", "]", ")", "return", "doc" ]
format the parameters according to the nipy style conventions .
train
false
51,384
def WriteHeaders(headers, outfile, content_len=None): wrote_content_length = False for (header, value) in headers.iteritems(): if ((header.lower() == 'content-length') and (content_len is not None)): value = content_len wrote_content_length = True outfile.write(('%s: %s\r\n' % (header, value))) if ((not wrote_content_length) and content_len): outfile.write(('Content-Length: %s\r\n' % content_len))
[ "def", "WriteHeaders", "(", "headers", ",", "outfile", ",", "content_len", "=", "None", ")", ":", "wrote_content_length", "=", "False", "for", "(", "header", ",", "value", ")", "in", "headers", ".", "iteritems", "(", ")", ":", "if", "(", "(", "header", ".", "lower", "(", ")", "==", "'content-length'", ")", "and", "(", "content_len", "is", "not", "None", ")", ")", ":", "value", "=", "content_len", "wrote_content_length", "=", "True", "outfile", ".", "write", "(", "(", "'%s: %s\\r\\n'", "%", "(", "header", ",", "value", ")", ")", ")", "if", "(", "(", "not", "wrote_content_length", ")", "and", "content_len", ")", ":", "outfile", ".", "write", "(", "(", "'Content-Length: %s\\r\\n'", "%", "content_len", ")", ")" ]
write headers to the output file .
train
false
51,385
def _update_on_fields(prev_ele, new_ele): return dict(((prop_name, prop_val) for (prop_name, prop_val) in six.iteritems(new_ele) if ((new_ele.get(prop_name) != prev_ele.get(prop_name)) or (prop_name in _MATCH_KEYS))))
[ "def", "_update_on_fields", "(", "prev_ele", ",", "new_ele", ")", ":", "return", "dict", "(", "(", "(", "prop_name", ",", "prop_val", ")", "for", "(", "prop_name", ",", "prop_val", ")", "in", "six", ".", "iteritems", "(", "new_ele", ")", "if", "(", "(", "new_ele", ".", "get", "(", "prop_name", ")", "!=", "prev_ele", ".", "get", "(", "prop_name", ")", ")", "or", "(", "prop_name", "in", "_MATCH_KEYS", ")", ")", ")", ")" ]
return a dict with fields that differ between two dicts .
train
false
51,386
def get_vm_extra_config_spec(client_factory, extra_opts): config_spec = client_factory.create('ns0:VirtualMachineConfigSpec') extra_config = [] for (key, value) in extra_opts.items(): opt = client_factory.create('ns0:OptionValue') opt.key = key opt.value = value extra_config.append(opt) config_spec.extraConfig = extra_config return config_spec
[ "def", "get_vm_extra_config_spec", "(", "client_factory", ",", "extra_opts", ")", ":", "config_spec", "=", "client_factory", ".", "create", "(", "'ns0:VirtualMachineConfigSpec'", ")", "extra_config", "=", "[", "]", "for", "(", "key", ",", "value", ")", "in", "extra_opts", ".", "items", "(", ")", ":", "opt", "=", "client_factory", ".", "create", "(", "'ns0:OptionValue'", ")", "opt", ".", "key", "=", "key", "opt", ".", "value", "=", "value", "extra_config", ".", "append", "(", "opt", ")", "config_spec", ".", "extraConfig", "=", "extra_config", "return", "config_spec" ]
builds extra spec fields from a dictionary .
train
false
51,388
def delete_saml_provider(name, region=None, key=None, keyid=None, profile=None): conn = _get_conn(region=region, key=key, keyid=keyid, profile=profile) try: saml_provider_arn = get_saml_provider_arn(name, region=region, key=key, keyid=keyid, profile=profile) if (not saml_provider_arn): msg = 'SAML provider {0} not found.' log.info(msg.format(name)) return True conn.delete_saml_provider(saml_provider_arn) msg = 'Successfully deleted {0} SAML provider.' log.info(msg.format(name)) return True except boto.exception.BotoServerError as e: aws = __utils__['boto.get_error'](e) log.debug(aws) msg = 'Failed to delete {0} SAML provider.' log.error(msg.format(name)) return False
[ "def", "delete_saml_provider", "(", "name", ",", "region", "=", "None", ",", "key", "=", "None", ",", "keyid", "=", "None", ",", "profile", "=", "None", ")", ":", "conn", "=", "_get_conn", "(", "region", "=", "region", ",", "key", "=", "key", ",", "keyid", "=", "keyid", ",", "profile", "=", "profile", ")", "try", ":", "saml_provider_arn", "=", "get_saml_provider_arn", "(", "name", ",", "region", "=", "region", ",", "key", "=", "key", ",", "keyid", "=", "keyid", ",", "profile", "=", "profile", ")", "if", "(", "not", "saml_provider_arn", ")", ":", "msg", "=", "'SAML provider {0} not found.'", "log", ".", "info", "(", "msg", ".", "format", "(", "name", ")", ")", "return", "True", "conn", ".", "delete_saml_provider", "(", "saml_provider_arn", ")", "msg", "=", "'Successfully deleted {0} SAML provider.'", "log", ".", "info", "(", "msg", ".", "format", "(", "name", ")", ")", "return", "True", "except", "boto", ".", "exception", ".", "BotoServerError", "as", "e", ":", "aws", "=", "__utils__", "[", "'boto.get_error'", "]", "(", "e", ")", "log", ".", "debug", "(", "aws", ")", "msg", "=", "'Failed to delete {0} SAML provider.'", "log", ".", "error", "(", "msg", ".", "format", "(", "name", ")", ")", "return", "False" ]
delete saml provider cli example: .
train
true
51,389
def get_user_specific_rating_for_exploration(user_id, exploration_id): exp_user_data_model = user_models.ExplorationUserDataModel.get(user_id, exploration_id) return (exp_user_data_model.rating if exp_user_data_model else None)
[ "def", "get_user_specific_rating_for_exploration", "(", "user_id", ",", "exploration_id", ")", ":", "exp_user_data_model", "=", "user_models", ".", "ExplorationUserDataModel", ".", "get", "(", "user_id", ",", "exploration_id", ")", "return", "(", "exp_user_data_model", ".", "rating", "if", "exp_user_data_model", "else", "None", ")" ]
fetches a rating for the specified exploration from the specified user if one exists .
train
false
51,390
def datetime_to_utc_timestamp(timeval): if (timeval is not None): return (timegm(timeval.utctimetuple()) + (timeval.microsecond / 1000000))
[ "def", "datetime_to_utc_timestamp", "(", "timeval", ")", ":", "if", "(", "timeval", "is", "not", "None", ")", ":", "return", "(", "timegm", "(", "timeval", ".", "utctimetuple", "(", ")", ")", "+", "(", "timeval", ".", "microsecond", "/", "1000000", ")", ")" ]
converts a datetime instance to a timestamp .
train
false
51,393
def update_char_format(baseformat, color=None, background=None, weight=None, italic=None, underline=None, font=None): charformat = QTextCharFormat(baseformat) if (color is not None): charformat.setForeground(color) if (background is not None): charformat.setBackground(background) if (font is not None): charformat.setFont(font) else: font = update_font(baseformat.font(), weight, italic, underline) charformat.setFont(font) return charformat
[ "def", "update_char_format", "(", "baseformat", ",", "color", "=", "None", ",", "background", "=", "None", ",", "weight", "=", "None", ",", "italic", "=", "None", ",", "underline", "=", "None", ",", "font", "=", "None", ")", ":", "charformat", "=", "QTextCharFormat", "(", "baseformat", ")", "if", "(", "color", "is", "not", "None", ")", ":", "charformat", ".", "setForeground", "(", "color", ")", "if", "(", "background", "is", "not", "None", ")", ":", "charformat", ".", "setBackground", "(", "background", ")", "if", "(", "font", "is", "not", "None", ")", ":", "charformat", ".", "setFont", "(", "font", ")", "else", ":", "font", "=", "update_font", "(", "baseformat", ".", "font", "(", ")", ",", "weight", ",", "italic", ",", "underline", ")", "charformat", ".", "setFont", "(", "font", ")", "return", "charformat" ]
return a copy of baseformat :class:qtextcharformat with updated color .
train
false
51,396
def get_date_regex(timestring): prev = '' curr = '' regex = '' for s in range(0, len(timestring)): curr = timestring[s] if (curr == '%'): pass elif ((curr in settings.date_regex()) and (prev == '%')): regex += (('\\d{' + settings.date_regex()[curr]) + '}') elif (curr in ['.', '-']): regex += ('\\' + curr) else: regex += curr prev = curr logger.debug('regex = {0}'.format(regex)) return regex
[ "def", "get_date_regex", "(", "timestring", ")", ":", "prev", "=", "''", "curr", "=", "''", "regex", "=", "''", "for", "s", "in", "range", "(", "0", ",", "len", "(", "timestring", ")", ")", ":", "curr", "=", "timestring", "[", "s", "]", "if", "(", "curr", "==", "'%'", ")", ":", "pass", "elif", "(", "(", "curr", "in", "settings", ".", "date_regex", "(", ")", ")", "and", "(", "prev", "==", "'%'", ")", ")", ":", "regex", "+=", "(", "(", "'\\\\d{'", "+", "settings", ".", "date_regex", "(", ")", "[", "curr", "]", ")", "+", "'}'", ")", "elif", "(", "curr", "in", "[", "'.'", ",", "'-'", "]", ")", ":", "regex", "+=", "(", "'\\\\'", "+", "curr", ")", "else", ":", "regex", "+=", "curr", "prev", "=", "curr", "logger", ".", "debug", "(", "'regex = {0}'", ".", "format", "(", "regex", ")", ")", "return", "regex" ]
return a regex string based on a provided strftime timestring .
train
false
51,397
def load_parser(grammar_url, trace=0, parser=None, chart_class=None, beam_size=0, **load_args): grammar = load(grammar_url, **load_args) if (not isinstance(grammar, CFG)): raise ValueError('The grammar must be a CFG, or a subclass thereof.') if isinstance(grammar, PCFG): if (parser is None): parser = InsideChartParser return parser(grammar, trace=trace, beam_size=beam_size) elif isinstance(grammar, FeatureGrammar): if (parser is None): parser = FeatureChartParser if (chart_class is None): chart_class = FeatureChart return parser(grammar, trace=trace, chart_class=chart_class) else: if (parser is None): parser = ChartParser if (chart_class is None): chart_class = Chart return parser(grammar, trace=trace, chart_class=chart_class)
[ "def", "load_parser", "(", "grammar_url", ",", "trace", "=", "0", ",", "parser", "=", "None", ",", "chart_class", "=", "None", ",", "beam_size", "=", "0", ",", "**", "load_args", ")", ":", "grammar", "=", "load", "(", "grammar_url", ",", "**", "load_args", ")", "if", "(", "not", "isinstance", "(", "grammar", ",", "CFG", ")", ")", ":", "raise", "ValueError", "(", "'The grammar must be a CFG, or a subclass thereof.'", ")", "if", "isinstance", "(", "grammar", ",", "PCFG", ")", ":", "if", "(", "parser", "is", "None", ")", ":", "parser", "=", "InsideChartParser", "return", "parser", "(", "grammar", ",", "trace", "=", "trace", ",", "beam_size", "=", "beam_size", ")", "elif", "isinstance", "(", "grammar", ",", "FeatureGrammar", ")", ":", "if", "(", "parser", "is", "None", ")", ":", "parser", "=", "FeatureChartParser", "if", "(", "chart_class", "is", "None", ")", ":", "chart_class", "=", "FeatureChart", "return", "parser", "(", "grammar", ",", "trace", "=", "trace", ",", "chart_class", "=", "chart_class", ")", "else", ":", "if", "(", "parser", "is", "None", ")", ":", "parser", "=", "ChartParser", "if", "(", "chart_class", "is", "None", ")", ":", "chart_class", "=", "Chart", "return", "parser", "(", "grammar", ",", "trace", "=", "trace", ",", "chart_class", "=", "chart_class", ")" ]
load a grammar from a file .
train
false
51,398
def dev_from_index(if_index): return IFACES.dev_from_index(if_index)
[ "def", "dev_from_index", "(", "if_index", ")", ":", "return", "IFACES", ".", "dev_from_index", "(", "if_index", ")" ]
return windows adapter name for given windows interface index .
train
false
51,399
def convert_TextProperty(model, prop, kwargs): return f.TextAreaField(**kwargs)
[ "def", "convert_TextProperty", "(", "model", ",", "prop", ",", "kwargs", ")", ":", "return", "f", ".", "TextAreaField", "(", "**", "kwargs", ")" ]
returns a form field for a db .
train
false
51,402
def collect_command(host, command, dest_path): logging.info("Collecting '%s' ...", command) devnull = open('/dev/null', 'w') try: result = host.run(command, stdout_tee=devnull).stdout utils.open_write_close(dest_path, result) except Exception as e: logging.warning("Collection of '%s' failed:\n%s", command, e) finally: devnull.close()
[ "def", "collect_command", "(", "host", ",", "command", ",", "dest_path", ")", ":", "logging", ".", "info", "(", "\"Collecting '%s' ...\"", ",", "command", ")", "devnull", "=", "open", "(", "'/dev/null'", ",", "'w'", ")", "try", ":", "result", "=", "host", ".", "run", "(", "command", ",", "stdout_tee", "=", "devnull", ")", ".", "stdout", "utils", ".", "open_write_close", "(", "dest_path", ",", "result", ")", "except", "Exception", "as", "e", ":", "logging", ".", "warning", "(", "\"Collection of '%s' failed:\\n%s\"", ",", "command", ",", "e", ")", "finally", ":", "devnull", ".", "close", "(", ")" ]
collects the result of a command on the remote machine .
train
false
51,403
def ValidatePropertyNothing(name, value): pass
[ "def", "ValidatePropertyNothing", "(", "name", ",", "value", ")", ":", "pass" ]
no-op validation function .
train
false
51,404
def set_process_title(progname, info=None): proctitle = ('[%s]' % progname) proctitle = (('%s %s' % (proctitle, info)) if info else proctitle) if _setproctitle: _setproctitle.setproctitle(proctitle) return proctitle
[ "def", "set_process_title", "(", "progname", ",", "info", "=", "None", ")", ":", "proctitle", "=", "(", "'[%s]'", "%", "progname", ")", "proctitle", "=", "(", "(", "'%s %s'", "%", "(", "proctitle", ",", "info", ")", ")", "if", "info", "else", "proctitle", ")", "if", "_setproctitle", ":", "_setproctitle", ".", "setproctitle", "(", "proctitle", ")", "return", "proctitle" ]
set the ps name for the currently running process .
train
false
51,405
def strip_ansi_codes(s): return re.sub(u'\x1b\\[([0-9]+)(;[0-9]+)*m', u'', s)
[ "def", "strip_ansi_codes", "(", "s", ")", ":", "return", "re", ".", "sub", "(", "u'\\x1b\\\\[([0-9]+)(;[0-9]+)*m'", ",", "u''", ",", "s", ")" ]
remove ansi color codes from the string .
train
false
51,407
@contextlib.contextmanager def fpopen(*args, **kwargs): uid = kwargs.pop('uid', (-1)) gid = kwargs.pop('gid', (-1)) mode = kwargs.pop('mode', None) with fopen(*args, **kwargs) as fhandle: path = args[0] d_stat = os.stat(path) if hasattr(os, 'chown'): if (((d_stat.st_uid != uid) or (d_stat.st_gid != gid)) and [i for i in (uid, gid) if (i != (-1))]): os.chown(path, uid, gid) if (mode is not None): mode_part = S_IMODE(d_stat.st_mode) if (mode_part != mode): os.chmod(path, ((d_stat.st_mode ^ mode_part) | mode)) (yield fhandle)
[ "@", "contextlib", ".", "contextmanager", "def", "fpopen", "(", "*", "args", ",", "**", "kwargs", ")", ":", "uid", "=", "kwargs", ".", "pop", "(", "'uid'", ",", "(", "-", "1", ")", ")", "gid", "=", "kwargs", ".", "pop", "(", "'gid'", ",", "(", "-", "1", ")", ")", "mode", "=", "kwargs", ".", "pop", "(", "'mode'", ",", "None", ")", "with", "fopen", "(", "*", "args", ",", "**", "kwargs", ")", "as", "fhandle", ":", "path", "=", "args", "[", "0", "]", "d_stat", "=", "os", ".", "stat", "(", "path", ")", "if", "hasattr", "(", "os", ",", "'chown'", ")", ":", "if", "(", "(", "(", "d_stat", ".", "st_uid", "!=", "uid", ")", "or", "(", "d_stat", ".", "st_gid", "!=", "gid", ")", ")", "and", "[", "i", "for", "i", "in", "(", "uid", ",", "gid", ")", "if", "(", "i", "!=", "(", "-", "1", ")", ")", "]", ")", ":", "os", ".", "chown", "(", "path", ",", "uid", ",", "gid", ")", "if", "(", "mode", "is", "not", "None", ")", ":", "mode_part", "=", "S_IMODE", "(", "d_stat", ".", "st_mode", ")", "if", "(", "mode_part", "!=", "mode", ")", ":", "os", ".", "chmod", "(", "path", ",", "(", "(", "d_stat", ".", "st_mode", "^", "mode_part", ")", "|", "mode", ")", ")", "(", "yield", "fhandle", ")" ]
shortcut for fopen with extra uid .
train
true
51,408
def call_tesseract(input_filename, output_filename, bool_digits=False): if bool_digits: args = (((((tesseract_exe_name + ' ') + input_filename) + ' ') + output_filename) + ' -l test_digits -psm 7 nobatch') else: args = (((((tesseract_exe_name + ' ') + input_filename) + ' ') + output_filename) + ' -l eng -psm 7 nobatch eng_characters') proc = subprocess.Popen(args, shell=True) retcode = proc.wait() if (retcode != 0): errors.check_for_errors()
[ "def", "call_tesseract", "(", "input_filename", ",", "output_filename", ",", "bool_digits", "=", "False", ")", ":", "if", "bool_digits", ":", "args", "=", "(", "(", "(", "(", "(", "tesseract_exe_name", "+", "' '", ")", "+", "input_filename", ")", "+", "' '", ")", "+", "output_filename", ")", "+", "' -l test_digits -psm 7 nobatch'", ")", "else", ":", "args", "=", "(", "(", "(", "(", "(", "tesseract_exe_name", "+", "' '", ")", "+", "input_filename", ")", "+", "' '", ")", "+", "output_filename", ")", "+", "' -l eng -psm 7 nobatch eng_characters'", ")", "proc", "=", "subprocess", ".", "Popen", "(", "args", ",", "shell", "=", "True", ")", "retcode", "=", "proc", ".", "wait", "(", ")", "if", "(", "retcode", "!=", "0", ")", ":", "errors", ".", "check_for_errors", "(", ")" ]
calls external tesseract .
train
false
51,409
def replace_pyzzer_entry_point_shebang(all_data, placeholder, new_prefix): launcher = shebang = None pos = all_data.rfind('PK\x05\x06') if (pos >= 0): end_cdr = all_data[(pos + 12):(pos + 20)] (cdr_size, cdr_offset) = struct.unpack(u'<LL', end_cdr) arc_pos = ((pos - cdr_size) - cdr_offset) data = all_data[arc_pos:] if (arc_pos > 0): pos = all_data.rfind('#!', 0, arc_pos) if (pos >= 0): shebang = all_data[pos:arc_pos] if (pos > 0): launcher = all_data[:pos] if (data and shebang and launcher): if hasattr(placeholder, u'encode'): placeholder = placeholder.encode(u'utf-8') if hasattr(new_prefix, u'encode'): new_prefix = new_prefix.encode(u'utf-8') shebang = shebang.replace(placeholder, new_prefix) all_data = ''.join([launcher, shebang, data]) return all_data
[ "def", "replace_pyzzer_entry_point_shebang", "(", "all_data", ",", "placeholder", ",", "new_prefix", ")", ":", "launcher", "=", "shebang", "=", "None", "pos", "=", "all_data", ".", "rfind", "(", "'PK\\x05\\x06'", ")", "if", "(", "pos", ">=", "0", ")", ":", "end_cdr", "=", "all_data", "[", "(", "pos", "+", "12", ")", ":", "(", "pos", "+", "20", ")", "]", "(", "cdr_size", ",", "cdr_offset", ")", "=", "struct", ".", "unpack", "(", "u'<LL'", ",", "end_cdr", ")", "arc_pos", "=", "(", "(", "pos", "-", "cdr_size", ")", "-", "cdr_offset", ")", "data", "=", "all_data", "[", "arc_pos", ":", "]", "if", "(", "arc_pos", ">", "0", ")", ":", "pos", "=", "all_data", ".", "rfind", "(", "'#!'", ",", "0", ",", "arc_pos", ")", "if", "(", "pos", ">=", "0", ")", ":", "shebang", "=", "all_data", "[", "pos", ":", "arc_pos", "]", "if", "(", "pos", ">", "0", ")", ":", "launcher", "=", "all_data", "[", ":", "pos", "]", "if", "(", "data", "and", "shebang", "and", "launcher", ")", ":", "if", "hasattr", "(", "placeholder", ",", "u'encode'", ")", ":", "placeholder", "=", "placeholder", ".", "encode", "(", "u'utf-8'", ")", "if", "hasattr", "(", "new_prefix", ",", "u'encode'", ")", ":", "new_prefix", "=", "new_prefix", ".", "encode", "(", "u'utf-8'", ")", "shebang", "=", "shebang", ".", "replace", "(", "placeholder", ",", "new_prefix", ")", "all_data", "=", "''", ".", "join", "(", "[", "launcher", ",", "shebang", ",", "data", "]", ")", "return", "all_data" ]
code adapted from pyzzer .
train
false
51,410
def _parse_timespan(timespan): if (timespan in ('', 'NOT_IMPLEMENTED', None)): return None else: return sum((((60 ** x[0]) * int(x[1])) for x in enumerate(reversed(timespan.split(':')))))
[ "def", "_parse_timespan", "(", "timespan", ")", ":", "if", "(", "timespan", "in", "(", "''", ",", "'NOT_IMPLEMENTED'", ",", "None", ")", ")", ":", "return", "None", "else", ":", "return", "sum", "(", "(", "(", "(", "60", "**", "x", "[", "0", "]", ")", "*", "int", "(", "x", "[", "1", "]", ")", ")", "for", "x", "in", "enumerate", "(", "reversed", "(", "timespan", ".", "split", "(", "':'", ")", ")", ")", ")", ")" ]
parse a time-span into number of seconds .
train
false
51,412
def has_database_privileges(cursor, user, db, privs): cur_privs = get_database_privileges(cursor, user, db) have_currently = cur_privs.intersection(privs) other_current = cur_privs.difference(privs) desired = privs.difference(cur_privs) return (have_currently, other_current, desired)
[ "def", "has_database_privileges", "(", "cursor", ",", "user", ",", "db", ",", "privs", ")", ":", "cur_privs", "=", "get_database_privileges", "(", "cursor", ",", "user", ",", "db", ")", "have_currently", "=", "cur_privs", ".", "intersection", "(", "privs", ")", "other_current", "=", "cur_privs", ".", "difference", "(", "privs", ")", "desired", "=", "privs", ".", "difference", "(", "cur_privs", ")", "return", "(", "have_currently", ",", "other_current", ",", "desired", ")" ]
return the difference between the privileges that a user already has and the privileges that they desire to have .
train
false
51,413
@status('Docs modified', modal=True) def docs_modified(file_paths): for path in file_paths: if path.startswith('Doc'): return True return False
[ "@", "status", "(", "'Docs modified'", ",", "modal", "=", "True", ")", "def", "docs_modified", "(", "file_paths", ")", ":", "for", "path", "in", "file_paths", ":", "if", "path", ".", "startswith", "(", "'Doc'", ")", ":", "return", "True", "return", "False" ]
report if any files in the docs directory .
train
false
51,415
def extract_num(buf, start, length): val = 0 for i in range(start, (start + length)): val <<= 8 val += ord(buf[i]) return val
[ "def", "extract_num", "(", "buf", ",", "start", ",", "length", ")", ":", "val", "=", "0", "for", "i", "in", "range", "(", "start", ",", "(", "start", "+", "length", ")", ")", ":", "val", "<<=", "8", "val", "+=", "ord", "(", "buf", "[", "i", "]", ")", "return", "val" ]
extracts a number from a raw byte string .
train
false
51,416
def _save_forum_role(course_key, name): (role, created) = Role.objects.get_or_create(name=name, course_id=course_key) if (created is False): role.course_id = course_key role.save() return role
[ "def", "_save_forum_role", "(", "course_key", ",", "name", ")", ":", "(", "role", ",", "created", ")", "=", "Role", ".", "objects", ".", "get_or_create", "(", "name", "=", "name", ",", "course_id", "=", "course_key", ")", "if", "(", "created", "is", "False", ")", ":", "role", ".", "course_id", "=", "course_key", "role", ".", "save", "(", ")", "return", "role" ]
save and update course_key for all roles which are already created to keep course_id same as actual passed course key .
train
false
51,418
def _formatSchema(data, incoming): if incoming: param = '<json' else: param = '>json' if (data['type'] == 'array'): param += 'arr' for (prop, attr) in sorted(data[u'properties'].iteritems()): if attr['required']: required = '*(required)* ' else: required = '' if isinstance(attr['type'], list): types = '|'.join(attr['type']) else: types = attr['type'] (yield (':%s %s %s: %s%s' % (param, types, prop, required, attr['title']))) (yield '') for line in attr['description']: (yield (' ' + line))
[ "def", "_formatSchema", "(", "data", ",", "incoming", ")", ":", "if", "incoming", ":", "param", "=", "'<json'", "else", ":", "param", "=", "'>json'", "if", "(", "data", "[", "'type'", "]", "==", "'array'", ")", ":", "param", "+=", "'arr'", "for", "(", "prop", ",", "attr", ")", "in", "sorted", "(", "data", "[", "u'properties'", "]", ".", "iteritems", "(", ")", ")", ":", "if", "attr", "[", "'required'", "]", ":", "required", "=", "'*(required)* '", "else", ":", "required", "=", "''", "if", "isinstance", "(", "attr", "[", "'type'", "]", ",", "list", ")", ":", "types", "=", "'|'", ".", "join", "(", "attr", "[", "'type'", "]", ")", "else", ":", "types", "=", "attr", "[", "'type'", "]", "(", "yield", "(", "':%s %s %s: %s%s'", "%", "(", "param", ",", "types", ",", "prop", ",", "required", ",", "attr", "[", "'title'", "]", ")", ")", ")", "(", "yield", "''", ")", "for", "line", "in", "attr", "[", "'description'", "]", ":", "(", "yield", "(", "' '", "+", "line", ")", ")" ]
generate the rst associated to a json schema .
train
false
51,419
def copula_bv_max(u, v): return np.maximum(((u + v) - 1), 0)
[ "def", "copula_bv_max", "(", "u", ",", "v", ")", ":", "return", "np", ".", "maximum", "(", "(", "(", "u", "+", "v", ")", "-", "1", ")", ",", "0", ")" ]
countermonotonic bivariate copula .
train
false
51,420
def goto(dNorth, dEast, gotoFunction=vehicle.simple_goto): currentLocation = vehicle.location.global_relative_frame targetLocation = get_location_metres(currentLocation, dNorth, dEast) targetDistance = get_distance_metres(currentLocation, targetLocation) gotoFunction(targetLocation) while (vehicle.mode.name == 'GUIDED'): remainingDistance = get_distance_metres(vehicle.location.global_relative_frame, targetLocation) print 'Distance to target: ', remainingDistance if (remainingDistance <= (targetDistance * 0.01)): print 'Reached target' break time.sleep(2)
[ "def", "goto", "(", "dNorth", ",", "dEast", ",", "gotoFunction", "=", "vehicle", ".", "simple_goto", ")", ":", "currentLocation", "=", "vehicle", ".", "location", ".", "global_relative_frame", "targetLocation", "=", "get_location_metres", "(", "currentLocation", ",", "dNorth", ",", "dEast", ")", "targetDistance", "=", "get_distance_metres", "(", "currentLocation", ",", "targetLocation", ")", "gotoFunction", "(", "targetLocation", ")", "while", "(", "vehicle", ".", "mode", ".", "name", "==", "'GUIDED'", ")", ":", "remainingDistance", "=", "get_distance_metres", "(", "vehicle", ".", "location", ".", "global_relative_frame", ",", "targetLocation", ")", "print", "'Distance to target: '", ",", "remainingDistance", "if", "(", "remainingDistance", "<=", "(", "targetDistance", "*", "0.01", ")", ")", ":", "print", "'Reached target'", "break", "time", ".", "sleep", "(", "2", ")" ]
moves the vehicle to a position dnorth metres north and deast metres east of the current position .
train
true
51,421
def load_pixbuf(fname, size=0): image = gtk.Image() image.set_from_file(os.path.join(_pixmap_path, fname)) image = image.get_pixbuf() if size: aspect = (float(image.get_height()) / image.get_width()) image = image.scale_simple(size, int((aspect * size)), 2) return image
[ "def", "load_pixbuf", "(", "fname", ",", "size", "=", "0", ")", ":", "image", "=", "gtk", ".", "Image", "(", ")", "image", ".", "set_from_file", "(", "os", ".", "path", ".", "join", "(", "_pixmap_path", ",", "fname", ")", ")", "image", "=", "image", ".", "get_pixbuf", "(", ")", "if", "size", ":", "aspect", "=", "(", "float", "(", "image", ".", "get_height", "(", ")", ")", "/", "image", ".", "get_width", "(", ")", ")", "image", "=", "image", ".", "scale_simple", "(", "size", ",", "int", "(", "(", "aspect", "*", "size", ")", ")", ",", "2", ")", "return", "image" ]
load an image from a file as a pixbuf .
train
false
51,423
def read_simple_binding(jboss_config, binding_name, profile=None): log.debug('======================== MODULE FUNCTION: jboss7.read_simple_binding, %s', binding_name) return __read_simple_binding(jboss_config, binding_name, profile=profile)
[ "def", "read_simple_binding", "(", "jboss_config", ",", "binding_name", ",", "profile", "=", "None", ")", ":", "log", ".", "debug", "(", "'======================== MODULE FUNCTION: jboss7.read_simple_binding, %s'", ",", "binding_name", ")", "return", "__read_simple_binding", "(", "jboss_config", ",", "binding_name", ",", "profile", "=", "profile", ")" ]
read jndi binding in the running jboss instance jboss_config configuration dictionary with properties specified above .
train
true
51,424
def ttest_1samp_no_p(X, sigma=0, method='relative'): if (method not in ['absolute', 'relative']): raise ValueError(('method must be "absolute" or "relative", not %s' % method)) var = np.var(X, axis=0, ddof=1) if (sigma > 0): limit = ((sigma * np.max(var)) if (method == 'relative') else sigma) var += limit return (np.mean(X, axis=0) / np.sqrt((var / X.shape[0])))
[ "def", "ttest_1samp_no_p", "(", "X", ",", "sigma", "=", "0", ",", "method", "=", "'relative'", ")", ":", "if", "(", "method", "not", "in", "[", "'absolute'", ",", "'relative'", "]", ")", ":", "raise", "ValueError", "(", "(", "'method must be \"absolute\" or \"relative\", not %s'", "%", "method", ")", ")", "var", "=", "np", ".", "var", "(", "X", ",", "axis", "=", "0", ",", "ddof", "=", "1", ")", "if", "(", "sigma", ">", "0", ")", ":", "limit", "=", "(", "(", "sigma", "*", "np", ".", "max", "(", "var", ")", ")", "if", "(", "method", "==", "'relative'", ")", "else", "sigma", ")", "var", "+=", "limit", "return", "(", "np", ".", "mean", "(", "X", ",", "axis", "=", "0", ")", "/", "np", ".", "sqrt", "(", "(", "var", "/", "X", ".", "shape", "[", "0", "]", ")", ")", ")" ]
perform t-test with variance adjustment and no p-value calculation .
train
false
51,426
def cert_from_instance(instance): if instance.signature: if instance.signature.key_info: return cert_from_key_info(instance.signature.key_info, ignore_age=True) return []
[ "def", "cert_from_instance", "(", "instance", ")", ":", "if", "instance", ".", "signature", ":", "if", "instance", ".", "signature", ".", "key_info", ":", "return", "cert_from_key_info", "(", "instance", ".", "signature", ".", "key_info", ",", "ignore_age", "=", "True", ")", "return", "[", "]" ]
find certificates that are part of an instance .
train
true
51,428
def _makep(array, descr_output, format, nrows=None): _offset = 0 if (not nrows): nrows = len(array) data_output = _VLF(([None] * nrows), dtype=format.dtype) if (format.dtype == 'a'): _nbytes = 1 else: _nbytes = np.array([], dtype=format.dtype).itemsize for idx in range(nrows): if (idx < len(array)): rowval = array[idx] elif (format.dtype == 'a'): rowval = (' ' * data_output.max) else: rowval = ([0] * data_output.max) if (format.dtype == 'a'): data_output[idx] = chararray.array(encode_ascii(rowval), itemsize=1) else: data_output[idx] = np.array(rowval, dtype=format.dtype) descr_output[(idx, 0)] = len(data_output[idx]) descr_output[(idx, 1)] = _offset _offset += (len(data_output[idx]) * _nbytes) return data_output
[ "def", "_makep", "(", "array", ",", "descr_output", ",", "format", ",", "nrows", "=", "None", ")", ":", "_offset", "=", "0", "if", "(", "not", "nrows", ")", ":", "nrows", "=", "len", "(", "array", ")", "data_output", "=", "_VLF", "(", "(", "[", "None", "]", "*", "nrows", ")", ",", "dtype", "=", "format", ".", "dtype", ")", "if", "(", "format", ".", "dtype", "==", "'a'", ")", ":", "_nbytes", "=", "1", "else", ":", "_nbytes", "=", "np", ".", "array", "(", "[", "]", ",", "dtype", "=", "format", ".", "dtype", ")", ".", "itemsize", "for", "idx", "in", "range", "(", "nrows", ")", ":", "if", "(", "idx", "<", "len", "(", "array", ")", ")", ":", "rowval", "=", "array", "[", "idx", "]", "elif", "(", "format", ".", "dtype", "==", "'a'", ")", ":", "rowval", "=", "(", "' '", "*", "data_output", ".", "max", ")", "else", ":", "rowval", "=", "(", "[", "0", "]", "*", "data_output", ".", "max", ")", "if", "(", "format", ".", "dtype", "==", "'a'", ")", ":", "data_output", "[", "idx", "]", "=", "chararray", ".", "array", "(", "encode_ascii", "(", "rowval", ")", ",", "itemsize", "=", "1", ")", "else", ":", "data_output", "[", "idx", "]", "=", "np", ".", "array", "(", "rowval", ",", "dtype", "=", "format", ".", "dtype", ")", "descr_output", "[", "(", "idx", ",", "0", ")", "]", "=", "len", "(", "data_output", "[", "idx", "]", ")", "descr_output", "[", "(", "idx", ",", "1", ")", "]", "=", "_offset", "_offset", "+=", "(", "len", "(", "data_output", "[", "idx", "]", ")", "*", "_nbytes", ")", "return", "data_output" ]
construct the p format column array .
train
false
51,429
def create_from_options(options, name='unknown'): raise NotImplementedError
[ "def", "create_from_options", "(", "options", ",", "name", "=", "'unknown'", ")", ":", "raise", "NotImplementedError" ]
factory using an options dictionary .
train
false
51,430
def is_started(): return (status() == 'running')
[ "def", "is_started", "(", ")", ":", "return", "(", "status", "(", ")", "==", "'running'", ")" ]
check if the firewall is started .
train
false
51,432
def delslice(model, start, end): if isinstance(model, PyListModel): del model[start:end] elif isinstance(model, QAbstractItemModel): model.removeRows(start, (end - start)) else: raise TypeError(type(model))
[ "def", "delslice", "(", "model", ",", "start", ",", "end", ")", ":", "if", "isinstance", "(", "model", ",", "PyListModel", ")", ":", "del", "model", "[", "start", ":", "end", "]", "elif", "isinstance", "(", "model", ",", "QAbstractItemModel", ")", ":", "model", ".", "removeRows", "(", "start", ",", "(", "end", "-", "start", ")", ")", "else", ":", "raise", "TypeError", "(", "type", "(", "model", ")", ")" ]
delete the start .
train
false
51,435
def _SynchronizeTxn(function): def sync(txn, *args, **kwargs): txn._lock.acquire() try: Check((txn._state is LiveTxn.ACTIVE), 'transaction closed') return function(txn, *args, **kwargs) finally: txn._lock.release() return sync
[ "def", "_SynchronizeTxn", "(", "function", ")", ":", "def", "sync", "(", "txn", ",", "*", "args", ",", "**", "kwargs", ")", ":", "txn", ".", "_lock", ".", "acquire", "(", ")", "try", ":", "Check", "(", "(", "txn", ".", "_state", "is", "LiveTxn", ".", "ACTIVE", ")", ",", "'transaction closed'", ")", "return", "function", "(", "txn", ",", "*", "args", ",", "**", "kwargs", ")", "finally", ":", "txn", ".", "_lock", ".", "release", "(", ")", "return", "sync" ]
a decorator that locks a transaction during the function call .
train
false