id_within_dataset
int64
1
55.5k
snippet
stringlengths
19
14.2k
tokens
listlengths
6
1.63k
nl
stringlengths
6
352
split_within_dataset
stringclasses
1 value
is_duplicated
bool
2 classes
50,485
def isdevice(dev): try: os.stat(dev) return str(dev) except OSError: raise vol.Invalid('No device found!')
[ "def", "isdevice", "(", "dev", ")", ":", "try", ":", "os", ".", "stat", "(", "dev", ")", "return", "str", "(", "dev", ")", "except", "OSError", ":", "raise", "vol", ".", "Invalid", "(", "'No device found!'", ")" ]
check if dev a real device .
train
false
50,486
def get_global(key): global _global_data if (_global_data is None): dirname = os.path.join(os.path.dirname(__file__)) filename = os.path.join(dirname, 'global.dat') if (not os.path.isfile(filename)): _raise_no_data_error() with open(filename, 'rb') as fileobj: _global_data = pickle.load(fileobj) return _global_data.get(key, {})
[ "def", "get_global", "(", "key", ")", ":", "global", "_global_data", "if", "(", "_global_data", "is", "None", ")", ":", "dirname", "=", "os", ".", "path", ".", "join", "(", "os", ".", "path", ".", "dirname", "(", "__file__", ")", ")", "filename", "=", "os", ".", "path", ".", "join", "(", "dirname", ",", "'global.dat'", ")", "if", "(", "not", "os", ".", "path", ".", "isfile", "(", "filename", ")", ")", ":", "_raise_no_data_error", "(", ")", "with", "open", "(", "filename", ",", "'rb'", ")", "as", "fileobj", ":", "_global_data", "=", "pickle", ".", "load", "(", "fileobj", ")", "return", "_global_data", ".", "get", "(", "key", ",", "{", "}", ")" ]
return the dictionary for the given key in the global data .
train
false
50,487
def csrf_view_exempt(view_func): warnings.warn('csrf_view_exempt is deprecated. Use csrf_exempt instead.', PendingDeprecationWarning) return csrf_exempt(view_func)
[ "def", "csrf_view_exempt", "(", "view_func", ")", ":", "warnings", ".", "warn", "(", "'csrf_view_exempt is deprecated. Use csrf_exempt instead.'", ",", "PendingDeprecationWarning", ")", "return", "csrf_exempt", "(", "view_func", ")" ]
marks a view function as being exempt from csrf view protection .
train
false
50,489
def get_member_names(group): return [r.name for r in get_members(group)]
[ "def", "get_member_names", "(", "group", ")", ":", "return", "[", "r", ".", "name", "for", "r", "in", "get_members", "(", "group", ")", "]" ]
get a list of resource names of the resources in the specified group .
train
false
50,493
def systemInformationType6(): a = L2PseudoLength(l2pLength=11) b = TpPd(pd=6) c = MessageType(mesType=30) d = CellIdentity() e = LocalAreaId() f = CellOptionsBCCH() g = NccPermitted() h = Si6RestOctets() packet = ((((((a / b) / c) / d) / e) / f) / g) return packet
[ "def", "systemInformationType6", "(", ")", ":", "a", "=", "L2PseudoLength", "(", "l2pLength", "=", "11", ")", "b", "=", "TpPd", "(", "pd", "=", "6", ")", "c", "=", "MessageType", "(", "mesType", "=", "30", ")", "d", "=", "CellIdentity", "(", ")", "e", "=", "LocalAreaId", "(", ")", "f", "=", "CellOptionsBCCH", "(", ")", "g", "=", "NccPermitted", "(", ")", "h", "=", "Si6RestOctets", "(", ")", "packet", "=", "(", "(", "(", "(", "(", "(", "a", "/", "b", ")", "/", "c", ")", "/", "d", ")", "/", "e", ")", "/", "f", ")", "/", "g", ")", "return", "packet" ]
system information type 6 section 9 .
train
true
50,494
def test_clone_should_raise_if_vcs_not_installed(mocker, clone_dir): mocker.patch('cookiecutter.vcs.is_vcs_installed', autospec=True, return_value=False) repo_url = 'https://github.com/pytest-dev/cookiecutter-pytest-plugin.git' with pytest.raises(exceptions.VCSNotInstalled): vcs.clone(repo_url, clone_to_dir=clone_dir)
[ "def", "test_clone_should_raise_if_vcs_not_installed", "(", "mocker", ",", "clone_dir", ")", ":", "mocker", ".", "patch", "(", "'cookiecutter.vcs.is_vcs_installed'", ",", "autospec", "=", "True", ",", "return_value", "=", "False", ")", "repo_url", "=", "'https://github.com/pytest-dev/cookiecutter-pytest-plugin.git'", "with", "pytest", ".", "raises", "(", "exceptions", ".", "VCSNotInstalled", ")", ":", "vcs", ".", "clone", "(", "repo_url", ",", "clone_to_dir", "=", "clone_dir", ")" ]
in clone() .
train
false
50,496
def change_lists_to_sets(iterable): if isinstance(iterable, dict): for key in iterable: if isinstance(iterable[key], (list, tuple)): try: iterable[key] = set(iterable[key]) except TypeError: pass elif getattr(iterable[key], '__iter__', False): change_lists_to_sets(iterable[key]) elif isinstance(iterable, (list, tuple)): for item in iterable: if isinstance(item, (list, tuple)): iterable.pop(item) iterable.append(set(item)) elif getattr(item, '__iter__', False): change_lists_to_sets(item) else: raise NotImplementedError
[ "def", "change_lists_to_sets", "(", "iterable", ")", ":", "if", "isinstance", "(", "iterable", ",", "dict", ")", ":", "for", "key", "in", "iterable", ":", "if", "isinstance", "(", "iterable", "[", "key", "]", ",", "(", "list", ",", "tuple", ")", ")", ":", "try", ":", "iterable", "[", "key", "]", "=", "set", "(", "iterable", "[", "key", "]", ")", "except", "TypeError", ":", "pass", "elif", "getattr", "(", "iterable", "[", "key", "]", ",", "'__iter__'", ",", "False", ")", ":", "change_lists_to_sets", "(", "iterable", "[", "key", "]", ")", "elif", "isinstance", "(", "iterable", ",", "(", "list", ",", "tuple", ")", ")", ":", "for", "item", "in", "iterable", ":", "if", "isinstance", "(", "item", ",", "(", "list", ",", "tuple", ")", ")", ":", "iterable", ".", "pop", "(", "item", ")", "iterable", ".", "append", "(", "set", "(", "item", ")", ")", "elif", "getattr", "(", "item", ",", "'__iter__'", ",", "False", ")", ":", "change_lists_to_sets", "(", "item", ")", "else", ":", "raise", "NotImplementedError" ]
convert any lists or tuples in iterable into sets .
train
false
50,497
def eliminate_implications(expr): return to_nnf(expr)
[ "def", "eliminate_implications", "(", "expr", ")", ":", "return", "to_nnf", "(", "expr", ")" ]
change >> .
train
false
50,498
def wheel(): client = salt.wheel.Wheel(__opts__) ret = client.get_docs() return ret
[ "def", "wheel", "(", ")", ":", "client", "=", "salt", ".", "wheel", ".", "Wheel", "(", "__opts__", ")", "ret", "=", "client", ".", "get_docs", "(", ")", "return", "ret" ]
return all inline documentation for wheel modules cli example: .
train
true
50,500
def collect_members(module_to_name): members = {} for (module, module_name) in module_to_name.items(): all_names = getattr(module, '__all__', None) for (name, member) in inspect.getmembers(module): if ((inspect.isfunction(member) or inspect.isclass(member)) and (not _always_drop_symbol_re.match(name)) and ((all_names is None) or (name in all_names))): fullname = ('%s.%s' % (module_name, name)) if (name in members): (other_fullname, other_member) = members[name] if (member is not other_member): raise RuntimeError(('Short name collision between %s and %s' % (fullname, other_fullname))) if (len(fullname) == len(other_fullname)): raise RuntimeError(("Can't decide whether to use %s or %s for %s: both full names have length %d" % (fullname, other_fullname, name, len(fullname)))) if (len(fullname) > len(other_fullname)): continue members[name] = (fullname, member) return members
[ "def", "collect_members", "(", "module_to_name", ")", ":", "members", "=", "{", "}", "for", "(", "module", ",", "module_name", ")", "in", "module_to_name", ".", "items", "(", ")", ":", "all_names", "=", "getattr", "(", "module", ",", "'__all__'", ",", "None", ")", "for", "(", "name", ",", "member", ")", "in", "inspect", ".", "getmembers", "(", "module", ")", ":", "if", "(", "(", "inspect", ".", "isfunction", "(", "member", ")", "or", "inspect", ".", "isclass", "(", "member", ")", ")", "and", "(", "not", "_always_drop_symbol_re", ".", "match", "(", "name", ")", ")", "and", "(", "(", "all_names", "is", "None", ")", "or", "(", "name", "in", "all_names", ")", ")", ")", ":", "fullname", "=", "(", "'%s.%s'", "%", "(", "module_name", ",", "name", ")", ")", "if", "(", "name", "in", "members", ")", ":", "(", "other_fullname", ",", "other_member", ")", "=", "members", "[", "name", "]", "if", "(", "member", "is", "not", "other_member", ")", ":", "raise", "RuntimeError", "(", "(", "'Short name collision between %s and %s'", "%", "(", "fullname", ",", "other_fullname", ")", ")", ")", "if", "(", "len", "(", "fullname", ")", "==", "len", "(", "other_fullname", ")", ")", ":", "raise", "RuntimeError", "(", "(", "\"Can't decide whether to use %s or %s for %s: both full names have length %d\"", "%", "(", "fullname", ",", "other_fullname", ",", "name", ",", "len", "(", "fullname", ")", ")", ")", ")", "if", "(", "len", "(", "fullname", ")", ">", "len", "(", "other_fullname", ")", ")", ":", "continue", "members", "[", "name", "]", "=", "(", "fullname", ",", "member", ")", "return", "members" ]
collect all symbols from a list of modules .
train
true
50,502
def TR2(rv): def f(rv): if (rv.func is tan): a = rv.args[0] return (sin(a) / cos(a)) elif (rv.func is cot): a = rv.args[0] return (cos(a) / sin(a)) return rv return bottom_up(rv, f)
[ "def", "TR2", "(", "rv", ")", ":", "def", "f", "(", "rv", ")", ":", "if", "(", "rv", ".", "func", "is", "tan", ")", ":", "a", "=", "rv", ".", "args", "[", "0", "]", "return", "(", "sin", "(", "a", ")", "/", "cos", "(", "a", ")", ")", "elif", "(", "rv", ".", "func", "is", "cot", ")", ":", "a", "=", "rv", ".", "args", "[", "0", "]", "return", "(", "cos", "(", "a", ")", "/", "sin", "(", "a", ")", ")", "return", "rv", "return", "bottom_up", "(", "rv", ",", "f", ")" ]
replace tan and cot with sin/cos and cos/sin examples .
train
false
50,503
def is_string_like(obj): if isinstance(obj, (str, unicode)): return True if ma.isMaskedArray(obj): if ((obj.ndim == 0) and (obj.dtype.kind in 'SU')): return True else: return False try: (obj + '') except (TypeError, ValueError): return False return True
[ "def", "is_string_like", "(", "obj", ")", ":", "if", "isinstance", "(", "obj", ",", "(", "str", ",", "unicode", ")", ")", ":", "return", "True", "if", "ma", ".", "isMaskedArray", "(", "obj", ")", ":", "if", "(", "(", "obj", ".", "ndim", "==", "0", ")", "and", "(", "obj", ".", "dtype", ".", "kind", "in", "'SU'", ")", ")", ":", "return", "True", "else", ":", "return", "False", "try", ":", "(", "obj", "+", "''", ")", "except", "(", "TypeError", ",", "ValueError", ")", ":", "return", "False", "return", "True" ]
return true if *obj* looks like a string .
train
false
50,504
def check_headers(request): return HttpResponse(('HTTP_X_ARG_CHECK: %s' % request.META.get('HTTP_X_ARG_CHECK', 'Undefined')))
[ "def", "check_headers", "(", "request", ")", ":", "return", "HttpResponse", "(", "(", "'HTTP_X_ARG_CHECK: %s'", "%", "request", ".", "META", ".", "get", "(", "'HTTP_X_ARG_CHECK'", ",", "'Undefined'", ")", ")", ")" ]
a view that responds with value of the x-arg-check header .
train
false
50,505
def isPackage(module): if (not isinstance(module, types.ModuleType)): return False basename = os.path.splitext(os.path.basename(module.__file__))[0] return (basename == '__init__')
[ "def", "isPackage", "(", "module", ")", ":", "if", "(", "not", "isinstance", "(", "module", ",", "types", ".", "ModuleType", ")", ")", ":", "return", "False", "basename", "=", "os", ".", "path", ".", "splitext", "(", "os", ".", "path", ".", "basename", "(", "module", ".", "__file__", ")", ")", "[", "0", "]", "return", "(", "basename", "==", "'__init__'", ")" ]
given an object return true if the object looks like a package .
train
false
50,506
def strip_self(f): def wrapper(self, *args, **kwargs): return f(*args, **kwargs) return wrapper
[ "def", "strip_self", "(", "f", ")", ":", "def", "wrapper", "(", "self", ",", "*", "args", ",", "**", "kwargs", ")", ":", "return", "f", "(", "*", "args", ",", "**", "kwargs", ")", "return", "wrapper" ]
wrapper to attach module level functions to base class .
train
false
50,508
def invalidate_table_constraints(func): def _cache_clear(self, table, *args, **opts): db_name = self._get_setting('NAME') if (db_name in self._constraint_cache): del self._constraint_cache[db_name] if (db_name in self._reverse_cache): del self._reverse_cache[db_name] if (db_name in self._constraint_references): del self._constraint_references[db_name] return func(self, table, *args, **opts) return _cache_clear
[ "def", "invalidate_table_constraints", "(", "func", ")", ":", "def", "_cache_clear", "(", "self", ",", "table", ",", "*", "args", ",", "**", "opts", ")", ":", "db_name", "=", "self", ".", "_get_setting", "(", "'NAME'", ")", "if", "(", "db_name", "in", "self", ".", "_constraint_cache", ")", ":", "del", "self", ".", "_constraint_cache", "[", "db_name", "]", "if", "(", "db_name", "in", "self", ".", "_reverse_cache", ")", ":", "del", "self", ".", "_reverse_cache", "[", "db_name", "]", "if", "(", "db_name", "in", "self", ".", "_constraint_references", ")", ":", "del", "self", ".", "_constraint_references", "[", "db_name", "]", "return", "func", "(", "self", ",", "table", ",", "*", "args", ",", "**", "opts", ")", "return", "_cache_clear" ]
for mysql we grab all table constraints simultaneously .
train
false
50,509
def ulp(x): x = abs(float(x)) if (math.isnan(x) or math.isinf(x)): return x n = struct.unpack('<q', struct.pack('<d', x))[0] x_next = struct.unpack('<d', struct.pack('<q', (n + 1)))[0] if math.isinf(x_next): x_prev = struct.unpack('<d', struct.pack('<q', (n - 1)))[0] return (x - x_prev) else: return (x_next - x)
[ "def", "ulp", "(", "x", ")", ":", "x", "=", "abs", "(", "float", "(", "x", ")", ")", "if", "(", "math", ".", "isnan", "(", "x", ")", "or", "math", ".", "isinf", "(", "x", ")", ")", ":", "return", "x", "n", "=", "struct", ".", "unpack", "(", "'<q'", ",", "struct", ".", "pack", "(", "'<d'", ",", "x", ")", ")", "[", "0", "]", "x_next", "=", "struct", ".", "unpack", "(", "'<d'", ",", "struct", ".", "pack", "(", "'<q'", ",", "(", "n", "+", "1", ")", ")", ")", "[", "0", "]", "if", "math", ".", "isinf", "(", "x_next", ")", ":", "x_prev", "=", "struct", ".", "unpack", "(", "'<d'", ",", "struct", ".", "pack", "(", "'<q'", ",", "(", "n", "-", "1", ")", ")", ")", "[", "0", "]", "return", "(", "x", "-", "x_prev", ")", "else", ":", "return", "(", "x_next", "-", "x", ")" ]
return the value of the least significant bit of a float x .
train
false
50,510
def _retry_on_port_collision(reason, add, cleanup): reason.trap(AddressInUse) (ip, port) = reason.value.address used_by = _find_process_name(port) ADDRESS_IN_USE(ip=ip, port=port, name=used_by).write() d = cleanup() d.addCallback((lambda ignored: add())) return d
[ "def", "_retry_on_port_collision", "(", "reason", ",", "add", ",", "cleanup", ")", ":", "reason", ".", "trap", "(", "AddressInUse", ")", "(", "ip", ",", "port", ")", "=", "reason", ".", "value", ".", "address", "used_by", "=", "_find_process_name", "(", "port", ")", "ADDRESS_IN_USE", "(", "ip", "=", "ip", ",", "port", "=", "port", ",", "name", "=", "used_by", ")", ".", "write", "(", ")", "d", "=", "cleanup", "(", ")", "d", ".", "addCallback", "(", "(", "lambda", "ignored", ":", "add", "(", ")", ")", ")", "return", "d" ]
cleanup and re-add a container if it failed to start because of a port collision .
train
false
50,511
def set_access_token(access_token): global _access_token _access_token = access_token
[ "def", "set_access_token", "(", "access_token", ")", ":", "global", "_access_token", "_access_token", "=", "access_token" ]
set the shared access token to authenticate with cloud storage .
train
false
50,512
def validate_vector_shape(name, shape, nrows, nobs): ndim = len(shape) if (ndim not in [1, 2]): raise ValueError(('Invalid value for %s vector. Requires a 1- or 2-dimensional array, got %d dimensions' % (name, ndim))) if (not (shape[0] == nrows)): raise ValueError(('Invalid dimensions for %s vector: requires %d rows, got %d' % (name, nrows, shape[0]))) if ((nobs is None) and (not ((ndim == 1) or (shape[(-1)] == 1)))): raise ValueError(('Invalid dimensions for %s vector: time-varying vectors cannot be given unless `nobs` is specified (implicitly when a dataset is bound or else set explicity)' % name)) if ((ndim == 2) and (not (shape[1] in [1, nobs]))): raise ValueError(('Invalid dimensions for time-varying %s vector. Requires shape (*,%d), got %s' % (name, nobs, str(shape))))
[ "def", "validate_vector_shape", "(", "name", ",", "shape", ",", "nrows", ",", "nobs", ")", ":", "ndim", "=", "len", "(", "shape", ")", "if", "(", "ndim", "not", "in", "[", "1", ",", "2", "]", ")", ":", "raise", "ValueError", "(", "(", "'Invalid value for %s vector. Requires a 1- or 2-dimensional array, got %d dimensions'", "%", "(", "name", ",", "ndim", ")", ")", ")", "if", "(", "not", "(", "shape", "[", "0", "]", "==", "nrows", ")", ")", ":", "raise", "ValueError", "(", "(", "'Invalid dimensions for %s vector: requires %d rows, got %d'", "%", "(", "name", ",", "nrows", ",", "shape", "[", "0", "]", ")", ")", ")", "if", "(", "(", "nobs", "is", "None", ")", "and", "(", "not", "(", "(", "ndim", "==", "1", ")", "or", "(", "shape", "[", "(", "-", "1", ")", "]", "==", "1", ")", ")", ")", ")", ":", "raise", "ValueError", "(", "(", "'Invalid dimensions for %s vector: time-varying vectors cannot be given unless `nobs` is specified (implicitly when a dataset is bound or else set explicity)'", "%", "name", ")", ")", "if", "(", "(", "ndim", "==", "2", ")", "and", "(", "not", "(", "shape", "[", "1", "]", "in", "[", "1", ",", "nobs", "]", ")", ")", ")", ":", "raise", "ValueError", "(", "(", "'Invalid dimensions for time-varying %s vector. Requires shape (*,%d), got %s'", "%", "(", "name", ",", "nobs", ",", "str", "(", "shape", ")", ")", ")", ")" ]
validate the shape of a possibly time-varying vector .
train
false
50,514
def _check_stats(conn): stats = conn.get_stats() if (not stats): raise CommandExecutionError('memcached server is down or does not exist') return stats
[ "def", "_check_stats", "(", "conn", ")", ":", "stats", "=", "conn", ".", "get_stats", "(", ")", "if", "(", "not", "stats", ")", ":", "raise", "CommandExecutionError", "(", "'memcached server is down or does not exist'", ")", "return", "stats" ]
helper function to check the stats data passed into it .
train
false
50,515
def test_simple_upload(): group = worker.WalTransferGroup(FakeWalUploader()) seg = FakeWalSegment((('1' * 8) * 3), explicit=True) group.start(seg) group.join() assert success(seg)
[ "def", "test_simple_upload", "(", ")", ":", "group", "=", "worker", ".", "WalTransferGroup", "(", "FakeWalUploader", "(", ")", ")", "seg", "=", "FakeWalSegment", "(", "(", "(", "'1'", "*", "8", ")", "*", "3", ")", ",", "explicit", "=", "True", ")", "group", ".", "start", "(", "seg", ")", "group", ".", "join", "(", ")", "assert", "success", "(", "seg", ")" ]
model a case where there is no concurrency while uploading .
train
false
50,516
@when(u'we save a named query') def step_save_named_query(context): context.cli.sendline(u'\\ns foo SELECT 12345')
[ "@", "when", "(", "u'we save a named query'", ")", "def", "step_save_named_query", "(", "context", ")", ":", "context", ".", "cli", ".", "sendline", "(", "u'\\\\ns foo SELECT 12345'", ")" ]
send s command .
train
false
50,518
def get_resource_path(venv, package=None, resource=None): _verify_safe_py_code(package, resource) bin_path = _verify_virtualenv(venv) ret = __salt__['cmd.exec_code_all'](bin_path, "import pkg_resources; print(pkg_resources.resource_filename('{0}', '{1}'))".format(package, resource)) if (ret['retcode'] != 0): raise CommandExecutionError('{stdout}\n{stderr}'.format(**ret)) return ret['stdout']
[ "def", "get_resource_path", "(", "venv", ",", "package", "=", "None", ",", "resource", "=", "None", ")", ":", "_verify_safe_py_code", "(", "package", ",", "resource", ")", "bin_path", "=", "_verify_virtualenv", "(", "venv", ")", "ret", "=", "__salt__", "[", "'cmd.exec_code_all'", "]", "(", "bin_path", ",", "\"import pkg_resources; print(pkg_resources.resource_filename('{0}', '{1}'))\"", ".", "format", "(", "package", ",", "resource", ")", ")", "if", "(", "ret", "[", "'retcode'", "]", "!=", "0", ")", ":", "raise", "CommandExecutionError", "(", "'{stdout}\\n{stderr}'", ".", "format", "(", "**", "ret", ")", ")", "return", "ret", "[", "'stdout'", "]" ]
return the full path of a resource of the given module .
train
true
50,520
def fill_model_instance(instance, data): for field in instance._meta.fields: if ((not field.editable) or (field.name not in data)): continue field.save_form_data(instance, data[field.name]) return instance
[ "def", "fill_model_instance", "(", "instance", ",", "data", ")", ":", "for", "field", "in", "instance", ".", "_meta", ".", "fields", ":", "if", "(", "(", "not", "field", ".", "editable", ")", "or", "(", "field", ".", "name", "not", "in", "data", ")", ")", ":", "continue", "field", ".", "save_form_data", "(", "instance", ",", "data", "[", "field", ".", "name", "]", ")", "return", "instance" ]
fill whatever fields possible in instance using the data dict .
train
false
50,521
def RequestID(): return os.environ.get(REQUEST_LOG_ID, None)
[ "def", "RequestID", "(", ")", ":", "return", "os", ".", "environ", ".", "get", "(", "REQUEST_LOG_ID", ",", "None", ")" ]
returns the id of the current request assigned by app engine .
train
false
50,525
def in6_isincluded(addr, prefix, plen): temp = inet_pton(socket.AF_INET6, addr) pref = in6_cidr2mask(plen) zero = inet_pton(socket.AF_INET6, prefix) return (zero == in6_and(temp, pref))
[ "def", "in6_isincluded", "(", "addr", ",", "prefix", ",", "plen", ")", ":", "temp", "=", "inet_pton", "(", "socket", ".", "AF_INET6", ",", "addr", ")", "pref", "=", "in6_cidr2mask", "(", "plen", ")", "zero", "=", "inet_pton", "(", "socket", ".", "AF_INET6", ",", "prefix", ")", "return", "(", "zero", "==", "in6_and", "(", "temp", ",", "pref", ")", ")" ]
returns true when addr belongs to prefix/plen .
train
true
50,526
def getCourseNameFromId(courseid): if (courseid == 1): return '' else: q = (db.courses.id == courseid) course_name = db(q).select()[0].course_name return course_name
[ "def", "getCourseNameFromId", "(", "courseid", ")", ":", "if", "(", "courseid", "==", "1", ")", ":", "return", "''", "else", ":", "q", "=", "(", "db", ".", "courses", ".", "id", "==", "courseid", ")", "course_name", "=", "db", "(", "q", ")", ".", "select", "(", ")", "[", "0", "]", ".", "course_name", "return", "course_name" ]
used to compute auth .
train
false
50,527
def url_to_s3_info(url): parsed_url = parse_url(url) assert (parsed_url.scheme == u's3'), (u'You can only use s3: urls (not %r)' % url) (bucket, key) = (parsed_url.host, parsed_url.path) return (bucket, key)
[ "def", "url_to_s3_info", "(", "url", ")", ":", "parsed_url", "=", "parse_url", "(", "url", ")", "assert", "(", "parsed_url", ".", "scheme", "==", "u's3'", ")", ",", "(", "u'You can only use s3: urls (not %r)'", "%", "url", ")", "(", "bucket", ",", "key", ")", "=", "(", "parsed_url", ".", "host", ",", "parsed_url", ".", "path", ")", "return", "(", "bucket", ",", "key", ")" ]
convert an s3 url to a tuple of bucket and key .
train
false
50,529
def p_definition_start(p): print ("start = '%s'" % p[2])
[ "def", "p_definition_start", "(", "p", ")", ":", "print", "(", "\"start = '%s'\"", "%", "p", "[", "2", "]", ")" ]
definition : start id .
train
false
50,530
def create_artifact(entity_id, message_handle, endpoint_index=0): sourceid = sha1(entity_id) ter = ('%s%.2x%s%s' % (ARTIFACT_TYPECODE, endpoint_index, sourceid.digest(), message_handle)) return base64.b64encode(ter)
[ "def", "create_artifact", "(", "entity_id", ",", "message_handle", ",", "endpoint_index", "=", "0", ")", ":", "sourceid", "=", "sha1", "(", "entity_id", ")", "ter", "=", "(", "'%s%.2x%s%s'", "%", "(", "ARTIFACT_TYPECODE", ",", "endpoint_index", ",", "sourceid", ".", "digest", "(", ")", ",", "message_handle", ")", ")", "return", "base64", ".", "b64encode", "(", "ter", ")" ]
saml_artifact := b64 typecode := byte1byte2 endpointindex := byte1byte2 remainingartifact := sourceid messagehandle sourceid := 20-byte_sequence messagehandle := 20-byte_sequence .
train
false
50,531
def _ver_from_ver_str(ver_str): ver = [] for s in ver_str.split('.'): try: ver.append(int(s)) except ValueError: ver.append(s) return tuple(ver)
[ "def", "_ver_from_ver_str", "(", "ver_str", ")", ":", "ver", "=", "[", "]", "for", "s", "in", "ver_str", ".", "split", "(", "'.'", ")", ":", "try", ":", "ver", ".", "append", "(", "int", "(", "s", ")", ")", "except", "ValueError", ":", "ver", ".", "append", "(", "s", ")", "return", "tuple", "(", "ver", ")" ]
convert a version string to a version object as used internally for the "stdlibs" area of the database .
train
false
50,532
def _setDBMSAuthentication(): if (not conf.dbmsCred): return debugMsg = 'setting the DBMS authentication credentials' logger.debug(debugMsg) match = re.search('^(.+?):(.*?)$', conf.dbmsCred) if (not match): errMsg = 'DBMS authentication credentials value must be in format ' errMsg += 'username:password' raise SqlmapSyntaxException(errMsg) conf.dbmsUsername = match.group(1) conf.dbmsPassword = match.group(2)
[ "def", "_setDBMSAuthentication", "(", ")", ":", "if", "(", "not", "conf", ".", "dbmsCred", ")", ":", "return", "debugMsg", "=", "'setting the DBMS authentication credentials'", "logger", ".", "debug", "(", "debugMsg", ")", "match", "=", "re", ".", "search", "(", "'^(.+?):(.*?)$'", ",", "conf", ".", "dbmsCred", ")", "if", "(", "not", "match", ")", ":", "errMsg", "=", "'DBMS authentication credentials value must be in format '", "errMsg", "+=", "'username:password'", "raise", "SqlmapSyntaxException", "(", "errMsg", ")", "conf", ".", "dbmsUsername", "=", "match", ".", "group", "(", "1", ")", "conf", ".", "dbmsPassword", "=", "match", ".", "group", "(", "2", ")" ]
check and set the dbms authentication credentials to run statements as another user .
train
false
50,533
def tuple_variable(t): raise NotImplementedError()
[ "def", "tuple_variable", "(", "t", ")", ":", "raise", "NotImplementedError", "(", ")" ]
make a tuplevariable from a tuple t of theanovariables .
train
false
50,534
def update_output_levels(show, hide): if show: for key in show.split(','): state.output[key] = True if hide: for key in hide.split(','): state.output[key] = False
[ "def", "update_output_levels", "(", "show", ",", "hide", ")", ":", "if", "show", ":", "for", "key", "in", "show", ".", "split", "(", "','", ")", ":", "state", ".", "output", "[", "key", "]", "=", "True", "if", "hide", ":", "for", "key", "in", "hide", ".", "split", "(", "','", ")", ":", "state", ".", "output", "[", "key", "]", "=", "False" ]
update state .
train
false
50,535
def re_unescape(s): return _re_unescape_pattern.sub(_re_unescape_replacement, s)
[ "def", "re_unescape", "(", "s", ")", ":", "return", "_re_unescape_pattern", ".", "sub", "(", "_re_unescape_replacement", ",", "s", ")" ]
unescape a string escaped by re .
train
false
50,537
@open_file(1, mode='wt') def write_graph6(G, path, nodes=None, header=True): path.write(generate_graph6(G, nodes=nodes, header=header)) path.write('\n')
[ "@", "open_file", "(", "1", ",", "mode", "=", "'wt'", ")", "def", "write_graph6", "(", "G", ",", "path", ",", "nodes", "=", "None", ",", "header", "=", "True", ")", ":", "path", ".", "write", "(", "generate_graph6", "(", "G", ",", "nodes", "=", "nodes", ",", "header", "=", "header", ")", ")", "path", ".", "write", "(", "'\\n'", ")" ]
write a simple undirected graph to path in graph6 format .
train
false
50,538
def c_help(client): cmds = ('help', 'help @teleport', 'help look', 'help @tunnel', 'help @dig') return cmds
[ "def", "c_help", "(", "client", ")", ":", "cmds", "=", "(", "'help'", ",", "'help @teleport'", ",", "'help look'", ",", "'help @tunnel'", ",", "'help @dig'", ")", "return", "cmds" ]
reads help files .
train
false
50,540
def find_changes(accounts, monitor_names, debug=True): for account_name in accounts: monitors = get_monitors(account_name, monitor_names, debug) for mon in monitors: cw = mon.watcher (items, exception_map) = cw.slurp() cw.find_changes(current=items, exception_map=exception_map) cw.save() audit_changes(accounts, monitor_names, False, debug) db.session.close()
[ "def", "find_changes", "(", "accounts", ",", "monitor_names", ",", "debug", "=", "True", ")", ":", "for", "account_name", "in", "accounts", ":", "monitors", "=", "get_monitors", "(", "account_name", ",", "monitor_names", ",", "debug", ")", "for", "mon", "in", "monitors", ":", "cw", "=", "mon", ".", "watcher", "(", "items", ",", "exception_map", ")", "=", "cw", ".", "slurp", "(", ")", "cw", ".", "find_changes", "(", "current", "=", "items", ",", "exception_map", "=", "exception_map", ")", "cw", ".", "save", "(", ")", "audit_changes", "(", "accounts", ",", "monitor_names", ",", "False", ",", "debug", ")", "db", ".", "session", ".", "close", "(", ")" ]
runs watchers .
train
false
50,541
def filer_folder_from_path(path): if (path is None): return None folder = None for component in six.text_type(path).split('/'): if component: folder = Folder.objects.get_or_create(name=component, parent=folder)[0] return folder
[ "def", "filer_folder_from_path", "(", "path", ")", ":", "if", "(", "path", "is", "None", ")", ":", "return", "None", "folder", "=", "None", "for", "component", "in", "six", ".", "text_type", "(", "path", ")", ".", "split", "(", "'/'", ")", ":", "if", "component", ":", "folder", "=", "Folder", ".", "objects", ".", "get_or_create", "(", "name", "=", "component", ",", "parent", "=", "folder", ")", "[", "0", "]", "return", "folder" ]
split path by slashes and create a hierarchy of filer folder objects accordingly .
train
false
50,542
def get_user_email(user): return user.email()
[ "def", "get_user_email", "(", "user", ")", ":", "return", "user", ".", "email", "(", ")" ]
given an user object .
train
false
50,543
def gf_sub_mul(f, g, h, p, K): return gf_sub(f, gf_mul(g, h, p, K), p, K)
[ "def", "gf_sub_mul", "(", "f", ",", "g", ",", "h", ",", "p", ",", "K", ")", ":", "return", "gf_sub", "(", "f", ",", "gf_mul", "(", "g", ",", "h", ",", "p", ",", "K", ")", ",", "p", ",", "K", ")" ]
compute f - g*h where f .
train
false
50,545
def print_report_info(results): print('Report Infos:') print(('Contains Sampled Data = %s' % results.get('containsSampledData'))) print(('Kind = %s' % results.get('kind'))) print(('ID = %s' % results.get('id'))) print(('Self Link = %s' % results.get('selfLink'))) print()
[ "def", "print_report_info", "(", "results", ")", ":", "print", "(", "'Report Infos:'", ")", "print", "(", "(", "'Contains Sampled Data = %s'", "%", "results", ".", "get", "(", "'containsSampledData'", ")", ")", ")", "print", "(", "(", "'Kind = %s'", "%", "results", ".", "get", "(", "'kind'", ")", ")", ")", "print", "(", "(", "'ID = %s'", "%", "results", ".", "get", "(", "'id'", ")", ")", ")", "print", "(", "(", "'Self Link = %s'", "%", "results", ".", "get", "(", "'selfLink'", ")", ")", ")", "print", "(", ")" ]
prints general information about this report .
train
false
50,546
def test_lambda_list_keywords_key(): can_compile(u'(fn (x &key {foo True}) (list x foo))') cant_compile(u'(fn (x &key {bar "baz"} &key {foo 42}) (list x bar foo))') cant_compile(u'(fn (x &key {1 2 3 4}) (list x))')
[ "def", "test_lambda_list_keywords_key", "(", ")", ":", "can_compile", "(", "u'(fn (x &key {foo True}) (list x foo))'", ")", "cant_compile", "(", "u'(fn (x &key {bar \"baz\"} &key {foo 42}) (list x bar foo))'", ")", "cant_compile", "(", "u'(fn (x &key {1 2 3 4}) (list x))'", ")" ]
ensure we can compile functions with &key .
train
false
50,547
def _add_directive(block, directive, replace): directive = nginxparser.UnspacedList(directive) if ((len(directive) == 0) or (directive[0] == '#')): block.append(directive) return location = next((index for (index, line) in enumerate(block) if (line and (line[0] == directive[0]))), None) if replace: if (location is None): raise errors.MisconfigurationError('expected directive for {0} in the Nginx config but did not find it.'.format(directive[0])) block[location] = directive _comment_directive(block, location) else: directive_name = directive[0] directive_value = directive[1] if ((location is None) or (isinstance(directive_name, str) and (directive_name in REPEATABLE_DIRECTIVES))): block.append(directive) _comment_directive(block, (len(block) - 1)) elif (block[location][1] != directive_value): raise errors.MisconfigurationError('tried to insert directive "{0}" but found conflicting "{1}".'.format(directive, block[location]))
[ "def", "_add_directive", "(", "block", ",", "directive", ",", "replace", ")", ":", "directive", "=", "nginxparser", ".", "UnspacedList", "(", "directive", ")", "if", "(", "(", "len", "(", "directive", ")", "==", "0", ")", "or", "(", "directive", "[", "0", "]", "==", "'#'", ")", ")", ":", "block", ".", "append", "(", "directive", ")", "return", "location", "=", "next", "(", "(", "index", "for", "(", "index", ",", "line", ")", "in", "enumerate", "(", "block", ")", "if", "(", "line", "and", "(", "line", "[", "0", "]", "==", "directive", "[", "0", "]", ")", ")", ")", ",", "None", ")", "if", "replace", ":", "if", "(", "location", "is", "None", ")", ":", "raise", "errors", ".", "MisconfigurationError", "(", "'expected directive for {0} in the Nginx config but did not find it.'", ".", "format", "(", "directive", "[", "0", "]", ")", ")", "block", "[", "location", "]", "=", "directive", "_comment_directive", "(", "block", ",", "location", ")", "else", ":", "directive_name", "=", "directive", "[", "0", "]", "directive_value", "=", "directive", "[", "1", "]", "if", "(", "(", "location", "is", "None", ")", "or", "(", "isinstance", "(", "directive_name", ",", "str", ")", "and", "(", "directive_name", "in", "REPEATABLE_DIRECTIVES", ")", ")", ")", ":", "block", ".", "append", "(", "directive", ")", "_comment_directive", "(", "block", ",", "(", "len", "(", "block", ")", "-", "1", ")", ")", "elif", "(", "block", "[", "location", "]", "[", "1", "]", "!=", "directive_value", ")", ":", "raise", "errors", ".", "MisconfigurationError", "(", "'tried to insert directive \"{0}\" but found conflicting \"{1}\".'", ".", "format", "(", "directive", ",", "block", "[", "location", "]", ")", ")" ]
adds or replaces a single directive in a config block .
train
false
50,548
def test_dict_key_completion_invalids(): ip = get_ipython() complete = ip.Completer.complete ip.user_ns['no_getitem'] = None ip.user_ns['no_keys'] = [] ip.user_ns['cant_call_keys'] = dict ip.user_ns['empty'] = {} ip.user_ns['d'] = {'abc': 5} (_, matches) = complete(line_buffer="no_getitem['") (_, matches) = complete(line_buffer="no_keys['") (_, matches) = complete(line_buffer="cant_call_keys['") (_, matches) = complete(line_buffer="empty['") (_, matches) = complete(line_buffer="name_error['") (_, matches) = complete(line_buffer="d['\\")
[ "def", "test_dict_key_completion_invalids", "(", ")", ":", "ip", "=", "get_ipython", "(", ")", "complete", "=", "ip", ".", "Completer", ".", "complete", "ip", ".", "user_ns", "[", "'no_getitem'", "]", "=", "None", "ip", ".", "user_ns", "[", "'no_keys'", "]", "=", "[", "]", "ip", ".", "user_ns", "[", "'cant_call_keys'", "]", "=", "dict", "ip", ".", "user_ns", "[", "'empty'", "]", "=", "{", "}", "ip", ".", "user_ns", "[", "'d'", "]", "=", "{", "'abc'", ":", "5", "}", "(", "_", ",", "matches", ")", "=", "complete", "(", "line_buffer", "=", "\"no_getitem['\"", ")", "(", "_", ",", "matches", ")", "=", "complete", "(", "line_buffer", "=", "\"no_keys['\"", ")", "(", "_", ",", "matches", ")", "=", "complete", "(", "line_buffer", "=", "\"cant_call_keys['\"", ")", "(", "_", ",", "matches", ")", "=", "complete", "(", "line_buffer", "=", "\"empty['\"", ")", "(", "_", ",", "matches", ")", "=", "complete", "(", "line_buffer", "=", "\"name_error['\"", ")", "(", "_", ",", "matches", ")", "=", "complete", "(", "line_buffer", "=", "\"d['\\\\\"", ")" ]
smoke test cases dict key completion cant handle .
train
false
50,549
def zeroize(name): ret = {'name': name, 'changes': {}, 'result': True, 'comment': ''} ret['changes'] = __salt__['junos.zeroize']() return ret
[ "def", "zeroize", "(", "name", ")", ":", "ret", "=", "{", "'name'", ":", "name", ",", "'changes'", ":", "{", "}", ",", "'result'", ":", "True", ",", "'comment'", ":", "''", "}", "ret", "[", "'changes'", "]", "=", "__salt__", "[", "'junos.zeroize'", "]", "(", ")", "return", "ret" ]
resets the device to default factory settings usage: .
train
false
50,550
@register.filter(is_safe=True) def markdown(value, arg=''): import warnings warnings.warn('The markdown filter has been deprecated', category=DeprecationWarning) try: import markdown except ImportError: if settings.DEBUG: raise template.TemplateSyntaxError("Error in 'markdown' filter: The Python markdown library isn't installed.") return force_text(value) else: markdown_vers = getattr(markdown, 'version_info', 0) if (markdown_vers < (2, 1)): if settings.DEBUG: raise template.TemplateSyntaxError("Error in 'markdown' filter: Django does not support versions of the Python markdown library < 2.1.") return force_text(value) else: extensions = [e for e in arg.split(',') if e] if (extensions and (extensions[0] == 'safe')): extensions = extensions[1:] return mark_safe(markdown.markdown(force_text(value), extensions, safe_mode=True, enable_attributes=False)) else: return mark_safe(markdown.markdown(force_text(value), extensions, safe_mode=False))
[ "@", "register", ".", "filter", "(", "is_safe", "=", "True", ")", "def", "markdown", "(", "value", ",", "arg", "=", "''", ")", ":", "import", "warnings", "warnings", ".", "warn", "(", "'The markdown filter has been deprecated'", ",", "category", "=", "DeprecationWarning", ")", "try", ":", "import", "markdown", "except", "ImportError", ":", "if", "settings", ".", "DEBUG", ":", "raise", "template", ".", "TemplateSyntaxError", "(", "\"Error in 'markdown' filter: The Python markdown library isn't installed.\"", ")", "return", "force_text", "(", "value", ")", "else", ":", "markdown_vers", "=", "getattr", "(", "markdown", ",", "'version_info'", ",", "0", ")", "if", "(", "markdown_vers", "<", "(", "2", ",", "1", ")", ")", ":", "if", "settings", ".", "DEBUG", ":", "raise", "template", ".", "TemplateSyntaxError", "(", "\"Error in 'markdown' filter: Django does not support versions of the Python markdown library < 2.1.\"", ")", "return", "force_text", "(", "value", ")", "else", ":", "extensions", "=", "[", "e", "for", "e", "in", "arg", ".", "split", "(", "','", ")", "if", "e", "]", "if", "(", "extensions", "and", "(", "extensions", "[", "0", "]", "==", "'safe'", ")", ")", ":", "extensions", "=", "extensions", "[", "1", ":", "]", "return", "mark_safe", "(", "markdown", ".", "markdown", "(", "force_text", "(", "value", ")", ",", "extensions", ",", "safe_mode", "=", "True", ",", "enable_attributes", "=", "False", ")", ")", "else", ":", "return", "mark_safe", "(", "markdown", ".", "markdown", "(", "force_text", "(", "value", ")", ",", "extensions", ",", "safe_mode", "=", "False", ")", ")" ]
runs markdown over a given value .
train
true
50,551
def decode_jpeg(image_buffer, scope=None): with tf.op_scope([image_buffer], scope, 'decode_jpeg'): image = tf.image.decode_jpeg(image_buffer, channels=3) image = tf.image.convert_image_dtype(image, dtype=tf.float32) return image
[ "def", "decode_jpeg", "(", "image_buffer", ",", "scope", "=", "None", ")", ":", "with", "tf", ".", "op_scope", "(", "[", "image_buffer", "]", ",", "scope", ",", "'decode_jpeg'", ")", ":", "image", "=", "tf", ".", "image", ".", "decode_jpeg", "(", "image_buffer", ",", "channels", "=", "3", ")", "image", "=", "tf", ".", "image", ".", "convert_image_dtype", "(", "image", ",", "dtype", "=", "tf", ".", "float32", ")", "return", "image" ]
decode a jpeg string into one 3-d float image tensor .
train
false
50,552
def set_default_fetch_deadline(value): _thread_local_settings.default_fetch_deadline = value
[ "def", "set_default_fetch_deadline", "(", "value", ")", ":", "_thread_local_settings", ".", "default_fetch_deadline", "=", "value" ]
set the default value for create_rpc()s deadline parameter .
train
false
50,553
def test_two_active_items(): items = [Item(QUrl(), '', active=True), Item(QUrl(), ''), Item(QUrl(), '', active=True)] with pytest.raises(ValueError): tabhistory.serialize(items)
[ "def", "test_two_active_items", "(", ")", ":", "items", "=", "[", "Item", "(", "QUrl", "(", ")", ",", "''", ",", "active", "=", "True", ")", ",", "Item", "(", "QUrl", "(", ")", ",", "''", ")", ",", "Item", "(", "QUrl", "(", ")", ",", "''", ",", "active", "=", "True", ")", "]", "with", "pytest", ".", "raises", "(", "ValueError", ")", ":", "tabhistory", ".", "serialize", "(", "items", ")" ]
check tabhistory .
train
false
50,554
def test_rename_channels(): info = read_info(raw_fname) mapping = {'EEG 160': 'EEG060'} assert_raises(ValueError, rename_channels, info, mapping) mapping = {'MEG 2641': 1.0} assert_raises(ValueError, rename_channels, info, mapping) mapping = {'MEG 2641': 'MEG 2642'} assert_raises(ValueError, rename_channels, info, mapping) assert_raises(ValueError, rename_channels, info, 1.0) info2 = deepcopy(info) info2['bads'] = ['EEG 060', 'EOG 061'] mapping = {'EEG 060': 'EEG060', 'EOG 061': 'EOG061'} rename_channels(info2, mapping) assert_true((info2['chs'][374]['ch_name'] == 'EEG060')) assert_true((info2['ch_names'][374] == 'EEG060')) assert_true((info2['chs'][375]['ch_name'] == 'EOG061')) assert_true((info2['ch_names'][375] == 'EOG061')) assert_array_equal(['EEG060', 'EOG061'], info2['bads']) info2 = deepcopy(info) rename_channels(info2, (lambda x: x.replace(' ', ''))) assert_true((info2['chs'][373]['ch_name'] == 'EEG059')) info2 = deepcopy(info) info2['bads'] = ['EEG 060', 'EEG 060'] rename_channels(info2, mapping) assert_array_equal(['EEG060', 'EEG060'], info2['bads'])
[ "def", "test_rename_channels", "(", ")", ":", "info", "=", "read_info", "(", "raw_fname", ")", "mapping", "=", "{", "'EEG 160'", ":", "'EEG060'", "}", "assert_raises", "(", "ValueError", ",", "rename_channels", ",", "info", ",", "mapping", ")", "mapping", "=", "{", "'MEG 2641'", ":", "1.0", "}", "assert_raises", "(", "ValueError", ",", "rename_channels", ",", "info", ",", "mapping", ")", "mapping", "=", "{", "'MEG 2641'", ":", "'MEG 2642'", "}", "assert_raises", "(", "ValueError", ",", "rename_channels", ",", "info", ",", "mapping", ")", "assert_raises", "(", "ValueError", ",", "rename_channels", ",", "info", ",", "1.0", ")", "info2", "=", "deepcopy", "(", "info", ")", "info2", "[", "'bads'", "]", "=", "[", "'EEG 060'", ",", "'EOG 061'", "]", "mapping", "=", "{", "'EEG 060'", ":", "'EEG060'", ",", "'EOG 061'", ":", "'EOG061'", "}", "rename_channels", "(", "info2", ",", "mapping", ")", "assert_true", "(", "(", "info2", "[", "'chs'", "]", "[", "374", "]", "[", "'ch_name'", "]", "==", "'EEG060'", ")", ")", "assert_true", "(", "(", "info2", "[", "'ch_names'", "]", "[", "374", "]", "==", "'EEG060'", ")", ")", "assert_true", "(", "(", "info2", "[", "'chs'", "]", "[", "375", "]", "[", "'ch_name'", "]", "==", "'EOG061'", ")", ")", "assert_true", "(", "(", "info2", "[", "'ch_names'", "]", "[", "375", "]", "==", "'EOG061'", ")", ")", "assert_array_equal", "(", "[", "'EEG060'", ",", "'EOG061'", "]", ",", "info2", "[", "'bads'", "]", ")", "info2", "=", "deepcopy", "(", "info", ")", "rename_channels", "(", "info2", ",", "(", "lambda", "x", ":", "x", ".", "replace", "(", "' '", ",", "''", ")", ")", ")", "assert_true", "(", "(", "info2", "[", "'chs'", "]", "[", "373", "]", "[", "'ch_name'", "]", "==", "'EEG059'", ")", ")", "info2", "=", "deepcopy", "(", "info", ")", "info2", "[", "'bads'", "]", "=", "[", "'EEG 060'", ",", "'EEG 060'", "]", "rename_channels", "(", "info2", ",", "mapping", ")", "assert_array_equal", "(", "[", "'EEG060'", ",", "'EEG060'", "]", ",", "info2", "[", "'bads'", "]", ")" ]
test rename channels .
train
false
50,555
def ndd_prefix_for_region(region_code, strip_non_digits): if (region_code is None): return None metadata = PhoneMetadata.metadata_for_region(region_code.upper(), None) if (metadata is None): return None national_prefix = metadata.national_prefix if ((national_prefix is None) or (len(national_prefix) == 0)): return None if strip_non_digits: national_prefix = re.sub(U_TILDE, U_EMPTY_STRING, national_prefix) return national_prefix
[ "def", "ndd_prefix_for_region", "(", "region_code", ",", "strip_non_digits", ")", ":", "if", "(", "region_code", "is", "None", ")", ":", "return", "None", "metadata", "=", "PhoneMetadata", ".", "metadata_for_region", "(", "region_code", ".", "upper", "(", ")", ",", "None", ")", "if", "(", "metadata", "is", "None", ")", ":", "return", "None", "national_prefix", "=", "metadata", ".", "national_prefix", "if", "(", "(", "national_prefix", "is", "None", ")", "or", "(", "len", "(", "national_prefix", ")", "==", "0", ")", ")", ":", "return", "None", "if", "strip_non_digits", ":", "national_prefix", "=", "re", ".", "sub", "(", "U_TILDE", ",", "U_EMPTY_STRING", ",", "national_prefix", ")", "return", "national_prefix" ]
returns the national dialling prefix for a specific region .
train
true
50,556
def test_run_method_should_return_success_when_command_name_not_specified(): options_mock = Mock() args = () help_cmd = HelpCommand() status = help_cmd.run(options_mock, args) assert (status == SUCCESS)
[ "def", "test_run_method_should_return_success_when_command_name_not_specified", "(", ")", ":", "options_mock", "=", "Mock", "(", ")", "args", "=", "(", ")", "help_cmd", "=", "HelpCommand", "(", ")", "status", "=", "help_cmd", ".", "run", "(", "options_mock", ",", "args", ")", "assert", "(", "status", "==", "SUCCESS", ")" ]
test helpcommand .
train
false
50,558
def mkNonce(when=None): salt = cryptutil.randomString(6, NONCE_CHARS) if (when is None): t = gmtime() else: t = gmtime(when) time_str = strftime(time_fmt, t) return (time_str + salt)
[ "def", "mkNonce", "(", "when", "=", "None", ")", ":", "salt", "=", "cryptutil", ".", "randomString", "(", "6", ",", "NONCE_CHARS", ")", "if", "(", "when", "is", "None", ")", ":", "t", "=", "gmtime", "(", ")", "else", ":", "t", "=", "gmtime", "(", "when", ")", "time_str", "=", "strftime", "(", "time_fmt", ",", "t", ")", "return", "(", "time_str", "+", "salt", ")" ]
generate a nonce with the current timestamp .
train
true
50,559
def Run(arg_dict, oauth2_parameters=None): arg_dict = ProcessArguments(arg_dict) SetupLogging(arg_dict) return _PerformBulkload(arg_dict, oauth2_parameters)
[ "def", "Run", "(", "arg_dict", ",", "oauth2_parameters", "=", "None", ")", ":", "arg_dict", "=", "ProcessArguments", "(", "arg_dict", ")", "SetupLogging", "(", "arg_dict", ")", "return", "_PerformBulkload", "(", "arg_dict", ",", "oauth2_parameters", ")" ]
standalone command line access point .
train
false
50,560
def SetPlatformArchContext(): CONFIG.AddContext(('Platform:%s' % platform.system().title())) machine = platform.uname()[4] if (machine in ['x86_64', 'AMD64', 'i686']): if (platform.architecture()[0] == '32bit'): arch = 'i386' else: arch = 'amd64' elif (machine == 'x86'): arch = 'i386' else: arch = machine CONFIG.AddContext(('Arch:%s' % arch))
[ "def", "SetPlatformArchContext", "(", ")", ":", "CONFIG", ".", "AddContext", "(", "(", "'Platform:%s'", "%", "platform", ".", "system", "(", ")", ".", "title", "(", ")", ")", ")", "machine", "=", "platform", ".", "uname", "(", ")", "[", "4", "]", "if", "(", "machine", "in", "[", "'x86_64'", ",", "'AMD64'", ",", "'i686'", "]", ")", ":", "if", "(", "platform", ".", "architecture", "(", ")", "[", "0", "]", "==", "'32bit'", ")", ":", "arch", "=", "'i386'", "else", ":", "arch", "=", "'amd64'", "elif", "(", "machine", "==", "'x86'", ")", ":", "arch", "=", "'i386'", "else", ":", "arch", "=", "machine", "CONFIG", ".", "AddContext", "(", "(", "'Arch:%s'", "%", "arch", ")", ")" ]
add the running contexts to the config system .
train
true
50,562
def overrideNative(): file_name = (Directory.GetCurrentDirectory() + '\\sys.cs') file = open(file_name, 'w') print >>file, cs_native file.close() compileAssembly(file_name) file_name = (Directory.GetCurrentDirectory() + '\\re.cs') file = open(file_name, 'w') print >>file, cs_native_re file.close() compileAssembly(file_name)
[ "def", "overrideNative", "(", ")", ":", "file_name", "=", "(", "Directory", ".", "GetCurrentDirectory", "(", ")", "+", "'\\\\sys.cs'", ")", "file", "=", "open", "(", "file_name", ",", "'w'", ")", "print", ">>", "file", ",", "cs_native", "file", ".", "close", "(", ")", "compileAssembly", "(", "file_name", ")", "file_name", "=", "(", "Directory", ".", "GetCurrentDirectory", "(", ")", "+", "'\\\\re.cs'", ")", "file", "=", "open", "(", "file_name", ",", "'w'", ")", "print", ">>", "file", ",", "cs_native_re", "file", ".", "close", "(", ")", "compileAssembly", "(", "file_name", ")" ]
tries to override the native ip sys module .
train
false
50,563
def pre_save_title(instance, raw, **kwargs): page = instance.page page_languages = page.get_languages() if (not (instance.language in page_languages)): page_languages.append(instance.language) page.update_languages(page_languages) if (not page.publisher_is_draft): menu_pool.clear(page.site_id) if (instance.pk and (not hasattr(instance, 'tmp_path'))): instance.tmp_path = None try: instance.tmp_path = Title.objects.filter(pk=instance.pk).values_list('path')[0][0] except IndexError: pass if (instance.has_url_overwrite and instance.path): instance.path = instance.path.strip(' /') else: update_title(instance) apphook_pre_title_checker(instance, **kwargs)
[ "def", "pre_save_title", "(", "instance", ",", "raw", ",", "**", "kwargs", ")", ":", "page", "=", "instance", ".", "page", "page_languages", "=", "page", ".", "get_languages", "(", ")", "if", "(", "not", "(", "instance", ".", "language", "in", "page_languages", ")", ")", ":", "page_languages", ".", "append", "(", "instance", ".", "language", ")", "page", ".", "update_languages", "(", "page_languages", ")", "if", "(", "not", "page", ".", "publisher_is_draft", ")", ":", "menu_pool", ".", "clear", "(", "page", ".", "site_id", ")", "if", "(", "instance", ".", "pk", "and", "(", "not", "hasattr", "(", "instance", ",", "'tmp_path'", ")", ")", ")", ":", "instance", ".", "tmp_path", "=", "None", "try", ":", "instance", ".", "tmp_path", "=", "Title", ".", "objects", ".", "filter", "(", "pk", "=", "instance", ".", "pk", ")", ".", "values_list", "(", "'path'", ")", "[", "0", "]", "[", "0", "]", "except", "IndexError", ":", "pass", "if", "(", "instance", ".", "has_url_overwrite", "and", "instance", ".", "path", ")", ":", "instance", ".", "path", "=", "instance", ".", "path", ".", "strip", "(", "' /'", ")", "else", ":", "update_title", "(", "instance", ")", "apphook_pre_title_checker", "(", "instance", ",", "**", "kwargs", ")" ]
save old state to instance and setup path .
train
false
50,564
def groovy(registry, xml_parent, data): root_tag = 'hudson.plugins.groovy.Groovy' groovy = XML.SubElement(xml_parent, root_tag) groovy.append(_groovy_common_scriptSource(data)) XML.SubElement(groovy, 'groovyName').text = str(data.get('version', '(Default)')) XML.SubElement(groovy, 'parameters').text = str(data.get('parameters', '')) XML.SubElement(groovy, 'scriptParameters').text = str(data.get('script-parameters', '')) XML.SubElement(groovy, 'properties').text = str(data.get('properties', '')) XML.SubElement(groovy, 'javaOpts').text = str(data.get('java-opts', '')) XML.SubElement(groovy, 'classPath').text = str(data.get('class-path', ''))
[ "def", "groovy", "(", "registry", ",", "xml_parent", ",", "data", ")", ":", "root_tag", "=", "'hudson.plugins.groovy.Groovy'", "groovy", "=", "XML", ".", "SubElement", "(", "xml_parent", ",", "root_tag", ")", "groovy", ".", "append", "(", "_groovy_common_scriptSource", "(", "data", ")", ")", "XML", ".", "SubElement", "(", "groovy", ",", "'groovyName'", ")", ".", "text", "=", "str", "(", "data", ".", "get", "(", "'version'", ",", "'(Default)'", ")", ")", "XML", ".", "SubElement", "(", "groovy", ",", "'parameters'", ")", ".", "text", "=", "str", "(", "data", ".", "get", "(", "'parameters'", ",", "''", ")", ")", "XML", ".", "SubElement", "(", "groovy", ",", "'scriptParameters'", ")", ".", "text", "=", "str", "(", "data", ".", "get", "(", "'script-parameters'", ",", "''", ")", ")", "XML", ".", "SubElement", "(", "groovy", ",", "'properties'", ")", ".", "text", "=", "str", "(", "data", ".", "get", "(", "'properties'", ",", "''", ")", ")", "XML", ".", "SubElement", "(", "groovy", ",", "'javaOpts'", ")", ".", "text", "=", "str", "(", "data", ".", "get", "(", "'java-opts'", ",", "''", ")", ")", "XML", ".", "SubElement", "(", "groovy", ",", "'classPath'", ")", ".", "text", "=", "str", "(", "data", ".", "get", "(", "'class-path'", ",", "''", ")", ")" ]
yaml: groovy execute a groovy script or command .
train
false
50,565
def construct_change_message(form, formsets, add): change_message = [] if add: change_message.append({'added': {}}) elif form.changed_data: change_message.append({'changed': {'fields': form.changed_data}}) if formsets: with translation_override(None): for formset in formsets: for added_object in formset.new_objects: change_message.append({'added': {'name': force_text(added_object._meta.verbose_name), 'object': force_text(added_object)}}) for (changed_object, changed_fields) in formset.changed_objects: change_message.append({'changed': {'name': force_text(changed_object._meta.verbose_name), 'object': force_text(changed_object), 'fields': changed_fields}}) for deleted_object in formset.deleted_objects: change_message.append({'deleted': {'name': force_text(deleted_object._meta.verbose_name), 'object': force_text(deleted_object)}}) return change_message
[ "def", "construct_change_message", "(", "form", ",", "formsets", ",", "add", ")", ":", "change_message", "=", "[", "]", "if", "add", ":", "change_message", ".", "append", "(", "{", "'added'", ":", "{", "}", "}", ")", "elif", "form", ".", "changed_data", ":", "change_message", ".", "append", "(", "{", "'changed'", ":", "{", "'fields'", ":", "form", ".", "changed_data", "}", "}", ")", "if", "formsets", ":", "with", "translation_override", "(", "None", ")", ":", "for", "formset", "in", "formsets", ":", "for", "added_object", "in", "formset", ".", "new_objects", ":", "change_message", ".", "append", "(", "{", "'added'", ":", "{", "'name'", ":", "force_text", "(", "added_object", ".", "_meta", ".", "verbose_name", ")", ",", "'object'", ":", "force_text", "(", "added_object", ")", "}", "}", ")", "for", "(", "changed_object", ",", "changed_fields", ")", "in", "formset", ".", "changed_objects", ":", "change_message", ".", "append", "(", "{", "'changed'", ":", "{", "'name'", ":", "force_text", "(", "changed_object", ".", "_meta", ".", "verbose_name", ")", ",", "'object'", ":", "force_text", "(", "changed_object", ")", ",", "'fields'", ":", "changed_fields", "}", "}", ")", "for", "deleted_object", "in", "formset", ".", "deleted_objects", ":", "change_message", ".", "append", "(", "{", "'deleted'", ":", "{", "'name'", ":", "force_text", "(", "deleted_object", ".", "_meta", ".", "verbose_name", ")", ",", "'object'", ":", "force_text", "(", "deleted_object", ")", "}", "}", ")", "return", "change_message" ]
construct a json structure describing changes from a changed object .
train
false
50,566
def parseChangeXML(raw_xml): data = dict() try: doc = xml.dom.minidom.parseString(raw_xml) except ExpatError: print('\nError: Got an empty response with an empty changeset.\n') raise log_entry = doc.getElementsByTagName('logentry')[0] data['revision'] = log_entry.getAttribute('revision') data['author'] = ''.join([t.data for t in log_entry.getElementsByTagName('author')[0].childNodes]) data['comments'] = ''.join([t.data for t in log_entry.getElementsByTagName('msg')[0].childNodes]) pathlist = log_entry.getElementsByTagName('paths')[0] paths = [] for path in pathlist.getElementsByTagName('path'): paths.append(''.join([t.data for t in path.childNodes])) data['paths'] = paths return data
[ "def", "parseChangeXML", "(", "raw_xml", ")", ":", "data", "=", "dict", "(", ")", "try", ":", "doc", "=", "xml", ".", "dom", ".", "minidom", ".", "parseString", "(", "raw_xml", ")", "except", "ExpatError", ":", "print", "(", "'\\nError: Got an empty response with an empty changeset.\\n'", ")", "raise", "log_entry", "=", "doc", ".", "getElementsByTagName", "(", "'logentry'", ")", "[", "0", "]", "data", "[", "'revision'", "]", "=", "log_entry", ".", "getAttribute", "(", "'revision'", ")", "data", "[", "'author'", "]", "=", "''", ".", "join", "(", "[", "t", ".", "data", "for", "t", "in", "log_entry", ".", "getElementsByTagName", "(", "'author'", ")", "[", "0", "]", ".", "childNodes", "]", ")", "data", "[", "'comments'", "]", "=", "''", ".", "join", "(", "[", "t", ".", "data", "for", "t", "in", "log_entry", ".", "getElementsByTagName", "(", "'msg'", ")", "[", "0", "]", ".", "childNodes", "]", ")", "pathlist", "=", "log_entry", ".", "getElementsByTagName", "(", "'paths'", ")", "[", "0", "]", "paths", "=", "[", "]", "for", "path", "in", "pathlist", ".", "getElementsByTagName", "(", "'path'", ")", ":", "paths", ".", "append", "(", "''", ".", "join", "(", "[", "t", ".", "data", "for", "t", "in", "path", ".", "childNodes", "]", ")", ")", "data", "[", "'paths'", "]", "=", "paths", "return", "data" ]
parse the raw xml and return a dict with key pairs set .
train
false
50,567
def _api_key_patch_replace(conn, apiKey, path, value): response = conn.update_api_key(apiKey=apiKey, patchOperations=[{'op': 'replace', 'path': path, 'value': value}]) return response
[ "def", "_api_key_patch_replace", "(", "conn", ",", "apiKey", ",", "path", ",", "value", ")", ":", "response", "=", "conn", ".", "update_api_key", "(", "apiKey", "=", "apiKey", ",", "patchOperations", "=", "[", "{", "'op'", ":", "'replace'", ",", "'path'", ":", "path", ",", "'value'", ":", "value", "}", "]", ")", "return", "response" ]
the replace patch operation on an apikey resource .
train
true
50,568
def f_1(): return __doc__
[ "def", "f_1", "(", ")", ":", "return", "__doc__" ]
f 1 doc .
train
false
50,569
def _parse_focus(arg_str): split_str = arg_str.split('|', 1) if (len(split_str) == 1): return {'id': split_str[0].rstrip(), 'annotation': ''} else: return {'id': split_str[0].rstrip(), 'annotation': split_str[1].lstrip()}
[ "def", "_parse_focus", "(", "arg_str", ")", ":", "split_str", "=", "arg_str", ".", "split", "(", "'|'", ",", "1", ")", "if", "(", "len", "(", "split_str", ")", "==", "1", ")", ":", "return", "{", "'id'", ":", "split_str", "[", "0", "]", ".", "rstrip", "(", ")", ",", "'annotation'", ":", "''", "}", "else", ":", "return", "{", "'id'", ":", "split_str", "[", "0", "]", ".", "rstrip", "(", ")", ",", "'annotation'", ":", "split_str", "[", "1", "]", ".", "lstrip", "(", ")", "}" ]
returns id and annotation after splitting input string .
train
false
50,572
def _check_epochs_input(epochs, y, picks=None): if (y is None): y = epochs.events[:, 2] elif isinstance(y, list): y = np.array(y) X = epochs.get_data() if (picks is None): picks = _pick_data_channels(epochs.info, with_ref_meg=False) if isinstance(picks, (list, np.ndarray)): picks = np.array(picks, dtype=np.int) else: raise ValueError('picks must be a list or a numpy.ndarray of int') X = X[:, picks, :] assert (X.shape[0] == y.shape[0]) return (X, y, picks)
[ "def", "_check_epochs_input", "(", "epochs", ",", "y", ",", "picks", "=", "None", ")", ":", "if", "(", "y", "is", "None", ")", ":", "y", "=", "epochs", ".", "events", "[", ":", ",", "2", "]", "elif", "isinstance", "(", "y", ",", "list", ")", ":", "y", "=", "np", ".", "array", "(", "y", ")", "X", "=", "epochs", ".", "get_data", "(", ")", "if", "(", "picks", "is", "None", ")", ":", "picks", "=", "_pick_data_channels", "(", "epochs", ".", "info", ",", "with_ref_meg", "=", "False", ")", "if", "isinstance", "(", "picks", ",", "(", "list", ",", "np", ".", "ndarray", ")", ")", ":", "picks", "=", "np", ".", "array", "(", "picks", ",", "dtype", "=", "np", ".", "int", ")", "else", ":", "raise", "ValueError", "(", "'picks must be a list or a numpy.ndarray of int'", ")", "X", "=", "X", "[", ":", ",", "picks", ",", ":", "]", "assert", "(", "X", ".", "shape", "[", "0", "]", "==", "y", ".", "shape", "[", "0", "]", ")", "return", "(", "X", ",", "y", ",", "picks", ")" ]
aux function of generalizationacrosstime .
train
false
50,573
def counter(): lock = compat.threading.Lock() counter = itertools.count(1) def _next(): lock.acquire() try: return next(counter) finally: lock.release() return _next
[ "def", "counter", "(", ")", ":", "lock", "=", "compat", ".", "threading", ".", "Lock", "(", ")", "counter", "=", "itertools", ".", "count", "(", "1", ")", "def", "_next", "(", ")", ":", "lock", ".", "acquire", "(", ")", "try", ":", "return", "next", "(", "counter", ")", "finally", ":", "lock", ".", "release", "(", ")", "return", "_next" ]
every time you reload .
train
false
50,574
def put_async(models, **kwargs): (models, multiple) = datastore.NormalizeAndTypeCheck(models, Model) entities = [model._populate_internal_entity() for model in models] def extra_hook(keys): if multiple: return keys assert (len(keys) == 1) return keys[0] return datastore.PutAsync(entities, extra_hook=extra_hook, **kwargs)
[ "def", "put_async", "(", "models", ",", "**", "kwargs", ")", ":", "(", "models", ",", "multiple", ")", "=", "datastore", ".", "NormalizeAndTypeCheck", "(", "models", ",", "Model", ")", "entities", "=", "[", "model", ".", "_populate_internal_entity", "(", ")", "for", "model", "in", "models", "]", "def", "extra_hook", "(", "keys", ")", ":", "if", "multiple", ":", "return", "keys", "assert", "(", "len", "(", "keys", ")", "==", "1", ")", "return", "keys", "[", "0", "]", "return", "datastore", ".", "PutAsync", "(", "entities", ",", "extra_hook", "=", "extra_hook", ",", "**", "kwargs", ")" ]
asynchronously store one or more model instances .
train
false
50,575
def getNS(s, count=1): ns = [] c = 0 for i in range(count): (l,) = struct.unpack('!L', s[c:(c + 4)]) ns.append(s[(c + 4):((4 + l) + c)]) c += (4 + l) return (tuple(ns) + (s[c:],))
[ "def", "getNS", "(", "s", ",", "count", "=", "1", ")", ":", "ns", "=", "[", "]", "c", "=", "0", "for", "i", "in", "range", "(", "count", ")", ":", "(", "l", ",", ")", "=", "struct", ".", "unpack", "(", "'!L'", ",", "s", "[", "c", ":", "(", "c", "+", "4", ")", "]", ")", "ns", ".", "append", "(", "s", "[", "(", "c", "+", "4", ")", ":", "(", "(", "4", "+", "l", ")", "+", "c", ")", "]", ")", "c", "+=", "(", "4", "+", "l", ")", "return", "(", "tuple", "(", "ns", ")", "+", "(", "s", "[", "c", ":", "]", ",", ")", ")" ]
get net string .
train
false
50,576
def plot_10_by_10_images(images): fig = plt.figure() images = [image[3:25, 3:25] for image in images] for x in range(10): for y in range(10): ax = fig.add_subplot(10, 10, ((10 * y) + x)) ax.matshow(images[((10 * y) + x)], cmap=matplotlib.cm.binary) plt.xticks(np.array([])) plt.yticks(np.array([])) plt.show()
[ "def", "plot_10_by_10_images", "(", "images", ")", ":", "fig", "=", "plt", ".", "figure", "(", ")", "images", "=", "[", "image", "[", "3", ":", "25", ",", "3", ":", "25", "]", "for", "image", "in", "images", "]", "for", "x", "in", "range", "(", "10", ")", ":", "for", "y", "in", "range", "(", "10", ")", ":", "ax", "=", "fig", ".", "add_subplot", "(", "10", ",", "10", ",", "(", "(", "10", "*", "y", ")", "+", "x", ")", ")", "ax", ".", "matshow", "(", "images", "[", "(", "(", "10", "*", "y", ")", "+", "x", ")", "]", ",", "cmap", "=", "matplotlib", ".", "cm", ".", "binary", ")", "plt", ".", "xticks", "(", "np", ".", "array", "(", "[", "]", ")", ")", "plt", ".", "yticks", "(", "np", ".", "array", "(", "[", "]", ")", ")", "plt", ".", "show", "(", ")" ]
plot 100 mnist images in a 10 by 10 table .
train
false
50,577
def cmd_status(args): if (len(args) == 0): mpstate.status.show(sys.stdout, pattern=None) else: for pattern in args: mpstate.status.show(sys.stdout, pattern=pattern)
[ "def", "cmd_status", "(", "args", ")", ":", "if", "(", "len", "(", "args", ")", "==", "0", ")", ":", "mpstate", ".", "status", ".", "show", "(", "sys", ".", "stdout", ",", "pattern", "=", "None", ")", "else", ":", "for", "pattern", "in", "args", ":", "mpstate", ".", "status", ".", "show", "(", "sys", ".", "stdout", ",", "pattern", "=", "pattern", ")" ]
show status .
train
true
50,579
def parse_timestamp(timestamp, time_format=None): if (time_format is None): time_format = '%a, %d %b %Y %H:%M:%S %Z' try: return time.strptime(timestamp, time_format) except ValueError: try: return time.strptime(re.sub(' \\w+$', '', timestamp), time_format.replace(' %Z', '')) except ValueError: LOG.error(('Failed to convert Oozie timestamp: %s' % time_format)) except Exception: LOG.error(('Failed to convert Oozie timestamp: %s' % time_format)) return None
[ "def", "parse_timestamp", "(", "timestamp", ",", "time_format", "=", "None", ")", ":", "if", "(", "time_format", "is", "None", ")", ":", "time_format", "=", "'%a, %d %b %Y %H:%M:%S %Z'", "try", ":", "return", "time", ".", "strptime", "(", "timestamp", ",", "time_format", ")", "except", "ValueError", ":", "try", ":", "return", "time", ".", "strptime", "(", "re", ".", "sub", "(", "' \\\\w+$'", ",", "''", ",", "timestamp", ")", ",", "time_format", ".", "replace", "(", "' %Z'", ",", "''", ")", ")", "except", "ValueError", ":", "LOG", ".", "error", "(", "(", "'Failed to convert Oozie timestamp: %s'", "%", "time_format", ")", ")", "except", "Exception", ":", "LOG", ".", "error", "(", "(", "'Failed to convert Oozie timestamp: %s'", "%", "time_format", ")", ")", "return", "None" ]
parse_timestamp -> struct_time does not raise valueerror .
train
false
50,580
@pytest.fixture(autouse=True) def remove_replay_dump(request, replay_file): def fin_remove_replay_file(): if os.path.exists(replay_file): os.remove(replay_file) request.addfinalizer(fin_remove_replay_file)
[ "@", "pytest", ".", "fixture", "(", "autouse", "=", "True", ")", "def", "remove_replay_dump", "(", "request", ",", "replay_file", ")", ":", "def", "fin_remove_replay_file", "(", ")", ":", "if", "os", ".", "path", ".", "exists", "(", "replay_file", ")", ":", "os", ".", "remove", "(", "replay_file", ")", "request", ".", "addfinalizer", "(", "fin_remove_replay_file", ")" ]
remove the replay file created by tests .
train
false
50,581
def get_settings(**kwargs): settings = DEFAULT_CONFIG.copy() for (key, value) in kwargs.items(): settings[key] = value return settings
[ "def", "get_settings", "(", "**", "kwargs", ")", ":", "settings", "=", "DEFAULT_CONFIG", ".", "copy", "(", ")", "for", "(", "key", ",", "value", ")", "in", "kwargs", ".", "items", "(", ")", ":", "settings", "[", "key", "]", "=", "value", "return", "settings" ]
provide tweaked setting dictionaries for testing set keyword arguments to override specific settings .
train
false
50,582
def p_statement_expr(): print t[1]
[ "def", "p_statement_expr", "(", ")", ":", "print", "t", "[", "1", "]" ]
statement : expression .
train
false
50,583
@cleanup def test_determinism_images(): _determinism_check(u'i', format=u'pdf')
[ "@", "cleanup", "def", "test_determinism_images", "(", ")", ":", "_determinism_check", "(", "u'i'", ",", "format", "=", "u'pdf'", ")" ]
test for reproducible pdf output: figure with different images .
train
false
50,584
def download_files_in_parallel(urls, cache=False, show_progress=True, timeout=None): from .console import ProgressBar if (timeout is None): timeout = conf.remote_timeout if show_progress: progress = sys.stdout else: progress = io.BytesIO() if (timeout is None): timeout = REMOTE_TIMEOUT() combined_urls = list(set(urls)) combined_paths = ProgressBar.map(_do_download_files_in_parallel, [(x, cache) for x in combined_urls], file=progress, multiprocess=True) paths = [] for url in urls: paths.append(combined_paths[combined_urls.index(url)]) return paths
[ "def", "download_files_in_parallel", "(", "urls", ",", "cache", "=", "False", ",", "show_progress", "=", "True", ",", "timeout", "=", "None", ")", ":", "from", ".", "console", "import", "ProgressBar", "if", "(", "timeout", "is", "None", ")", ":", "timeout", "=", "conf", ".", "remote_timeout", "if", "show_progress", ":", "progress", "=", "sys", ".", "stdout", "else", ":", "progress", "=", "io", ".", "BytesIO", "(", ")", "if", "(", "timeout", "is", "None", ")", ":", "timeout", "=", "REMOTE_TIMEOUT", "(", ")", "combined_urls", "=", "list", "(", "set", "(", "urls", ")", ")", "combined_paths", "=", "ProgressBar", ".", "map", "(", "_do_download_files_in_parallel", ",", "[", "(", "x", ",", "cache", ")", "for", "x", "in", "combined_urls", "]", ",", "file", "=", "progress", ",", "multiprocess", "=", "True", ")", "paths", "=", "[", "]", "for", "url", "in", "urls", ":", "paths", ".", "append", "(", "combined_paths", "[", "combined_urls", ".", "index", "(", "url", ")", "]", ")", "return", "paths" ]
downloads multiple files in parallel from the given urls .
train
false
50,585
def rot_to_quat(rot): rot = rot.reshape((rot.shape[:(-2)] + (9,))) return np.apply_along_axis(_one_rot_to_quat, (-1), rot)
[ "def", "rot_to_quat", "(", "rot", ")", ":", "rot", "=", "rot", ".", "reshape", "(", "(", "rot", ".", "shape", "[", ":", "(", "-", "2", ")", "]", "+", "(", "9", ",", ")", ")", ")", "return", "np", ".", "apply_along_axis", "(", "_one_rot_to_quat", ",", "(", "-", "1", ")", ",", "rot", ")" ]
convert a set of rotations to quaternions .
train
false
50,586
def xml_root_close(): return '</resource>'
[ "def", "xml_root_close", "(", ")", ":", "return", "'</resource>'" ]
returns the closing tag of the xml root node .
train
false
50,587
def extract_thumbnail_number(text): pattr = re.compile('^\\s*#\\s*sphinx_gallery_thumbnail_number\\s*=\\s*([0-9]+)\\s*$', flags=re.MULTILINE) match = pattr.search(text) if (match is None): thumbnail_number = 1 else: thumbnail_number = int(match.groups()[0]) return thumbnail_number
[ "def", "extract_thumbnail_number", "(", "text", ")", ":", "pattr", "=", "re", ".", "compile", "(", "'^\\\\s*#\\\\s*sphinx_gallery_thumbnail_number\\\\s*=\\\\s*([0-9]+)\\\\s*$'", ",", "flags", "=", "re", ".", "MULTILINE", ")", "match", "=", "pattr", ".", "search", "(", "text", ")", "if", "(", "match", "is", "None", ")", ":", "thumbnail_number", "=", "1", "else", ":", "thumbnail_number", "=", "int", "(", "match", ".", "groups", "(", ")", "[", "0", "]", ")", "return", "thumbnail_number" ]
pull out the thumbnail image number specified in the docstring .
train
false
50,588
def _ensure_systemd(version): try: version = int(version) except ValueError: raise CommandExecutionError("Invalid version '{0}'".format(version)) try: installed = _sd_version() log.debug('nspawn: detected systemd {0}'.format(installed)) except (IndexError, ValueError): raise CommandExecutionError('nspawn: Unable to get systemd version') if (installed < version): raise CommandExecutionError('This function requires systemd >= {0} (Detected version: {1}).'.format(version, installed))
[ "def", "_ensure_systemd", "(", "version", ")", ":", "try", ":", "version", "=", "int", "(", "version", ")", "except", "ValueError", ":", "raise", "CommandExecutionError", "(", "\"Invalid version '{0}'\"", ".", "format", "(", "version", ")", ")", "try", ":", "installed", "=", "_sd_version", "(", ")", "log", ".", "debug", "(", "'nspawn: detected systemd {0}'", ".", "format", "(", "installed", ")", ")", "except", "(", "IndexError", ",", "ValueError", ")", ":", "raise", "CommandExecutionError", "(", "'nspawn: Unable to get systemd version'", ")", "if", "(", "installed", "<", "version", ")", ":", "raise", "CommandExecutionError", "(", "'This function requires systemd >= {0} (Detected version: {1}).'", ".", "format", "(", "version", ",", "installed", ")", ")" ]
raises an exception if the systemd version is not greater than the passed version .
train
true
50,589
def parse_accept_language_header(string, pattern='([a-zA-Z]{1,8}(-[a-zA-Z0-9]{1,8})?)\\s*(;\\s*q\\s*=\\s*(1|0\\.[0-9]+))?'): res = {} if (not string): return None for match in re.finditer(pattern, string): if (None == match.group(4)): q = 1 else: q = match.group(4) l = match.group(1).replace('-', '_') if (len(l) == 2): l = l.lower() elif (len(l) == 5): l = ((l.split('_')[0].lower() + '_') + l.split('_')[1].upper()) else: l = None if l: res[l] = int((100 * float(q))) return res
[ "def", "parse_accept_language_header", "(", "string", ",", "pattern", "=", "'([a-zA-Z]{1,8}(-[a-zA-Z0-9]{1,8})?)\\\\s*(;\\\\s*q\\\\s*=\\\\s*(1|0\\\\.[0-9]+))?'", ")", ":", "res", "=", "{", "}", "if", "(", "not", "string", ")", ":", "return", "None", "for", "match", "in", "re", ".", "finditer", "(", "pattern", ",", "string", ")", ":", "if", "(", "None", "==", "match", ".", "group", "(", "4", ")", ")", ":", "q", "=", "1", "else", ":", "q", "=", "match", ".", "group", "(", "4", ")", "l", "=", "match", ".", "group", "(", "1", ")", ".", "replace", "(", "'-'", ",", "'_'", ")", "if", "(", "len", "(", "l", ")", "==", "2", ")", ":", "l", "=", "l", ".", "lower", "(", ")", "elif", "(", "len", "(", "l", ")", "==", "5", ")", ":", "l", "=", "(", "(", "l", ".", "split", "(", "'_'", ")", "[", "0", "]", ".", "lower", "(", ")", "+", "'_'", ")", "+", "l", ".", "split", "(", "'_'", ")", "[", "1", "]", ".", "upper", "(", ")", ")", "else", ":", "l", "=", "None", "if", "l", ":", "res", "[", "l", "]", "=", "int", "(", "(", "100", "*", "float", "(", "q", ")", ")", ")", "return", "res" ]
parse a dict from an accept-language header string example input: en-us .
train
false
50,590
@requires_application() def test_reactive_draw(): with TestingCanvas() as c: rpolygon = visuals.RegularPolygon(center=[50, 50, 0.0], radius=20, sides=8, color='yellow', parent=c.scene) rpolygon.center = [70, 40, 0.0] assert_image_approved(c.render(), 'visuals/reactive_regular_polygon1.png') rpolygon.radius = 25 assert_image_approved(c.render(), 'visuals/reactive_regular_polygon2.png') rpolygon.color = 'red' assert_image_approved(c.render(), 'visuals/reactive_regular_polygon3.png') rpolygon.border_color = 'yellow' assert_image_approved(c.render(), 'visuals/reactive_regular_polygon4.png') rpolygon.sides = 6 assert_image_approved(c.render(), 'visuals/reactive_regular_polygon5.png')
[ "@", "requires_application", "(", ")", "def", "test_reactive_draw", "(", ")", ":", "with", "TestingCanvas", "(", ")", "as", "c", ":", "rpolygon", "=", "visuals", ".", "RegularPolygon", "(", "center", "=", "[", "50", ",", "50", ",", "0.0", "]", ",", "radius", "=", "20", ",", "sides", "=", "8", ",", "color", "=", "'yellow'", ",", "parent", "=", "c", ".", "scene", ")", "rpolygon", ".", "center", "=", "[", "70", ",", "40", ",", "0.0", "]", "assert_image_approved", "(", "c", ".", "render", "(", ")", ",", "'visuals/reactive_regular_polygon1.png'", ")", "rpolygon", ".", "radius", "=", "25", "assert_image_approved", "(", "c", ".", "render", "(", ")", ",", "'visuals/reactive_regular_polygon2.png'", ")", "rpolygon", ".", "color", "=", "'red'", "assert_image_approved", "(", "c", ".", "render", "(", ")", ",", "'visuals/reactive_regular_polygon3.png'", ")", "rpolygon", ".", "border_color", "=", "'yellow'", "assert_image_approved", "(", "c", ".", "render", "(", ")", ",", "'visuals/reactive_regular_polygon4.png'", ")", "rpolygon", ".", "sides", "=", "6", "assert_image_approved", "(", "c", ".", "render", "(", ")", ",", "'visuals/reactive_regular_polygon5.png'", ")" ]
test reactive regular polygon attributes .
train
false
50,591
def update_sorcery(module): changed = False if module.check_mode: if ((not module.params['name']) and (not module.params['update_cache'])): module.exit_json(changed=True, msg='would have updated Sorcery') else: sorcery_ver = get_sorcery_ver(module) cmd_sorcery = ('%s update' % SORCERY['sorcery']) (rc, stdout, stderr) = module.run_command(cmd_sorcery) if (rc != 0): module.fail_json(msg=('unable to update Sorcery: ' + stdout)) if (sorcery_ver != get_sorcery_ver(module)): changed = True if ((not module.params['name']) and (not module.params['update_cache'])): module.exit_json(changed=changed, msg='successfully updated Sorcery')
[ "def", "update_sorcery", "(", "module", ")", ":", "changed", "=", "False", "if", "module", ".", "check_mode", ":", "if", "(", "(", "not", "module", ".", "params", "[", "'name'", "]", ")", "and", "(", "not", "module", ".", "params", "[", "'update_cache'", "]", ")", ")", ":", "module", ".", "exit_json", "(", "changed", "=", "True", ",", "msg", "=", "'would have updated Sorcery'", ")", "else", ":", "sorcery_ver", "=", "get_sorcery_ver", "(", "module", ")", "cmd_sorcery", "=", "(", "'%s update'", "%", "SORCERY", "[", "'sorcery'", "]", ")", "(", "rc", ",", "stdout", ",", "stderr", ")", "=", "module", ".", "run_command", "(", "cmd_sorcery", ")", "if", "(", "rc", "!=", "0", ")", ":", "module", ".", "fail_json", "(", "msg", "=", "(", "'unable to update Sorcery: '", "+", "stdout", ")", ")", "if", "(", "sorcery_ver", "!=", "get_sorcery_ver", "(", "module", ")", ")", ":", "changed", "=", "True", "if", "(", "(", "not", "module", ".", "params", "[", "'name'", "]", ")", "and", "(", "not", "module", ".", "params", "[", "'update_cache'", "]", ")", ")", ":", "module", ".", "exit_json", "(", "changed", "=", "changed", ",", "msg", "=", "'successfully updated Sorcery'", ")" ]
update sorcery scripts .
train
false
50,592
def recreate_token_data(context, token_data=None, expires=None, user_ref=None, project_ref=None): new_expires = expires project_id = None user_id = None domain_id = None methods = ['password', 'token'] extras = {} if token_data: token = token_data['token'] domain_id = (token['domain']['id'] if ('domain' in token) else None) project_id = (token['project']['id'] if ('project' in token) else None) if (not new_expires): new_expires = token.get('expires_at', token.get('expires')) user_id = token['user']['id'] methods = token['methods'] extras = token['extras'] else: token = None project_id = (project_ref['id'] if project_ref else None) user_id = user_ref['id'] token_data_helper = TokenDataHelper(context) return token_data_helper.get_token_data(user_id, methods, extras, domain_id, project_id, new_expires, token=token)
[ "def", "recreate_token_data", "(", "context", ",", "token_data", "=", "None", ",", "expires", "=", "None", ",", "user_ref", "=", "None", ",", "project_ref", "=", "None", ")", ":", "new_expires", "=", "expires", "project_id", "=", "None", "user_id", "=", "None", "domain_id", "=", "None", "methods", "=", "[", "'password'", ",", "'token'", "]", "extras", "=", "{", "}", "if", "token_data", ":", "token", "=", "token_data", "[", "'token'", "]", "domain_id", "=", "(", "token", "[", "'domain'", "]", "[", "'id'", "]", "if", "(", "'domain'", "in", "token", ")", "else", "None", ")", "project_id", "=", "(", "token", "[", "'project'", "]", "[", "'id'", "]", "if", "(", "'project'", "in", "token", ")", "else", "None", ")", "if", "(", "not", "new_expires", ")", ":", "new_expires", "=", "token", ".", "get", "(", "'expires_at'", ",", "token", ".", "get", "(", "'expires'", ")", ")", "user_id", "=", "token", "[", "'user'", "]", "[", "'id'", "]", "methods", "=", "token", "[", "'methods'", "]", "extras", "=", "token", "[", "'extras'", "]", "else", ":", "token", "=", "None", "project_id", "=", "(", "project_ref", "[", "'id'", "]", "if", "project_ref", "else", "None", ")", "user_id", "=", "user_ref", "[", "'id'", "]", "token_data_helper", "=", "TokenDataHelper", "(", "context", ")", "return", "token_data_helper", ".", "get_token_data", "(", "user_id", ",", "methods", ",", "extras", ",", "domain_id", ",", "project_id", ",", "new_expires", ",", "token", "=", "token", ")" ]
recreate token from an existing token .
train
false
50,593
def xmlencode(s): return _encre.sub((lambda m: ('&#%d;' % ord(m.group(1)))), s)
[ "def", "xmlencode", "(", "s", ")", ":", "return", "_encre", ".", "sub", "(", "(", "lambda", "m", ":", "(", "'&#%d;'", "%", "ord", "(", "m", ".", "group", "(", "1", ")", ")", ")", ")", ",", "s", ")" ]
encode the given string for inclusion in a utf-8 xml document .
train
false
50,595
def test_source(qtbot, py_proc, runner): (cmd, args) = py_proc("\n import os\n import json\n\n data = {\n 'html_file': os.environ['QUTE_HTML'],\n 'text_file': os.environ['QUTE_TEXT'],\n }\n\n with open(os.environ['QUTE_HTML'], 'r') as f:\n data['html'] = f.read()\n\n with open(os.environ['QUTE_TEXT'], 'r') as f:\n data['text'] = f.read()\n\n with open(os.environ['QUTE_FIFO'], 'w') as f:\n json.dump(data, f)\n f.write('\\n')\n ") with qtbot.waitSignal(runner.finished, timeout=10000): with qtbot.waitSignal(runner.got_cmd, timeout=10000) as blocker: runner.prepare_run(cmd, *args) runner.store_html('This is HTML') runner.store_text('This is text') data = blocker.args[0] parsed = json.loads(data) assert (parsed['text'] == 'This is text') assert (parsed['html'] == 'This is HTML') assert (not os.path.exists(parsed['text_file'])) assert (not os.path.exists(parsed['html_file']))
[ "def", "test_source", "(", "qtbot", ",", "py_proc", ",", "runner", ")", ":", "(", "cmd", ",", "args", ")", "=", "py_proc", "(", "\"\\n import os\\n import json\\n\\n data = {\\n 'html_file': os.environ['QUTE_HTML'],\\n 'text_file': os.environ['QUTE_TEXT'],\\n }\\n\\n with open(os.environ['QUTE_HTML'], 'r') as f:\\n data['html'] = f.read()\\n\\n with open(os.environ['QUTE_TEXT'], 'r') as f:\\n data['text'] = f.read()\\n\\n with open(os.environ['QUTE_FIFO'], 'w') as f:\\n json.dump(data, f)\\n f.write('\\\\n')\\n \"", ")", "with", "qtbot", ".", "waitSignal", "(", "runner", ".", "finished", ",", "timeout", "=", "10000", ")", ":", "with", "qtbot", ".", "waitSignal", "(", "runner", ".", "got_cmd", ",", "timeout", "=", "10000", ")", "as", "blocker", ":", "runner", ".", "prepare_run", "(", "cmd", ",", "*", "args", ")", "runner", ".", "store_html", "(", "'This is HTML'", ")", "runner", ".", "store_text", "(", "'This is text'", ")", "data", "=", "blocker", ".", "args", "[", "0", "]", "parsed", "=", "json", ".", "loads", "(", "data", ")", "assert", "(", "parsed", "[", "'text'", "]", "==", "'This is text'", ")", "assert", "(", "parsed", "[", "'html'", "]", "==", "'This is HTML'", ")", "assert", "(", "not", "os", ".", "path", ".", "exists", "(", "parsed", "[", "'text_file'", "]", ")", ")", "assert", "(", "not", "os", ".", "path", ".", "exists", "(", "parsed", "[", "'html_file'", "]", ")", ")" ]
make sure the page source is read and cleaned up correctly .
train
false
50,596
def assertIsSubdomainOf(testCase, descendant, ancestor): testCase.assertTrue(dns._isSubdomainOf(descendant, ancestor), ('%r is not a subdomain of %r' % (descendant, ancestor)))
[ "def", "assertIsSubdomainOf", "(", "testCase", ",", "descendant", ",", "ancestor", ")", ":", "testCase", ".", "assertTrue", "(", "dns", ".", "_isSubdomainOf", "(", "descendant", ",", "ancestor", ")", ",", "(", "'%r is not a subdomain of %r'", "%", "(", "descendant", ",", "ancestor", ")", ")", ")" ]
assert that c{descendant} *is* a subdomain of c{ancestor} .
train
false
50,597
def __int(value): (valid, _value) = (False, value) try: _value = int(value) valid = True except ValueError: pass return (valid, _value, 'integer')
[ "def", "__int", "(", "value", ")", ":", "(", "valid", ",", "_value", ")", "=", "(", "False", ",", "value", ")", "try", ":", "_value", "=", "int", "(", "value", ")", "valid", "=", "True", "except", "ValueError", ":", "pass", "return", "(", "valid", ",", "_value", ",", "'integer'", ")" ]
validate an integer .
train
true
50,598
def _get_hook(): hook = (__salt__['config.get']('mattermost.hook') or __salt__['config.get']('mattermost:hook')) if (not hook): raise SaltInvocationError('No Mattermost Hook found') return hook
[ "def", "_get_hook", "(", ")", ":", "hook", "=", "(", "__salt__", "[", "'config.get'", "]", "(", "'mattermost.hook'", ")", "or", "__salt__", "[", "'config.get'", "]", "(", "'mattermost:hook'", ")", ")", "if", "(", "not", "hook", ")", ":", "raise", "SaltInvocationError", "(", "'No Mattermost Hook found'", ")", "return", "hook" ]
retrieves and return the mattermosts configured hook :return: string: the hook string .
train
false
50,600
def vm_detach_nic(name, kwargs=None, call=None): if (call != 'action'): raise SaltCloudSystemExit('The vm_detach_nic action must be called with -a or --action.') if (kwargs is None): kwargs = {} nic_id = kwargs.get('nic_id', None) if (nic_id is None): raise SaltCloudSystemExit("The vm_detach_nic function requires a 'nic_id' to be provided.") (server, user, password) = _get_xml_rpc() auth = ':'.join([user, password]) vm_id = int(get_vm_id(kwargs={'name': name})) response = server.one.vm.detachnic(auth, vm_id, int(nic_id)) data = {'action': 'vm.detachnic', 'nic_detached': response[0], 'vm_id': response[1], 'error_code': response[2]} return data
[ "def", "vm_detach_nic", "(", "name", ",", "kwargs", "=", "None", ",", "call", "=", "None", ")", ":", "if", "(", "call", "!=", "'action'", ")", ":", "raise", "SaltCloudSystemExit", "(", "'The vm_detach_nic action must be called with -a or --action.'", ")", "if", "(", "kwargs", "is", "None", ")", ":", "kwargs", "=", "{", "}", "nic_id", "=", "kwargs", ".", "get", "(", "'nic_id'", ",", "None", ")", "if", "(", "nic_id", "is", "None", ")", ":", "raise", "SaltCloudSystemExit", "(", "\"The vm_detach_nic function requires a 'nic_id' to be provided.\"", ")", "(", "server", ",", "user", ",", "password", ")", "=", "_get_xml_rpc", "(", ")", "auth", "=", "':'", ".", "join", "(", "[", "user", ",", "password", "]", ")", "vm_id", "=", "int", "(", "get_vm_id", "(", "kwargs", "=", "{", "'name'", ":", "name", "}", ")", ")", "response", "=", "server", ".", "one", ".", "vm", ".", "detachnic", "(", "auth", ",", "vm_id", ",", "int", "(", "nic_id", ")", ")", "data", "=", "{", "'action'", ":", "'vm.detachnic'", ",", "'nic_detached'", ":", "response", "[", "0", "]", ",", "'vm_id'", ":", "response", "[", "1", "]", ",", "'error_code'", ":", "response", "[", "2", "]", "}", "return", "data" ]
detaches a disk from a virtual machine .
train
true
50,603
def unregister_unpack_format(name): del _UNPACK_FORMATS[name]
[ "def", "unregister_unpack_format", "(", "name", ")", ":", "del", "_UNPACK_FORMATS", "[", "name", "]" ]
removes the pack format from the registery .
train
false
50,604
def no_tvtk(): global _have_tvtk return (not _have_tvtk)
[ "def", "no_tvtk", "(", ")", ":", "global", "_have_tvtk", "return", "(", "not", "_have_tvtk", ")" ]
checks if tvtk was found .
train
false
50,605
def print_test(method): def run(*args, **kw): ts = time.time() print (' DCTB testing function %r' % method.__name__) method(*args, **kw) te = time.time() print (' DCTB [OK] in %r %2.2f sec' % (method.__name__, (te - ts))) return run
[ "def", "print_test", "(", "method", ")", ":", "def", "run", "(", "*", "args", ",", "**", "kw", ")", ":", "ts", "=", "time", ".", "time", "(", ")", "print", "(", "' DCTB testing function %r'", "%", "method", ".", "__name__", ")", "method", "(", "*", "args", ",", "**", "kw", ")", "te", "=", "time", ".", "time", "(", ")", "print", "(", "' DCTB [OK] in %r %2.2f sec'", "%", "(", "method", ".", "__name__", ",", "(", "te", "-", "ts", ")", ")", ")", "return", "run" ]
utility method for print verbalizing test suite .
train
false
50,606
@decorator.decorator def apply_to_audio(f, clip, *a, **k): newclip = f(clip, *a, **k) if (hasattr(newclip, 'audio') and (newclip.audio is not None)): newclip.audio = f(newclip.audio, *a, **k) return newclip
[ "@", "decorator", ".", "decorator", "def", "apply_to_audio", "(", "f", ",", "clip", ",", "*", "a", ",", "**", "k", ")", ":", "newclip", "=", "f", "(", "clip", ",", "*", "a", ",", "**", "k", ")", "if", "(", "hasattr", "(", "newclip", ",", "'audio'", ")", "and", "(", "newclip", ".", "audio", "is", "not", "None", ")", ")", ":", "newclip", ".", "audio", "=", "f", "(", "newclip", ".", "audio", ",", "*", "a", ",", "**", "k", ")", "return", "newclip" ]
this decorator will apply the function f to the audio of the clip created with f .
train
false
50,610
def get_master_ip(host=None, port=None, password=None): server = _connect(host, port, password) srv_info = server.info() ret = (srv_info.get('master_host', ''), srv_info.get('master_port', '')) return dict(list(zip(('master_host', 'master_port'), ret)))
[ "def", "get_master_ip", "(", "host", "=", "None", ",", "port", "=", "None", ",", "password", "=", "None", ")", ":", "server", "=", "_connect", "(", "host", ",", "port", ",", "password", ")", "srv_info", "=", "server", ".", "info", "(", ")", "ret", "=", "(", "srv_info", ".", "get", "(", "'master_host'", ",", "''", ")", ",", "srv_info", ".", "get", "(", "'master_port'", ",", "''", ")", ")", "return", "dict", "(", "list", "(", "zip", "(", "(", "'master_host'", ",", "'master_port'", ")", ",", "ret", ")", ")", ")" ]
get host information about slave .
train
true
50,612
def rotate_encryption_key(bucket_name, blob_name, base64_encryption_key, base64_new_encryption_key): storage_client = storage.Client() bucket = storage_client.get_bucket(bucket_name) current_encryption_key = base64.b64decode(base64_encryption_key) new_encryption_key = base64.b64decode(base64_new_encryption_key) source_blob = Blob(blob_name, bucket, encryption_key=current_encryption_key) destination_blob = Blob(blob_name, bucket, encryption_key=new_encryption_key) token = None while True: (token, bytes_rewritten, total_bytes) = destination_blob.rewrite(source_blob, token=token) if (token is None): break print 'Key rotation complete for Blob {}'.format(blob_name)
[ "def", "rotate_encryption_key", "(", "bucket_name", ",", "blob_name", ",", "base64_encryption_key", ",", "base64_new_encryption_key", ")", ":", "storage_client", "=", "storage", ".", "Client", "(", ")", "bucket", "=", "storage_client", ".", "get_bucket", "(", "bucket_name", ")", "current_encryption_key", "=", "base64", ".", "b64decode", "(", "base64_encryption_key", ")", "new_encryption_key", "=", "base64", ".", "b64decode", "(", "base64_new_encryption_key", ")", "source_blob", "=", "Blob", "(", "blob_name", ",", "bucket", ",", "encryption_key", "=", "current_encryption_key", ")", "destination_blob", "=", "Blob", "(", "blob_name", ",", "bucket", ",", "encryption_key", "=", "new_encryption_key", ")", "token", "=", "None", "while", "True", ":", "(", "token", ",", "bytes_rewritten", ",", "total_bytes", ")", "=", "destination_blob", ".", "rewrite", "(", "source_blob", ",", "token", "=", "token", ")", "if", "(", "token", "is", "None", ")", ":", "break", "print", "'Key rotation complete for Blob {}'", ".", "format", "(", "blob_name", ")" ]
performs a key rotation by re-writing an encrypted blob with a new encryption key .
train
false
50,613
@csrf_protect def account_register(request, next_url=u'dashboard'): siteconfig = SiteConfiguration.objects.get_current() auth_backends = get_enabled_auth_backends() if (auth_backends[0].supports_registration and siteconfig.get(u'auth_enable_registration')): response = register(request, next_page=reverse(next_url), form_class=RegistrationForm) return response return HttpResponseRedirect(reverse(u'login'))
[ "@", "csrf_protect", "def", "account_register", "(", "request", ",", "next_url", "=", "u'dashboard'", ")", ":", "siteconfig", "=", "SiteConfiguration", ".", "objects", ".", "get_current", "(", ")", "auth_backends", "=", "get_enabled_auth_backends", "(", ")", "if", "(", "auth_backends", "[", "0", "]", ".", "supports_registration", "and", "siteconfig", ".", "get", "(", "u'auth_enable_registration'", ")", ")", ":", "response", "=", "register", "(", "request", ",", "next_page", "=", "reverse", "(", "next_url", ")", ",", "form_class", "=", "RegistrationForm", ")", "return", "response", "return", "HttpResponseRedirect", "(", "reverse", "(", "u'login'", ")", ")" ]
display the appropriate registration page .
train
false
50,614
def jacobi(a, b): if (a == 0): return 0 result = 1 while (a > 1): if (a & 1): if ((((a - 1) * (b - 1)) >> 2) & 1): result = (- result) (a, b) = ((b % a), a) else: if ((((b * b) - 1) >> 3) & 1): result = (- result) a >>= 1 if (a == 0): return 0 return result
[ "def", "jacobi", "(", "a", ",", "b", ")", ":", "if", "(", "a", "==", "0", ")", ":", "return", "0", "result", "=", "1", "while", "(", "a", ">", "1", ")", ":", "if", "(", "a", "&", "1", ")", ":", "if", "(", "(", "(", "(", "a", "-", "1", ")", "*", "(", "b", "-", "1", ")", ")", ">>", "2", ")", "&", "1", ")", ":", "result", "=", "(", "-", "result", ")", "(", "a", ",", "b", ")", "=", "(", "(", "b", "%", "a", ")", ",", "a", ")", "else", ":", "if", "(", "(", "(", "(", "b", "*", "b", ")", "-", "1", ")", ">>", "3", ")", "&", "1", ")", ":", "result", "=", "(", "-", "result", ")", "a", ">>=", "1", "if", "(", "a", "==", "0", ")", ":", "return", "0", "return", "result" ]
calculates the value of the jacobi symbol where both a and b are positive integers .
train
true
50,615
def get_amount(ref_doc, dt): if (dt == u'Sales Order'): grand_total = (flt(ref_doc.grand_total) - flt(ref_doc.advance_paid)) if (dt == u'Sales Invoice'): if (ref_doc.party_account_currency == ref_doc.currency): grand_total = flt(ref_doc.outstanding_amount) else: grand_total = (flt(ref_doc.outstanding_amount) / ref_doc.conversion_rate) if (grand_total > 0): return grand_total else: frappe.throw(_(u'Payment Entry is already created'))
[ "def", "get_amount", "(", "ref_doc", ",", "dt", ")", ":", "if", "(", "dt", "==", "u'Sales Order'", ")", ":", "grand_total", "=", "(", "flt", "(", "ref_doc", ".", "grand_total", ")", "-", "flt", "(", "ref_doc", ".", "advance_paid", ")", ")", "if", "(", "dt", "==", "u'Sales Invoice'", ")", ":", "if", "(", "ref_doc", ".", "party_account_currency", "==", "ref_doc", ".", "currency", ")", ":", "grand_total", "=", "flt", "(", "ref_doc", ".", "outstanding_amount", ")", "else", ":", "grand_total", "=", "(", "flt", "(", "ref_doc", ".", "outstanding_amount", ")", "/", "ref_doc", ".", "conversion_rate", ")", "if", "(", "grand_total", ">", "0", ")", ":", "return", "grand_total", "else", ":", "frappe", ".", "throw", "(", "_", "(", "u'Payment Entry is already created'", ")", ")" ]
get amount based on doctype .
train
false