id_within_dataset
int64
1
55.5k
snippet
stringlengths
19
14.2k
tokens
listlengths
6
1.63k
nl
stringlengths
6
352
split_within_dataset
stringclasses
1 value
is_duplicated
bool
2 classes
42,531
def clear_subscribers(): _SUBSCRIBERS.clear()
[ "def", "clear_subscribers", "(", ")", ":", "_SUBSCRIBERS", ".", "clear", "(", ")" ]
empty subscribers dictionary .
train
false
42,533
def list_pipelines(region=None, key=None, keyid=None, profile=None): client = _get_client(region, key, keyid, profile) r = {} try: paginator = client.get_paginator('list_pipelines') pipelines = [] for page in paginator.paginate(): pipelines += page['pipelineIdList'] r['result'] = pipelines except (botocore.exceptions.BotoCoreError, botocore.exceptions.ClientError) as e: r['error'] = str(e) return r
[ "def", "list_pipelines", "(", "region", "=", "None", ",", "key", "=", "None", ",", "keyid", "=", "None", ",", "profile", "=", "None", ")", ":", "client", "=", "_get_client", "(", "region", ",", "key", ",", "keyid", ",", "profile", ")", "r", "=", "{", "}", "try", ":", "paginator", "=", "client", ".", "get_paginator", "(", "'list_pipelines'", ")", "pipelines", "=", "[", "]", "for", "page", "in", "paginator", ".", "paginate", "(", ")", ":", "pipelines", "+=", "page", "[", "'pipelineIdList'", "]", "r", "[", "'result'", "]", "=", "pipelines", "except", "(", "botocore", ".", "exceptions", ".", "BotoCoreError", ",", "botocore", ".", "exceptions", ".", "ClientError", ")", "as", "e", ":", "r", "[", "'error'", "]", "=", "str", "(", "e", ")", "return", "r" ]
get a list of pipeline ids and names for all pipelines .
train
true
42,534
def gen_diff_header(paths, modes, shas): (old_path, new_path) = paths (old_mode, new_mode) = modes (old_sha, new_sha) = shas (yield (((('diff --git ' + old_path) + ' ') + new_path) + '\n')) if (old_mode != new_mode): if (new_mode is not None): if (old_mode is not None): (yield ('old mode %o\n' % old_mode).encode('ascii')) (yield ('new mode %o\n' % new_mode).encode('ascii')) else: (yield ('deleted mode %o\n' % old_mode).encode('ascii')) (yield ((('index ' + shortid(old_sha)) + '..') + shortid(new_sha))) if (new_mode is not None): (yield (' %o' % new_mode).encode('ascii')) (yield '\n')
[ "def", "gen_diff_header", "(", "paths", ",", "modes", ",", "shas", ")", ":", "(", "old_path", ",", "new_path", ")", "=", "paths", "(", "old_mode", ",", "new_mode", ")", "=", "modes", "(", "old_sha", ",", "new_sha", ")", "=", "shas", "(", "yield", "(", "(", "(", "(", "'diff --git '", "+", "old_path", ")", "+", "' '", ")", "+", "new_path", ")", "+", "'\\n'", ")", ")", "if", "(", "old_mode", "!=", "new_mode", ")", ":", "if", "(", "new_mode", "is", "not", "None", ")", ":", "if", "(", "old_mode", "is", "not", "None", ")", ":", "(", "yield", "(", "'old mode %o\\n'", "%", "old_mode", ")", ".", "encode", "(", "'ascii'", ")", ")", "(", "yield", "(", "'new mode %o\\n'", "%", "new_mode", ")", ".", "encode", "(", "'ascii'", ")", ")", "else", ":", "(", "yield", "(", "'deleted mode %o\\n'", "%", "old_mode", ")", ".", "encode", "(", "'ascii'", ")", ")", "(", "yield", "(", "(", "(", "'index '", "+", "shortid", "(", "old_sha", ")", ")", "+", "'..'", ")", "+", "shortid", "(", "new_sha", ")", ")", ")", "if", "(", "new_mode", "is", "not", "None", ")", ":", "(", "yield", "(", "' %o'", "%", "new_mode", ")", ".", "encode", "(", "'ascii'", ")", ")", "(", "yield", "'\\n'", ")" ]
write a blob diff header .
train
false
42,536
def collaboration_weighted_projected_graph(B, nodes): if B.is_multigraph(): raise nx.NetworkXError('not defined for multigraphs') if B.is_directed(): pred = B.pred G = nx.DiGraph() else: pred = B.adj G = nx.Graph() G.graph.update(B.graph) G.add_nodes_from(((n, B.node[n]) for n in nodes)) for u in nodes: unbrs = set(B[u]) nbrs2 = (set((n for nbr in unbrs for n in B[nbr])) - set([u])) for v in nbrs2: vnbrs = set(pred[v]) common = (unbrs & vnbrs) weight = sum([(1.0 / (len(B[n]) - 1)) for n in common if (len(B[n]) > 1)]) G.add_edge(u, v, weight=weight) return G
[ "def", "collaboration_weighted_projected_graph", "(", "B", ",", "nodes", ")", ":", "if", "B", ".", "is_multigraph", "(", ")", ":", "raise", "nx", ".", "NetworkXError", "(", "'not defined for multigraphs'", ")", "if", "B", ".", "is_directed", "(", ")", ":", "pred", "=", "B", ".", "pred", "G", "=", "nx", ".", "DiGraph", "(", ")", "else", ":", "pred", "=", "B", ".", "adj", "G", "=", "nx", ".", "Graph", "(", ")", "G", ".", "graph", ".", "update", "(", "B", ".", "graph", ")", "G", ".", "add_nodes_from", "(", "(", "(", "n", ",", "B", ".", "node", "[", "n", "]", ")", "for", "n", "in", "nodes", ")", ")", "for", "u", "in", "nodes", ":", "unbrs", "=", "set", "(", "B", "[", "u", "]", ")", "nbrs2", "=", "(", "set", "(", "(", "n", "for", "nbr", "in", "unbrs", "for", "n", "in", "B", "[", "nbr", "]", ")", ")", "-", "set", "(", "[", "u", "]", ")", ")", "for", "v", "in", "nbrs2", ":", "vnbrs", "=", "set", "(", "pred", "[", "v", "]", ")", "common", "=", "(", "unbrs", "&", "vnbrs", ")", "weight", "=", "sum", "(", "[", "(", "1.0", "/", "(", "len", "(", "B", "[", "n", "]", ")", "-", "1", ")", ")", "for", "n", "in", "common", "if", "(", "len", "(", "B", "[", "n", "]", ")", ">", "1", ")", "]", ")", "G", ".", "add_edge", "(", "u", ",", "v", ",", "weight", "=", "weight", ")", "return", "G" ]
newmans weighted projection of b onto one of its node sets .
train
false
42,539
def _sort_fields(elem, orders_dicts): try: order = orders_dicts[elem.tag] except KeyError: pass else: tmp = sorted([((order.get(child.tag, 1000000000.0), i), child) for (i, child) in enumerate(elem)]) elem[:] = [child for (key, child) in tmp] for child in elem: if len(child): _sort_fields(child, orders_dicts)
[ "def", "_sort_fields", "(", "elem", ",", "orders_dicts", ")", ":", "try", ":", "order", "=", "orders_dicts", "[", "elem", ".", "tag", "]", "except", "KeyError", ":", "pass", "else", ":", "tmp", "=", "sorted", "(", "[", "(", "(", "order", ".", "get", "(", "child", ".", "tag", ",", "1000000000.0", ")", ",", "i", ")", ",", "child", ")", "for", "(", "i", ",", "child", ")", "in", "enumerate", "(", "elem", ")", "]", ")", "elem", "[", ":", "]", "=", "[", "child", "for", "(", "key", ",", "child", ")", "in", "tmp", "]", "for", "child", "in", "elem", ":", "if", "len", "(", "child", ")", ":", "_sort_fields", "(", "child", ",", "orders_dicts", ")" ]
sort the children of elem .
train
false
42,541
@LocalProxy def _proxy_stream(): ctx = _request_ctx_stack.top if (ctx is not None): return ctx.request.environ['wsgi.errors'] return sys.stderr
[ "@", "LocalProxy", "def", "_proxy_stream", "(", ")", ":", "ctx", "=", "_request_ctx_stack", ".", "top", "if", "(", "ctx", "is", "not", "None", ")", ":", "return", "ctx", ".", "request", ".", "environ", "[", "'wsgi.errors'", "]", "return", "sys", ".", "stderr" ]
finds the most appropriate error stream for the application .
train
false
42,542
def get_project_from_domain(domain_ref): project_ref = domain_ref.copy() project_ref['is_domain'] = True project_ref['domain_id'] = None project_ref['parent_id'] = None return project_ref
[ "def", "get_project_from_domain", "(", "domain_ref", ")", ":", "project_ref", "=", "domain_ref", ".", "copy", "(", ")", "project_ref", "[", "'is_domain'", "]", "=", "True", "project_ref", "[", "'domain_id'", "]", "=", "None", "project_ref", "[", "'parent_id'", "]", "=", "None", "return", "project_ref" ]
create a project ref from the provided domain ref .
train
false
42,543
def getRow(): try: db = mysql.connect(host=mysqlHost, user=mysqlUser, passwd=mysqlPasswd) cursor = db.cursor() cursor.execute("SHOW STATUS LIKE '%wsrep%'") result = cursor.fetchall() except: print 'Error: unable to fetch data - Check your configuration!' sys.exit(13) db.close() return result
[ "def", "getRow", "(", ")", ":", "try", ":", "db", "=", "mysql", ".", "connect", "(", "host", "=", "mysqlHost", ",", "user", "=", "mysqlUser", ",", "passwd", "=", "mysqlPasswd", ")", "cursor", "=", "db", ".", "cursor", "(", ")", "cursor", ".", "execute", "(", "\"SHOW STATUS LIKE '%wsrep%'\"", ")", "result", "=", "cursor", ".", "fetchall", "(", ")", "except", ":", "print", "'Error: unable to fetch data - Check your configuration!'", "sys", ".", "exit", "(", "13", ")", "db", ".", "close", "(", ")", "return", "result" ]
test connection .
train
false
42,545
def test_adjust_gamma_zero(): image = np.random.uniform(0, 255, (8, 8)) result = exposure.adjust_gamma(image, 0) dtype = image.dtype.type assert_array_equal(result, dtype_range[dtype][1])
[ "def", "test_adjust_gamma_zero", "(", ")", ":", "image", "=", "np", ".", "random", ".", "uniform", "(", "0", ",", "255", ",", "(", "8", ",", "8", ")", ")", "result", "=", "exposure", ".", "adjust_gamma", "(", "image", ",", "0", ")", "dtype", "=", "image", ".", "dtype", ".", "type", "assert_array_equal", "(", "result", ",", "dtype_range", "[", "dtype", "]", "[", "1", "]", ")" ]
white image should be returned for gamma equal to zero .
train
false
42,546
def s3_ownerstamp(): auth = current.auth utable = auth.settings.table_user s3_meta_owned_by_user = S3ReusableField('owned_by_user', utable, readable=False, writable=False, requires=None, default=(current.session.auth.user.id if auth.is_logged_in() else None), represent=(lambda id: ((id and s3_auth_user_represent(id)) or current.messages.UNKNOWN_OPT)), ondelete='RESTRICT') s3_meta_owned_by_group = S3ReusableField('owned_by_group', 'integer', readable=False, writable=False, requires=None, default=None, represent=S3Represent(lookup='auth_group', fields=['role'])) s3_meta_realm_entity = S3ReusableField('realm_entity', 'integer', default=None, readable=False, writable=False, requires=None, represent=(lambda val: current.s3db.pr_pentity_represent(val))) return (s3_meta_owned_by_user(), s3_meta_owned_by_group(), s3_meta_realm_entity())
[ "def", "s3_ownerstamp", "(", ")", ":", "auth", "=", "current", ".", "auth", "utable", "=", "auth", ".", "settings", ".", "table_user", "s3_meta_owned_by_user", "=", "S3ReusableField", "(", "'owned_by_user'", ",", "utable", ",", "readable", "=", "False", ",", "writable", "=", "False", ",", "requires", "=", "None", ",", "default", "=", "(", "current", ".", "session", ".", "auth", ".", "user", ".", "id", "if", "auth", ".", "is_logged_in", "(", ")", "else", "None", ")", ",", "represent", "=", "(", "lambda", "id", ":", "(", "(", "id", "and", "s3_auth_user_represent", "(", "id", ")", ")", "or", "current", ".", "messages", ".", "UNKNOWN_OPT", ")", ")", ",", "ondelete", "=", "'RESTRICT'", ")", "s3_meta_owned_by_group", "=", "S3ReusableField", "(", "'owned_by_group'", ",", "'integer'", ",", "readable", "=", "False", ",", "writable", "=", "False", ",", "requires", "=", "None", ",", "default", "=", "None", ",", "represent", "=", "S3Represent", "(", "lookup", "=", "'auth_group'", ",", "fields", "=", "[", "'role'", "]", ")", ")", "s3_meta_realm_entity", "=", "S3ReusableField", "(", "'realm_entity'", ",", "'integer'", ",", "default", "=", "None", ",", "readable", "=", "False", ",", "writable", "=", "False", ",", "requires", "=", "None", ",", "represent", "=", "(", "lambda", "val", ":", "current", ".", "s3db", ".", "pr_pentity_represent", "(", "val", ")", ")", ")", "return", "(", "s3_meta_owned_by_user", "(", ")", ",", "s3_meta_owned_by_group", "(", ")", ",", "s3_meta_realm_entity", "(", ")", ")" ]
record ownership meta-fields .
train
false
42,548
def getAroundLoop(begin, end, loop): aroundLoop = [] if (end <= begin): end += len(loop) for pointIndex in xrange(begin, end): aroundLoop.append(loop[(pointIndex % len(loop))]) return aroundLoop
[ "def", "getAroundLoop", "(", "begin", ",", "end", ",", "loop", ")", ":", "aroundLoop", "=", "[", "]", "if", "(", "end", "<=", "begin", ")", ":", "end", "+=", "len", "(", "loop", ")", "for", "pointIndex", "in", "xrange", "(", "begin", ",", "end", ")", ":", "aroundLoop", ".", "append", "(", "loop", "[", "(", "pointIndex", "%", "len", "(", "loop", ")", ")", "]", ")", "return", "aroundLoop" ]
get an arc around a loop .
train
false
42,549
def write_pid_to_pidfile(pidfile_path): open_flags = ((os.O_CREAT | os.O_EXCL) | os.O_WRONLY) open_mode = 420 pidfile_fd = os.open(pidfile_path, open_flags, open_mode) pidfile = os.fdopen(pidfile_fd, 'w') pid = os.getpid() pidfile.write(('%s\n' % pid)) pidfile.close()
[ "def", "write_pid_to_pidfile", "(", "pidfile_path", ")", ":", "open_flags", "=", "(", "(", "os", ".", "O_CREAT", "|", "os", ".", "O_EXCL", ")", "|", "os", ".", "O_WRONLY", ")", "open_mode", "=", "420", "pidfile_fd", "=", "os", ".", "open", "(", "pidfile_path", ",", "open_flags", ",", "open_mode", ")", "pidfile", "=", "os", ".", "fdopen", "(", "pidfile_fd", ",", "'w'", ")", "pid", "=", "os", ".", "getpid", "(", ")", "pidfile", ".", "write", "(", "(", "'%s\\n'", "%", "pid", ")", ")", "pidfile", ".", "close", "(", ")" ]
write the pid in the named pid file .
train
true
42,550
def find_filter_class(filtername): if (filtername in FILTERS): return FILTERS[filtername] for (name, cls) in find_plugin_filters(): if (name == filtername): return cls return None
[ "def", "find_filter_class", "(", "filtername", ")", ":", "if", "(", "filtername", "in", "FILTERS", ")", ":", "return", "FILTERS", "[", "filtername", "]", "for", "(", "name", ",", "cls", ")", "in", "find_plugin_filters", "(", ")", ":", "if", "(", "name", "==", "filtername", ")", ":", "return", "cls", "return", "None" ]
lookup a filter by name .
train
true
42,551
def has_tag(node, kind): for d in node['directory']: if (d.kind == kind): return True return False
[ "def", "has_tag", "(", "node", ",", "kind", ")", ":", "for", "d", "in", "node", "[", "'directory'", "]", ":", "if", "(", "d", ".", "kind", "==", "kind", ")", ":", "return", "True", "return", "False" ]
check if the node contains a tag of a given kind .
train
false
42,552
def setup_client(client_secret_file, server_public_file, endpoint='tcp://127.0.0.1:9000'): client = zmq.Context.instance().socket(zmq.DEALER) (client_public, client_secret) = zmq.auth.load_certificate(client_secret_file) client.curve_secretkey = client_secret client.curve_publickey = client_public (server_public, _) = zmq.auth.load_certificate(server_public_file) client.curve_serverkey = server_public client.connect(endpoint) client_stream = zmqstream.ZMQStream(client) client_stream.on_recv(client_msg_recvd) return client_stream
[ "def", "setup_client", "(", "client_secret_file", ",", "server_public_file", ",", "endpoint", "=", "'tcp://127.0.0.1:9000'", ")", ":", "client", "=", "zmq", ".", "Context", ".", "instance", "(", ")", ".", "socket", "(", "zmq", ".", "DEALER", ")", "(", "client_public", ",", "client_secret", ")", "=", "zmq", ".", "auth", ".", "load_certificate", "(", "client_secret_file", ")", "client", ".", "curve_secretkey", "=", "client_secret", "client", ".", "curve_publickey", "=", "client_public", "(", "server_public", ",", "_", ")", "=", "zmq", ".", "auth", ".", "load_certificate", "(", "server_public_file", ")", "client", ".", "curve_serverkey", "=", "server_public", "client", ".", "connect", "(", "endpoint", ")", "client_stream", "=", "zmqstream", ".", "ZMQStream", "(", "client", ")", "client_stream", ".", "on_recv", "(", "client_msg_recvd", ")", "return", "client_stream" ]
setup a simple client with curve auth .
train
false
42,556
def doNotFailOnNetworkError(func): @functools.wraps(func) def wrapper(*a, **kw): try: func(*a, **kw) except FailTest as e: if e.args[0].startswith("'Failed to get object inventory from "): raise SkipTest('This test is prone to intermittent network errors. See ticket 8753. Exception was: {!r}'.format(e)) raise return wrapper
[ "def", "doNotFailOnNetworkError", "(", "func", ")", ":", "@", "functools", ".", "wraps", "(", "func", ")", "def", "wrapper", "(", "*", "a", ",", "**", "kw", ")", ":", "try", ":", "func", "(", "*", "a", ",", "**", "kw", ")", "except", "FailTest", "as", "e", ":", "if", "e", ".", "args", "[", "0", "]", ".", "startswith", "(", "\"'Failed to get object inventory from \"", ")", ":", "raise", "SkipTest", "(", "'This test is prone to intermittent network errors. See ticket 8753. Exception was: {!r}'", ".", "format", "(", "e", ")", ")", "raise", "return", "wrapper" ]
a decorator which makes apibuilder tests not fail because of intermittent network failures -- mamely .
train
false
42,558
def get_all_stats(): all_stats = {} secret = appscale_info.get_secret() logging.debug('Retrieved deployment secret: {}'.format(secret)) for ip in appscale_info.get_all_ips(): appcontroller_endpoint = 'https://{}:{}'.format(ip, hermes_constants.APPCONTROLLER_PORT) logging.debug('Connecting to AC at: {}'.format(appcontroller_endpoint)) server = SOAPpy.SOAPProxy(appcontroller_endpoint) try: all_stats[ip] = json.loads(server.get_all_stats(secret)) except SOAPpy.SOAPException as soap_exception: logging.exception('Exception while performing SOAP call to {}'.format(appcontroller_endpoint)) logging.exception(soap_exception) all_stats[ip] = {JSONTags.ERROR: JSONTags.UNREACHABLE} except socket_error as serr: logging.error('Socket error while performing SOAP call to {}'.format(appcontroller_endpoint)) logging.error(serr) all_stats[ip] = {JSONTags.ERROR: JSONTags.UNREACHABLE} return all_stats
[ "def", "get_all_stats", "(", ")", ":", "all_stats", "=", "{", "}", "secret", "=", "appscale_info", ".", "get_secret", "(", ")", "logging", ".", "debug", "(", "'Retrieved deployment secret: {}'", ".", "format", "(", "secret", ")", ")", "for", "ip", "in", "appscale_info", ".", "get_all_ips", "(", ")", ":", "appcontroller_endpoint", "=", "'https://{}:{}'", ".", "format", "(", "ip", ",", "hermes_constants", ".", "APPCONTROLLER_PORT", ")", "logging", ".", "debug", "(", "'Connecting to AC at: {}'", ".", "format", "(", "appcontroller_endpoint", ")", ")", "server", "=", "SOAPpy", ".", "SOAPProxy", "(", "appcontroller_endpoint", ")", "try", ":", "all_stats", "[", "ip", "]", "=", "json", ".", "loads", "(", "server", ".", "get_all_stats", "(", "secret", ")", ")", "except", "SOAPpy", ".", "SOAPException", "as", "soap_exception", ":", "logging", ".", "exception", "(", "'Exception while performing SOAP call to {}'", ".", "format", "(", "appcontroller_endpoint", ")", ")", "logging", ".", "exception", "(", "soap_exception", ")", "all_stats", "[", "ip", "]", "=", "{", "JSONTags", ".", "ERROR", ":", "JSONTags", ".", "UNREACHABLE", "}", "except", "socket_error", "as", "serr", ":", "logging", ".", "error", "(", "'Socket error while performing SOAP call to {}'", ".", "format", "(", "appcontroller_endpoint", ")", ")", "logging", ".", "error", "(", "serr", ")", "all_stats", "[", "ip", "]", "=", "{", "JSONTags", ".", "ERROR", ":", "JSONTags", ".", "UNREACHABLE", "}", "return", "all_stats" ]
collects platform stats from all deployment nodes .
train
false
42,559
def expr_eval(expr): c = test_expr(expr, _EXPR_OPCODES) return unsafe_eval(c)
[ "def", "expr_eval", "(", "expr", ")", ":", "c", "=", "test_expr", "(", "expr", ",", "_EXPR_OPCODES", ")", "return", "unsafe_eval", "(", "c", ")" ]
expr_eval -> value restricted python expression evaluation evaluates a string that contains an expression that only uses python constants .
train
false
42,560
@environmentfilter def do_random(environment, seq): try: return choice(seq) except IndexError: return environment.undefined('No random item, sequence was empty.')
[ "@", "environmentfilter", "def", "do_random", "(", "environment", ",", "seq", ")", ":", "try", ":", "return", "choice", "(", "seq", ")", "except", "IndexError", ":", "return", "environment", ".", "undefined", "(", "'No random item, sequence was empty.'", ")" ]
return a random item from the sequence .
train
false
42,561
def resource_data_set(context, resource_id, key, value, redact=False): if redact: (method, value) = crypt.encrypt(value) else: method = '' try: current = resource_data_get_by_key(context, resource_id, key) except exception.NotFound: current = models.ResourceData() current.key = key current.resource_id = resource_id current.redact = redact current.value = value current.decrypt_method = method current.save(session=context.session) return current
[ "def", "resource_data_set", "(", "context", ",", "resource_id", ",", "key", ",", "value", ",", "redact", "=", "False", ")", ":", "if", "redact", ":", "(", "method", ",", "value", ")", "=", "crypt", ".", "encrypt", "(", "value", ")", "else", ":", "method", "=", "''", "try", ":", "current", "=", "resource_data_get_by_key", "(", "context", ",", "resource_id", ",", "key", ")", "except", "exception", ".", "NotFound", ":", "current", "=", "models", ".", "ResourceData", "(", ")", "current", ".", "key", "=", "key", "current", ".", "resource_id", "=", "resource_id", "current", ".", "redact", "=", "redact", "current", ".", "value", "=", "value", "current", ".", "decrypt_method", "=", "method", "current", ".", "save", "(", "session", "=", "context", ".", "session", ")", "return", "current" ]
save resources key/value pair to database .
train
false
42,563
def format_certificate_for_user(username, cert): return {'username': username, 'course_key': cert.course_id, 'type': cert.mode, 'status': cert.status, 'grade': cert.grade, 'created': cert.created_date, 'modified': cert.modified_date, 'download_url': ((cert.download_url or get_certificate_url(cert.user.id, cert.course_id)) if (cert.status == CertificateStatuses.downloadable) else None)}
[ "def", "format_certificate_for_user", "(", "username", ",", "cert", ")", ":", "return", "{", "'username'", ":", "username", ",", "'course_key'", ":", "cert", ".", "course_id", ",", "'type'", ":", "cert", ".", "mode", ",", "'status'", ":", "cert", ".", "status", ",", "'grade'", ":", "cert", ".", "grade", ",", "'created'", ":", "cert", ".", "created_date", ",", "'modified'", ":", "cert", ".", "modified_date", ",", "'download_url'", ":", "(", "(", "cert", ".", "download_url", "or", "get_certificate_url", "(", "cert", ".", "user", ".", "id", ",", "cert", ".", "course_id", ")", ")", "if", "(", "cert", ".", "status", "==", "CertificateStatuses", ".", "downloadable", ")", "else", "None", ")", "}" ]
helper function to serialize an user certificate .
train
false
42,564
def collect_dirs(dir_paths, ignored_dir_paths=None): valid_dirs = list(filter((lambda fname: os.path.isdir(fname[0])), icollect(dir_paths, ignored_dir_paths))) if valid_dirs: (collected_dirs, _) = zip(*valid_dirs) return list(collected_dirs) else: return []
[ "def", "collect_dirs", "(", "dir_paths", ",", "ignored_dir_paths", "=", "None", ")", ":", "valid_dirs", "=", "list", "(", "filter", "(", "(", "lambda", "fname", ":", "os", ".", "path", ".", "isdir", "(", "fname", "[", "0", "]", ")", ")", ",", "icollect", "(", "dir_paths", ",", "ignored_dir_paths", ")", ")", ")", "if", "valid_dirs", ":", "(", "collected_dirs", ",", "_", ")", "=", "zip", "(", "*", "valid_dirs", ")", "return", "list", "(", "collected_dirs", ")", "else", ":", "return", "[", "]" ]
evaluate globs in directory paths and return all matching directories .
train
false
42,566
def authorize_client(client, auth_type=None, service=None, source=None, scopes=None, oauth_type=None, consumer_key=None, consumer_secret=None): settings = SettingsUtil() return settings.authorize_client(client=client, auth_type=auth_type, service=service, source=source, scopes=scopes, oauth_type=oauth_type, consumer_key=consumer_key, consumer_secret=consumer_secret)
[ "def", "authorize_client", "(", "client", ",", "auth_type", "=", "None", ",", "service", "=", "None", ",", "source", "=", "None", ",", "scopes", "=", "None", ",", "oauth_type", "=", "None", ",", "consumer_key", "=", "None", ",", "consumer_secret", "=", "None", ")", ":", "settings", "=", "SettingsUtil", "(", ")", "return", "settings", ".", "authorize_client", "(", "client", "=", "client", ",", "auth_type", "=", "auth_type", ",", "service", "=", "service", ",", "source", "=", "source", ",", "scopes", "=", "scopes", ",", "oauth_type", "=", "oauth_type", ",", "consumer_key", "=", "consumer_key", ",", "consumer_secret", "=", "consumer_secret", ")" ]
uses command line arguments .
train
false
42,567
def _PackForSuds(obj, factory): if (obj in ({}, None)): return suds.null() elif isinstance(obj, dict): if ('xsi_type' in obj): try: new_obj = factory.create(obj['xsi_type']) except suds.TypeNotFound: new_obj = factory.create(':'.join(['ns0', obj['xsi_type']])) for (param, _) in new_obj: if param.endswith('.Type'): setattr(new_obj, param, obj['xsi_type']) else: setattr(new_obj, param, None) for key in obj: if (key == 'xsi_type'): continue setattr(new_obj, key, _PackForSuds(obj[key], factory)) else: new_obj = {} for key in obj: new_obj[key] = _PackForSuds(obj[key], factory) return new_obj elif isinstance(obj, (list, tuple)): return [_PackForSuds(item, factory) for item in obj] else: _RecurseOverObject(obj, factory) return obj
[ "def", "_PackForSuds", "(", "obj", ",", "factory", ")", ":", "if", "(", "obj", "in", "(", "{", "}", ",", "None", ")", ")", ":", "return", "suds", ".", "null", "(", ")", "elif", "isinstance", "(", "obj", ",", "dict", ")", ":", "if", "(", "'xsi_type'", "in", "obj", ")", ":", "try", ":", "new_obj", "=", "factory", ".", "create", "(", "obj", "[", "'xsi_type'", "]", ")", "except", "suds", ".", "TypeNotFound", ":", "new_obj", "=", "factory", ".", "create", "(", "':'", ".", "join", "(", "[", "'ns0'", ",", "obj", "[", "'xsi_type'", "]", "]", ")", ")", "for", "(", "param", ",", "_", ")", "in", "new_obj", ":", "if", "param", ".", "endswith", "(", "'.Type'", ")", ":", "setattr", "(", "new_obj", ",", "param", ",", "obj", "[", "'xsi_type'", "]", ")", "else", ":", "setattr", "(", "new_obj", ",", "param", ",", "None", ")", "for", "key", "in", "obj", ":", "if", "(", "key", "==", "'xsi_type'", ")", ":", "continue", "setattr", "(", "new_obj", ",", "key", ",", "_PackForSuds", "(", "obj", "[", "key", "]", ",", "factory", ")", ")", "else", ":", "new_obj", "=", "{", "}", "for", "key", "in", "obj", ":", "new_obj", "[", "key", "]", "=", "_PackForSuds", "(", "obj", "[", "key", "]", ",", "factory", ")", "return", "new_obj", "elif", "isinstance", "(", "obj", ",", "(", "list", ",", "tuple", ")", ")", ":", "return", "[", "_PackForSuds", "(", "item", ",", "factory", ")", "for", "item", "in", "obj", "]", "else", ":", "_RecurseOverObject", "(", "obj", ",", "factory", ")", "return", "obj" ]
packs soap input into the format we want for suds .
train
true
42,568
def rugplot(a, height=0.05, axis='x', ax=None, **kwargs): if (ax is None): ax = plt.gca() a = np.asarray(a) vertical = kwargs.pop('vertical', (axis == 'y')) func = (ax.axhline if vertical else ax.axvline) kwargs.setdefault('linewidth', 1) for pt in a: func(pt, 0, height, **kwargs) return ax
[ "def", "rugplot", "(", "a", ",", "height", "=", "0.05", ",", "axis", "=", "'x'", ",", "ax", "=", "None", ",", "**", "kwargs", ")", ":", "if", "(", "ax", "is", "None", ")", ":", "ax", "=", "plt", ".", "gca", "(", ")", "a", "=", "np", ".", "asarray", "(", "a", ")", "vertical", "=", "kwargs", ".", "pop", "(", "'vertical'", ",", "(", "axis", "==", "'y'", ")", ")", "func", "=", "(", "ax", ".", "axhline", "if", "vertical", "else", "ax", ".", "axvline", ")", "kwargs", ".", "setdefault", "(", "'linewidth'", ",", "1", ")", "for", "pt", "in", "a", ":", "func", "(", "pt", ",", "0", ",", "height", ",", "**", "kwargs", ")", "return", "ax" ]
plot datapoints in an array as sticks on an axis .
train
false
42,569
def send_schedule_change(email, session_name, link): message_settings = MessageSettings.query.filter_by(action=SESSION_SCHEDULE).first() if ((not message_settings) or (message_settings.mail_status == 1)): send_email(to=email, action=SESSION_SCHEDULE, subject=MAILS[SESSION_SCHEDULE]['subject'].format(session_name=session_name), html=MAILS[SESSION_SCHEDULE]['message'].format(email=str(email), session_name=str(session_name), link=link))
[ "def", "send_schedule_change", "(", "email", ",", "session_name", ",", "link", ")", ":", "message_settings", "=", "MessageSettings", ".", "query", ".", "filter_by", "(", "action", "=", "SESSION_SCHEDULE", ")", ".", "first", "(", ")", "if", "(", "(", "not", "message_settings", ")", "or", "(", "message_settings", ".", "mail_status", "==", "1", ")", ")", ":", "send_email", "(", "to", "=", "email", ",", "action", "=", "SESSION_SCHEDULE", ",", "subject", "=", "MAILS", "[", "SESSION_SCHEDULE", "]", "[", "'subject'", "]", ".", "format", "(", "session_name", "=", "session_name", ")", ",", "html", "=", "MAILS", "[", "SESSION_SCHEDULE", "]", "[", "'message'", "]", ".", "format", "(", "email", "=", "str", "(", "email", ")", ",", "session_name", "=", "str", "(", "session_name", ")", ",", "link", "=", "link", ")", ")" ]
send schedule change in session .
train
false
42,571
def assert_raises_with_msg(exc_class, expected_msg, callable_obj, *args, **kwargs): try: callable_obj(*args, **kwargs) except exc_class as err: assert_equal(expected_msg, unic(err), 'Correct exception but wrong message') else: if hasattr(exc_class, '__name__'): exc_name = exc_class.__name__ else: exc_name = str(exc_class) _report_failure(('%s not raised' % exc_name))
[ "def", "assert_raises_with_msg", "(", "exc_class", ",", "expected_msg", ",", "callable_obj", ",", "*", "args", ",", "**", "kwargs", ")", ":", "try", ":", "callable_obj", "(", "*", "args", ",", "**", "kwargs", ")", "except", "exc_class", "as", "err", ":", "assert_equal", "(", "expected_msg", ",", "unic", "(", "err", ")", ",", "'Correct exception but wrong message'", ")", "else", ":", "if", "hasattr", "(", "exc_class", ",", "'__name__'", ")", ":", "exc_name", "=", "exc_class", ".", "__name__", "else", ":", "exc_name", "=", "str", "(", "exc_class", ")", "_report_failure", "(", "(", "'%s not raised'", "%", "exc_name", ")", ")" ]
similar to fail_unless_raises but also checks the exception message .
train
false
42,572
def slicer(dimensions): def decorator(extractor): @wraps(extractor) def make_slicer(*a, **k): return Slicer(dimensions=dimensions, extractor=partial(extractor, *a, **k)) return make_slicer return decorator
[ "def", "slicer", "(", "dimensions", ")", ":", "def", "decorator", "(", "extractor", ")", ":", "@", "wraps", "(", "extractor", ")", "def", "make_slicer", "(", "*", "a", ",", "**", "k", ")", ":", "return", "Slicer", "(", "dimensions", "=", "dimensions", ",", "extractor", "=", "partial", "(", "extractor", ",", "*", "a", ",", "**", "k", ")", ")", "return", "make_slicer", "return", "decorator" ]
slicer decorator .
train
false
42,573
@deprecated('The functon distribute_covar_matrix_to_match_covariance_typeis deprecated in 0.18 and will be removed in 0.20.') def distribute_covar_matrix_to_match_covariance_type(tied_cv, covariance_type, n_components): if (covariance_type == 'spherical'): cv = np.tile((tied_cv.mean() * np.ones(tied_cv.shape[1])), (n_components, 1)) elif (covariance_type == 'tied'): cv = tied_cv elif (covariance_type == 'diag'): cv = np.tile(np.diag(tied_cv), (n_components, 1)) elif (covariance_type == 'full'): cv = np.tile(tied_cv, (n_components, 1, 1)) else: raise ValueError(('covariance_type must be one of ' + "'spherical', 'tied', 'diag', 'full'")) return cv
[ "@", "deprecated", "(", "'The functon distribute_covar_matrix_to_match_covariance_typeis deprecated in 0.18 and will be removed in 0.20.'", ")", "def", "distribute_covar_matrix_to_match_covariance_type", "(", "tied_cv", ",", "covariance_type", ",", "n_components", ")", ":", "if", "(", "covariance_type", "==", "'spherical'", ")", ":", "cv", "=", "np", ".", "tile", "(", "(", "tied_cv", ".", "mean", "(", ")", "*", "np", ".", "ones", "(", "tied_cv", ".", "shape", "[", "1", "]", ")", ")", ",", "(", "n_components", ",", "1", ")", ")", "elif", "(", "covariance_type", "==", "'tied'", ")", ":", "cv", "=", "tied_cv", "elif", "(", "covariance_type", "==", "'diag'", ")", ":", "cv", "=", "np", ".", "tile", "(", "np", ".", "diag", "(", "tied_cv", ")", ",", "(", "n_components", ",", "1", ")", ")", "elif", "(", "covariance_type", "==", "'full'", ")", ":", "cv", "=", "np", ".", "tile", "(", "tied_cv", ",", "(", "n_components", ",", "1", ",", "1", ")", ")", "else", ":", "raise", "ValueError", "(", "(", "'covariance_type must be one of '", "+", "\"'spherical', 'tied', 'diag', 'full'\"", ")", ")", "return", "cv" ]
create all the covariance matrices from a given template .
train
true
42,574
def skip_if_browser(browser): def decorator(test_function): @functools.wraps(test_function) def wrapper(self, *args, **kwargs): if (self.browser.name == browser): raise unittest.SkipTest('Skipping as this test will not work with {}'.format(browser)) test_function(self, *args, **kwargs) return wrapper return decorator
[ "def", "skip_if_browser", "(", "browser", ")", ":", "def", "decorator", "(", "test_function", ")", ":", "@", "functools", ".", "wraps", "(", "test_function", ")", "def", "wrapper", "(", "self", ",", "*", "args", ",", "**", "kwargs", ")", ":", "if", "(", "self", ".", "browser", ".", "name", "==", "browser", ")", ":", "raise", "unittest", ".", "SkipTest", "(", "'Skipping as this test will not work with {}'", ".", "format", "(", "browser", ")", ")", "test_function", "(", "self", ",", "*", "args", ",", "**", "kwargs", ")", "return", "wrapper", "return", "decorator" ]
method decorator that skips a test if browser is browser args: browser : name of internet browser returns: decorated function .
train
false
42,575
def scorers_processors(): scorers = [fuzz.ratio, fuzz.partial_ratio] processors = [(lambda x: x), partial(utils.full_process, force_ascii=False), partial(utils.full_process, force_ascii=True)] splist = list(product(scorers, processors)) splist.extend([(fuzz.WRatio, partial(utils.full_process, force_ascii=True)), (fuzz.QRatio, partial(utils.full_process, force_ascii=True)), (fuzz.UWRatio, partial(utils.full_process, force_ascii=False)), (fuzz.UQRatio, partial(utils.full_process, force_ascii=False)), (fuzz.token_set_ratio, partial(utils.full_process, force_ascii=True)), (fuzz.token_sort_ratio, partial(utils.full_process, force_ascii=True)), (fuzz.partial_token_set_ratio, partial(utils.full_process, force_ascii=True)), (fuzz.partial_token_sort_ratio, partial(utils.full_process, force_ascii=True))]) return splist
[ "def", "scorers_processors", "(", ")", ":", "scorers", "=", "[", "fuzz", ".", "ratio", ",", "fuzz", ".", "partial_ratio", "]", "processors", "=", "[", "(", "lambda", "x", ":", "x", ")", ",", "partial", "(", "utils", ".", "full_process", ",", "force_ascii", "=", "False", ")", ",", "partial", "(", "utils", ".", "full_process", ",", "force_ascii", "=", "True", ")", "]", "splist", "=", "list", "(", "product", "(", "scorers", ",", "processors", ")", ")", "splist", ".", "extend", "(", "[", "(", "fuzz", ".", "WRatio", ",", "partial", "(", "utils", ".", "full_process", ",", "force_ascii", "=", "True", ")", ")", ",", "(", "fuzz", ".", "QRatio", ",", "partial", "(", "utils", ".", "full_process", ",", "force_ascii", "=", "True", ")", ")", ",", "(", "fuzz", ".", "UWRatio", ",", "partial", "(", "utils", ".", "full_process", ",", "force_ascii", "=", "False", ")", ")", ",", "(", "fuzz", ".", "UQRatio", ",", "partial", "(", "utils", ".", "full_process", ",", "force_ascii", "=", "False", ")", ")", ",", "(", "fuzz", ".", "token_set_ratio", ",", "partial", "(", "utils", ".", "full_process", ",", "force_ascii", "=", "True", ")", ")", ",", "(", "fuzz", ".", "token_sort_ratio", ",", "partial", "(", "utils", ".", "full_process", ",", "force_ascii", "=", "True", ")", ")", ",", "(", "fuzz", ".", "partial_token_set_ratio", ",", "partial", "(", "utils", ".", "full_process", ",", "force_ascii", "=", "True", ")", ")", ",", "(", "fuzz", ".", "partial_token_sort_ratio", ",", "partial", "(", "utils", ".", "full_process", ",", "force_ascii", "=", "True", ")", ")", "]", ")", "return", "splist" ]
generate a list of pairs for testing :return: [ .
train
false
42,576
def shapeFromPath(path, pen): stroker = QPainterPathStroker() stroker.setCapStyle(pen.capStyle()) stroker.setJoinStyle(pen.joinStyle()) stroker.setMiterLimit(pen.miterLimit()) stroker.setWidth(max(pen.widthF(), 1e-09)) shape = stroker.createStroke(path) shape.addPath(path) return shape
[ "def", "shapeFromPath", "(", "path", ",", "pen", ")", ":", "stroker", "=", "QPainterPathStroker", "(", ")", "stroker", ".", "setCapStyle", "(", "pen", ".", "capStyle", "(", ")", ")", "stroker", ".", "setJoinStyle", "(", "pen", ".", "joinStyle", "(", ")", ")", "stroker", ".", "setMiterLimit", "(", "pen", ".", "miterLimit", "(", ")", ")", "stroker", ".", "setWidth", "(", "max", "(", "pen", ".", "widthF", "(", ")", ",", "1e-09", ")", ")", "shape", "=", "stroker", ".", "createStroke", "(", "path", ")", "shape", ".", "addPath", "(", "path", ")", "return", "shape" ]
create a qpainterpath shape from the path drawn with pen .
train
false
42,578
def parse_to_aware_datetime(value): if isinstance(value, datetime.datetime): datetime_obj = value else: datetime_obj = parse_timestamp(value) if (datetime_obj.tzinfo is None): datetime_obj = datetime_obj.replace(tzinfo=tzutc()) else: datetime_obj = datetime_obj.astimezone(tzutc()) return datetime_obj
[ "def", "parse_to_aware_datetime", "(", "value", ")", ":", "if", "isinstance", "(", "value", ",", "datetime", ".", "datetime", ")", ":", "datetime_obj", "=", "value", "else", ":", "datetime_obj", "=", "parse_timestamp", "(", "value", ")", "if", "(", "datetime_obj", ".", "tzinfo", "is", "None", ")", ":", "datetime_obj", "=", "datetime_obj", ".", "replace", "(", "tzinfo", "=", "tzutc", "(", ")", ")", "else", ":", "datetime_obj", "=", "datetime_obj", ".", "astimezone", "(", "tzutc", "(", ")", ")", "return", "datetime_obj" ]
converted the passed in value to a datetime object with tzinfo .
train
false
42,579
def whois_callers_caller(): import inspect frameObj = inspect.stack()[2][0] return inspect.getframeinfo(frameObj)
[ "def", "whois_callers_caller", "(", ")", ":", "import", "inspect", "frameObj", "=", "inspect", ".", "stack", "(", ")", "[", "2", "]", "[", "0", "]", "return", "inspect", ".", "getframeinfo", "(", "frameObj", ")" ]
returns: traceback namedtuple for our callers caller .
train
false
42,580
def update_settings(section, option): if ((section, option) == ('general', 'private-browsing')): cache_path = standarddir.cache() if (config.get('general', 'private-browsing') or (cache_path is None)): QWebSettings.setIconDatabasePath('') else: QWebSettings.setIconDatabasePath(cache_path) elif ((section == 'ui') and (option in ['hide-scrollbar', 'user-stylesheet'])): _set_user_stylesheet() websettings.update_mappings(MAPPINGS, section, option)
[ "def", "update_settings", "(", "section", ",", "option", ")", ":", "if", "(", "(", "section", ",", "option", ")", "==", "(", "'general'", ",", "'private-browsing'", ")", ")", ":", "cache_path", "=", "standarddir", ".", "cache", "(", ")", "if", "(", "config", ".", "get", "(", "'general'", ",", "'private-browsing'", ")", "or", "(", "cache_path", "is", "None", ")", ")", ":", "QWebSettings", ".", "setIconDatabasePath", "(", "''", ")", "else", ":", "QWebSettings", ".", "setIconDatabasePath", "(", "cache_path", ")", "elif", "(", "(", "section", "==", "'ui'", ")", "and", "(", "option", "in", "[", "'hide-scrollbar'", ",", "'user-stylesheet'", "]", ")", ")", ":", "_set_user_stylesheet", "(", ")", "websettings", ".", "update_mappings", "(", "MAPPINGS", ",", "section", ",", "option", ")" ]
update global settings when qwebsettings changed .
train
false
42,581
@requires_segment_info def paste_indicator(pl, segment_info, text=u'PASTE'): return (text if int(vim.eval(u'&paste')) else None)
[ "@", "requires_segment_info", "def", "paste_indicator", "(", "pl", ",", "segment_info", ",", "text", "=", "u'PASTE'", ")", ":", "return", "(", "text", "if", "int", "(", "vim", ".", "eval", "(", "u'&paste'", ")", ")", "else", "None", ")" ]
return a paste mode indicator .
train
false
42,582
@scopes.add_arg_scope def global_step(device=''): global_step_ref = tf.get_collection(tf.GraphKeys.GLOBAL_STEP) if global_step_ref: return global_step_ref[0] else: collections = [VARIABLES_TO_RESTORE, tf.GraphKeys.GLOBAL_VARIABLES, tf.GraphKeys.GLOBAL_STEP] with tf.device(variable_device(device, 'global_step')): return tf.get_variable('global_step', shape=[], dtype=tf.int64, initializer=tf.zeros_initializer, trainable=False, collections=collections)
[ "@", "scopes", ".", "add_arg_scope", "def", "global_step", "(", "device", "=", "''", ")", ":", "global_step_ref", "=", "tf", ".", "get_collection", "(", "tf", ".", "GraphKeys", ".", "GLOBAL_STEP", ")", "if", "global_step_ref", ":", "return", "global_step_ref", "[", "0", "]", "else", ":", "collections", "=", "[", "VARIABLES_TO_RESTORE", ",", "tf", ".", "GraphKeys", ".", "GLOBAL_VARIABLES", ",", "tf", ".", "GraphKeys", ".", "GLOBAL_STEP", "]", "with", "tf", ".", "device", "(", "variable_device", "(", "device", ",", "'global_step'", ")", ")", ":", "return", "tf", ".", "get_variable", "(", "'global_step'", ",", "shape", "=", "[", "]", ",", "dtype", "=", "tf", ".", "int64", ",", "initializer", "=", "tf", ".", "zeros_initializer", ",", "trainable", "=", "False", ",", "collections", "=", "collections", ")" ]
returns the global step variable .
train
true
42,583
def obj2dict(obj, processed=None): processed = (processed if (not (processed is None)) else set()) if (obj is None): return None if isinstance(obj, (int, long, str, unicode, float, bool)): return obj if (id(obj) in processed): return '<reference>' processed.add(id(obj)) if isinstance(obj, (list, tuple)): return [obj2dict(item, processed) for item in obj] if ((not isinstance(obj, dict)) and hasattr(obj, '__dict__')): obj = obj.__dict__ else: return repr(obj) return dict(((key, obj2dict(value, processed)) for (key, value) in obj.items() if ((not key.startswith('_')) and (not (type(value) in (types.FunctionType, types.LambdaType, types.BuiltinFunctionType, types.BuiltinMethodType))))))
[ "def", "obj2dict", "(", "obj", ",", "processed", "=", "None", ")", ":", "processed", "=", "(", "processed", "if", "(", "not", "(", "processed", "is", "None", ")", ")", "else", "set", "(", ")", ")", "if", "(", "obj", "is", "None", ")", ":", "return", "None", "if", "isinstance", "(", "obj", ",", "(", "int", ",", "long", ",", "str", ",", "unicode", ",", "float", ",", "bool", ")", ")", ":", "return", "obj", "if", "(", "id", "(", "obj", ")", "in", "processed", ")", ":", "return", "'<reference>'", "processed", ".", "add", "(", "id", "(", "obj", ")", ")", "if", "isinstance", "(", "obj", ",", "(", "list", ",", "tuple", ")", ")", ":", "return", "[", "obj2dict", "(", "item", ",", "processed", ")", "for", "item", "in", "obj", "]", "if", "(", "(", "not", "isinstance", "(", "obj", ",", "dict", ")", ")", "and", "hasattr", "(", "obj", ",", "'__dict__'", ")", ")", ":", "obj", "=", "obj", ".", "__dict__", "else", ":", "return", "repr", "(", "obj", ")", "return", "dict", "(", "(", "(", "key", ",", "obj2dict", "(", "value", ",", "processed", ")", ")", "for", "(", "key", ",", "value", ")", "in", "obj", ".", "items", "(", ")", "if", "(", "(", "not", "key", ".", "startswith", "(", "'_'", ")", ")", "and", "(", "not", "(", "type", "(", "value", ")", "in", "(", "types", ".", "FunctionType", ",", "types", ".", "LambdaType", ",", "types", ".", "BuiltinFunctionType", ",", "types", ".", "BuiltinMethodType", ")", ")", ")", ")", ")", ")" ]
converts any object into a dict .
train
false
42,584
def _ensure_has_len(seq): try: len(seq) except TypeError: return list(seq) else: return seq
[ "def", "_ensure_has_len", "(", "seq", ")", ":", "try", ":", "len", "(", "seq", ")", "except", "TypeError", ":", "return", "list", "(", "seq", ")", "else", ":", "return", "seq" ]
if seq is an iterator .
train
false
42,585
def authorize_user_context(context, user_id): if is_user_context(context): if (not context.user_id): raise exception.Forbidden() elif (context.user_id != user_id): raise exception.Forbidden()
[ "def", "authorize_user_context", "(", "context", ",", "user_id", ")", ":", "if", "is_user_context", "(", "context", ")", ":", "if", "(", "not", "context", ".", "user_id", ")", ":", "raise", "exception", ".", "Forbidden", "(", ")", "elif", "(", "context", ".", "user_id", "!=", "user_id", ")", ":", "raise", "exception", ".", "Forbidden", "(", ")" ]
ensures a request has permission to access the given user .
train
false
42,586
def reset(): _runtime.reset()
[ "def", "reset", "(", ")", ":", "_runtime", ".", "reset", "(", ")" ]
resets the db contents .
train
false
42,587
def set_caa_hostname(new_hostname): global hostname hostname = new_hostname
[ "def", "set_caa_hostname", "(", "new_hostname", ")", ":", "global", "hostname", "hostname", "=", "new_hostname" ]
set the base hostname for cover art archive requests .
train
false
42,588
@webob.dec.wsgify @microversion.version_handler(1.2) @util.check_accept('application/json') def get_resource_class(req): name = util.wsgi_path_item(req.environ, 'name') context = req.environ['placement.context'] rc = objects.ResourceClass.get_by_name(context, name) req.response.body = encodeutils.to_utf8(jsonutils.dumps(_serialize_resource_class(req.environ, rc))) req.response.content_type = 'application/json' return req.response
[ "@", "webob", ".", "dec", ".", "wsgify", "@", "microversion", ".", "version_handler", "(", "1.2", ")", "@", "util", ".", "check_accept", "(", "'application/json'", ")", "def", "get_resource_class", "(", "req", ")", ":", "name", "=", "util", ".", "wsgi_path_item", "(", "req", ".", "environ", ",", "'name'", ")", "context", "=", "req", ".", "environ", "[", "'placement.context'", "]", "rc", "=", "objects", ".", "ResourceClass", ".", "get_by_name", "(", "context", ",", "name", ")", "req", ".", "response", ".", "body", "=", "encodeutils", ".", "to_utf8", "(", "jsonutils", ".", "dumps", "(", "_serialize_resource_class", "(", "req", ".", "environ", ",", "rc", ")", ")", ")", "req", ".", "response", ".", "content_type", "=", "'application/json'", "return", "req", ".", "response" ]
get a single resource class .
train
false
42,589
def test_ipython_notebook(Chart, datas): chart = Chart() chart = make_data(chart, datas) assert chart._repr_svg_()
[ "def", "test_ipython_notebook", "(", "Chart", ",", "datas", ")", ":", "chart", "=", "Chart", "(", ")", "chart", "=", "make_data", "(", "chart", ",", "datas", ")", "assert", "chart", ".", "_repr_svg_", "(", ")" ]
test ipython notebook .
train
false
42,590
def read_user_dict(var_name, default_value): if (not isinstance(default_value, dict)): raise TypeError default_display = 'default' user_value = click.prompt(var_name, default=default_display, type=click.STRING, value_proc=process_json) if (user_value == default_display): return default_value return user_value
[ "def", "read_user_dict", "(", "var_name", ",", "default_value", ")", ":", "if", "(", "not", "isinstance", "(", "default_value", ",", "dict", ")", ")", ":", "raise", "TypeError", "default_display", "=", "'default'", "user_value", "=", "click", ".", "prompt", "(", "var_name", ",", "default", "=", "default_display", ",", "type", "=", "click", ".", "STRING", ",", "value_proc", "=", "process_json", ")", "if", "(", "user_value", "==", "default_display", ")", ":", "return", "default_value", "return", "user_value" ]
prompt the user to provide a dictionary of data .
train
true
42,591
def process_file(filename): try: checksum_errors = verify_checksums(filename) if OPTIONS.compliance: compliance_errors = verify_compliance(filename) else: compliance_errors = 0 if ((OPTIONS.write_file and (checksum_errors == 0)) or OPTIONS.force): update(filename) return (checksum_errors + compliance_errors) except Exception as e: log.error('EXCEPTION {!r} .. {}'.format(filename, e)) return 1
[ "def", "process_file", "(", "filename", ")", ":", "try", ":", "checksum_errors", "=", "verify_checksums", "(", "filename", ")", "if", "OPTIONS", ".", "compliance", ":", "compliance_errors", "=", "verify_compliance", "(", "filename", ")", "else", ":", "compliance_errors", "=", "0", "if", "(", "(", "OPTIONS", ".", "write_file", "and", "(", "checksum_errors", "==", "0", ")", ")", "or", "OPTIONS", ".", "force", ")", ":", "update", "(", "filename", ")", "return", "(", "checksum_errors", "+", "compliance_errors", ")", "except", "Exception", "as", "e", ":", "log", ".", "error", "(", "'EXCEPTION {!r} .. {}'", ".", "format", "(", "filename", ",", "e", ")", ")", "return", "1" ]
handle a single .
train
false
42,593
def line_prefix(prefix, string): return '\n'.join(((prefix + x) for x in string.splitlines()))
[ "def", "line_prefix", "(", "prefix", ",", "string", ")", ":", "return", "'\\n'", ".", "join", "(", "(", "(", "prefix", "+", "x", ")", "for", "x", "in", "string", ".", "splitlines", "(", ")", ")", ")" ]
return string with all lines prefixed by prefix .
train
false
42,594
def parse_config_h(fp, vars=None): if (vars is None): vars = {} import re define_rx = re.compile('#define ([A-Z][A-Za-z0-9_]+) (.*)\n') undef_rx = re.compile('/[*] #undef ([A-Z][A-Za-z0-9_]+) [*]/\n') while True: line = fp.readline() if (not line): break m = define_rx.match(line) if m: (n, v) = m.group(1, 2) try: v = int(v) except ValueError: pass vars[n] = v else: m = undef_rx.match(line) if m: vars[m.group(1)] = 0 return vars
[ "def", "parse_config_h", "(", "fp", ",", "vars", "=", "None", ")", ":", "if", "(", "vars", "is", "None", ")", ":", "vars", "=", "{", "}", "import", "re", "define_rx", "=", "re", ".", "compile", "(", "'#define ([A-Z][A-Za-z0-9_]+) (.*)\\n'", ")", "undef_rx", "=", "re", ".", "compile", "(", "'/[*] #undef ([A-Z][A-Za-z0-9_]+) [*]/\\n'", ")", "while", "True", ":", "line", "=", "fp", ".", "readline", "(", ")", "if", "(", "not", "line", ")", ":", "break", "m", "=", "define_rx", ".", "match", "(", "line", ")", "if", "m", ":", "(", "n", ",", "v", ")", "=", "m", ".", "group", "(", "1", ",", "2", ")", "try", ":", "v", "=", "int", "(", "v", ")", "except", "ValueError", ":", "pass", "vars", "[", "n", "]", "=", "v", "else", ":", "m", "=", "undef_rx", ".", "match", "(", "line", ")", "if", "m", ":", "vars", "[", "m", ".", "group", "(", "1", ")", "]", "=", "0", "return", "vars" ]
parse a config .
train
true
42,595
def strip_ns(xml, ns): it = ET.iterparse(StringIO(xml)) ns_prefix = ('{%s}' % XMLNS) for (_, el) in it: if el.tag.startswith(ns_prefix): el.tag = el.tag[len(ns_prefix):] return it.root
[ "def", "strip_ns", "(", "xml", ",", "ns", ")", ":", "it", "=", "ET", ".", "iterparse", "(", "StringIO", "(", "xml", ")", ")", "ns_prefix", "=", "(", "'{%s}'", "%", "XMLNS", ")", "for", "(", "_", ",", "el", ")", "in", "it", ":", "if", "el", ".", "tag", ".", "startswith", "(", "ns_prefix", ")", ":", "el", ".", "tag", "=", "el", ".", "tag", "[", "len", "(", "ns_prefix", ")", ":", "]", "return", "it", ".", "root" ]
strip the provided name from tag names .
train
false
42,597
@receiver((post_save, post_delete), sender=CircuitTermination) def update_circuit(instance, **kwargs): Circuit.objects.filter(pk=instance.circuit_id).update(last_updated=timezone.now())
[ "@", "receiver", "(", "(", "post_save", ",", "post_delete", ")", ",", "sender", "=", "CircuitTermination", ")", "def", "update_circuit", "(", "instance", ",", "**", "kwargs", ")", ":", "Circuit", ".", "objects", ".", "filter", "(", "pk", "=", "instance", ".", "circuit_id", ")", ".", "update", "(", "last_updated", "=", "timezone", ".", "now", "(", ")", ")" ]
when a circuittermination has been modified .
train
false
42,598
def get_sub_folders(session, ds_browser, ds_path): search_task = session._call_method(session.vim, 'SearchDatastore_Task', ds_browser, datastorePath=str(ds_path)) try: task_info = session._wait_for_task(search_task) except vexc.FileNotFoundException: return set() if hasattr(task_info.result, 'file'): return set([file.path for file in task_info.result.file]) return set()
[ "def", "get_sub_folders", "(", "session", ",", "ds_browser", ",", "ds_path", ")", ":", "search_task", "=", "session", ".", "_call_method", "(", "session", ".", "vim", ",", "'SearchDatastore_Task'", ",", "ds_browser", ",", "datastorePath", "=", "str", "(", "ds_path", ")", ")", "try", ":", "task_info", "=", "session", ".", "_wait_for_task", "(", "search_task", ")", "except", "vexc", ".", "FileNotFoundException", ":", "return", "set", "(", ")", "if", "hasattr", "(", "task_info", ".", "result", ",", "'file'", ")", ":", "return", "set", "(", "[", "file", ".", "path", "for", "file", "in", "task_info", ".", "result", ".", "file", "]", ")", "return", "set", "(", ")" ]
return a set of subfolders for a path on a datastore .
train
false
42,599
def _uniquify_projs(projs, check_active=True, sort=True): final_projs = [] for proj in projs: if (not any((_proj_equal(p, proj, check_active) for p in final_projs))): final_projs.append(proj) my_count = count(len(final_projs)) def sorter(x): 'Sort in a nice way.' digits = [s for s in x['desc'] if s.isdigit()] if digits: sort_idx = int(digits[(-1)]) else: sort_idx = next(my_count) return (sort_idx, x['desc']) return (sorted(final_projs, key=sorter) if sort else final_projs)
[ "def", "_uniquify_projs", "(", "projs", ",", "check_active", "=", "True", ",", "sort", "=", "True", ")", ":", "final_projs", "=", "[", "]", "for", "proj", "in", "projs", ":", "if", "(", "not", "any", "(", "(", "_proj_equal", "(", "p", ",", "proj", ",", "check_active", ")", "for", "p", "in", "final_projs", ")", ")", ")", ":", "final_projs", ".", "append", "(", "proj", ")", "my_count", "=", "count", "(", "len", "(", "final_projs", ")", ")", "def", "sorter", "(", "x", ")", ":", "digits", "=", "[", "s", "for", "s", "in", "x", "[", "'desc'", "]", "if", "s", ".", "isdigit", "(", ")", "]", "if", "digits", ":", "sort_idx", "=", "int", "(", "digits", "[", "(", "-", "1", ")", "]", ")", "else", ":", "sort_idx", "=", "next", "(", "my_count", ")", "return", "(", "sort_idx", ",", "x", "[", "'desc'", "]", ")", "return", "(", "sorted", "(", "final_projs", ",", "key", "=", "sorter", ")", "if", "sort", "else", "final_projs", ")" ]
make unique projs .
train
false
42,600
@deco.keyword('Set Name Using Robot Name Attribute') def name_set_in_method_signature(a, b, *args, **kwargs): pass
[ "@", "deco", ".", "keyword", "(", "'Set Name Using Robot Name Attribute'", ")", "def", "name_set_in_method_signature", "(", "a", ",", "b", ",", "*", "args", ",", "**", "kwargs", ")", ":", "pass" ]
this makes sure that @deco .
train
false
42,602
def setgroups(groups): max_groups = None try: max_groups = os.sysconf(u'SC_NGROUPS_MAX') except Exception: pass try: return _setgroups_hack(groups[:max_groups]) except OSError as exc: if (exc.errno != errno.EPERM): raise if any(((group not in groups) for group in os.getgroups())): raise
[ "def", "setgroups", "(", "groups", ")", ":", "max_groups", "=", "None", "try", ":", "max_groups", "=", "os", ".", "sysconf", "(", "u'SC_NGROUPS_MAX'", ")", "except", "Exception", ":", "pass", "try", ":", "return", "_setgroups_hack", "(", "groups", "[", ":", "max_groups", "]", ")", "except", "OSError", "as", "exc", ":", "if", "(", "exc", ".", "errno", "!=", "errno", ".", "EPERM", ")", ":", "raise", "if", "any", "(", "(", "(", "group", "not", "in", "groups", ")", "for", "group", "in", "os", ".", "getgroups", "(", ")", ")", ")", ":", "raise" ]
set active groups from a list of group ids .
train
false
42,605
def p_postfix_expression_5(t): pass
[ "def", "p_postfix_expression_5", "(", "t", ")", ":", "pass" ]
postfix_expression : postfix_expression period id .
train
false
42,606
def load_features(jobs, host, port, path, tiles): features = [] lock = _Lock() args = (lock, host, port, path, tiles, features) threads = [Thread(target=load_tile_features, args=args) for i in range(jobs)] for thread in threads: thread.start() for thread in threads: thread.join() logging.debug(('Loaded %d features' % len(features))) return features
[ "def", "load_features", "(", "jobs", ",", "host", ",", "port", ",", "path", ",", "tiles", ")", ":", "features", "=", "[", "]", "lock", "=", "_Lock", "(", ")", "args", "=", "(", "lock", ",", "host", ",", "port", ",", "path", ",", "tiles", ",", "features", ")", "threads", "=", "[", "Thread", "(", "target", "=", "load_tile_features", ",", "args", "=", "args", ")", "for", "i", "in", "range", "(", "jobs", ")", "]", "for", "thread", "in", "threads", ":", "thread", ".", "start", "(", ")", "for", "thread", "in", "threads", ":", "thread", ".", "join", "(", ")", "logging", ".", "debug", "(", "(", "'Loaded %d features'", "%", "len", "(", "features", ")", ")", ")", "return", "features" ]
load data from tiles to features .
train
false
42,607
def provider_uri_rewrites(providers): return _load_provider_feature('uri_rewrites', providers)
[ "def", "provider_uri_rewrites", "(", "providers", ")", ":", "return", "_load_provider_feature", "(", "'uri_rewrites'", ",", "providers", ")" ]
load tuples from an ordered list of dotted-notation modules which contain a uri_rewrites function uri_rewrites should accept a list of rewrites and returns an augmented list of rewrites: this allows the addition of .
train
false
42,609
def custom_object_scope(*args): return CustomObjectScope(*args)
[ "def", "custom_object_scope", "(", "*", "args", ")", ":", "return", "CustomObjectScope", "(", "*", "args", ")" ]
provides a scope that changes to _global_custom_objects cannot escape .
train
false
42,610
def showScore(score): scoreDigits = [int(x) for x in list(str(score))] totalWidth = 0 for digit in scoreDigits: totalWidth += IMAGES['numbers'][digit].get_width() Xoffset = ((SCREENWIDTH - totalWidth) / 2) for digit in scoreDigits: SCREEN.blit(IMAGES['numbers'][digit], (Xoffset, (SCREENHEIGHT * 0.1))) Xoffset += IMAGES['numbers'][digit].get_width()
[ "def", "showScore", "(", "score", ")", ":", "scoreDigits", "=", "[", "int", "(", "x", ")", "for", "x", "in", "list", "(", "str", "(", "score", ")", ")", "]", "totalWidth", "=", "0", "for", "digit", "in", "scoreDigits", ":", "totalWidth", "+=", "IMAGES", "[", "'numbers'", "]", "[", "digit", "]", ".", "get_width", "(", ")", "Xoffset", "=", "(", "(", "SCREENWIDTH", "-", "totalWidth", ")", "/", "2", ")", "for", "digit", "in", "scoreDigits", ":", "SCREEN", ".", "blit", "(", "IMAGES", "[", "'numbers'", "]", "[", "digit", "]", ",", "(", "Xoffset", ",", "(", "SCREENHEIGHT", "*", "0.1", ")", ")", ")", "Xoffset", "+=", "IMAGES", "[", "'numbers'", "]", "[", "digit", "]", ".", "get_width", "(", ")" ]
displays score in center of screen .
train
false
42,611
def before_new_template(force_new_urls): global new_style_url_tag, error_on_old_style_url_tag, errors new_style_url_tag = False error_on_old_style_url_tag = force_new_urls errors = []
[ "def", "before_new_template", "(", "force_new_urls", ")", ":", "global", "new_style_url_tag", ",", "error_on_old_style_url_tag", ",", "errors", "new_style_url_tag", "=", "False", "error_on_old_style_url_tag", "=", "force_new_urls", "errors", "=", "[", "]" ]
reset state ready for new template .
train
false
42,612
def run_all(plugin, args=''): data = _execute_cmd(plugin, args, 'cmd.run_all') return data
[ "def", "run_all", "(", "plugin", ",", "args", "=", "''", ")", ":", "data", "=", "_execute_cmd", "(", "plugin", ",", "args", ",", "'cmd.run_all'", ")", "return", "data" ]
run :mod:cmd .
train
false
42,613
def _normalize_query_parameter(value): if isinstance(value, Model): value = value.key() if (isinstance(value, datetime.date) and (not isinstance(value, datetime.datetime))): value = _date_to_datetime(value) elif isinstance(value, datetime.time): value = _time_to_datetime(value) return value
[ "def", "_normalize_query_parameter", "(", "value", ")", ":", "if", "isinstance", "(", "value", ",", "Model", ")", ":", "value", "=", "value", ".", "key", "(", ")", "if", "(", "isinstance", "(", "value", ",", "datetime", ".", "date", ")", "and", "(", "not", "isinstance", "(", "value", ",", "datetime", ".", "datetime", ")", ")", ")", ":", "value", "=", "_date_to_datetime", "(", "value", ")", "elif", "isinstance", "(", "value", ",", "datetime", ".", "time", ")", ":", "value", "=", "_time_to_datetime", "(", "value", ")", "return", "value" ]
make any necessary type conversions to a query parameter .
train
false
42,614
def _pack(x): try: return list(x) except TypeError: return [x]
[ "def", "_pack", "(", "x", ")", ":", "try", ":", "return", "list", "(", "x", ")", "except", "TypeError", ":", "return", "[", "x", "]" ]
pack parts into a ssh key blob .
train
false
42,615
def _estimate_gaussian_parameters(X, resp, reg_covar, covariance_type): nk = (resp.sum(axis=0) + (10 * np.finfo(resp.dtype).eps)) means = (np.dot(resp.T, X) / nk[:, np.newaxis]) covariances = {'full': _estimate_gaussian_covariances_full, 'tied': _estimate_gaussian_covariances_tied, 'diag': _estimate_gaussian_covariances_diag, 'spherical': _estimate_gaussian_covariances_spherical}[covariance_type](resp, X, nk, means, reg_covar) return (nk, means, covariances)
[ "def", "_estimate_gaussian_parameters", "(", "X", ",", "resp", ",", "reg_covar", ",", "covariance_type", ")", ":", "nk", "=", "(", "resp", ".", "sum", "(", "axis", "=", "0", ")", "+", "(", "10", "*", "np", ".", "finfo", "(", "resp", ".", "dtype", ")", ".", "eps", ")", ")", "means", "=", "(", "np", ".", "dot", "(", "resp", ".", "T", ",", "X", ")", "/", "nk", "[", ":", ",", "np", ".", "newaxis", "]", ")", "covariances", "=", "{", "'full'", ":", "_estimate_gaussian_covariances_full", ",", "'tied'", ":", "_estimate_gaussian_covariances_tied", ",", "'diag'", ":", "_estimate_gaussian_covariances_diag", ",", "'spherical'", ":", "_estimate_gaussian_covariances_spherical", "}", "[", "covariance_type", "]", "(", "resp", ",", "X", ",", "nk", ",", "means", ",", "reg_covar", ")", "return", "(", "nk", ",", "means", ",", "covariances", ")" ]
estimate the gaussian distribution parameters .
train
false
42,616
def setup_env(): os.environ['AUTH_DOMAIN'] = 'appscale.com' os.environ['USER_EMAIL'] = '' os.environ['USER_NICKNAME'] = '' os.environ['APPLICATION_ID'] = ''
[ "def", "setup_env", "(", ")", ":", "os", ".", "environ", "[", "'AUTH_DOMAIN'", "]", "=", "'appscale.com'", "os", ".", "environ", "[", "'USER_EMAIL'", "]", "=", "''", "os", ".", "environ", "[", "'USER_NICKNAME'", "]", "=", "''", "os", ".", "environ", "[", "'APPLICATION_ID'", "]", "=", "''" ]
setup virtualenv for bench .
train
false
42,617
def create_lti_user(lti_user_id, lti_consumer): edx_password = str(uuid.uuid4()) created = False while (not created): try: edx_user_id = generate_random_edx_username() edx_email = '{}@{}'.format(edx_user_id, settings.LTI_USER_EMAIL_DOMAIN) with transaction.atomic(): edx_user = User.objects.create_user(username=edx_user_id, password=edx_password, email=edx_email) edx_user_profile = UserProfile(user=edx_user) edx_user_profile.save() created = True except IntegrityError: pass lti_user = LtiUser(lti_consumer=lti_consumer, lti_user_id=lti_user_id, edx_user=edx_user) lti_user.save() return lti_user
[ "def", "create_lti_user", "(", "lti_user_id", ",", "lti_consumer", ")", ":", "edx_password", "=", "str", "(", "uuid", ".", "uuid4", "(", ")", ")", "created", "=", "False", "while", "(", "not", "created", ")", ":", "try", ":", "edx_user_id", "=", "generate_random_edx_username", "(", ")", "edx_email", "=", "'{}@{}'", ".", "format", "(", "edx_user_id", ",", "settings", ".", "LTI_USER_EMAIL_DOMAIN", ")", "with", "transaction", ".", "atomic", "(", ")", ":", "edx_user", "=", "User", ".", "objects", ".", "create_user", "(", "username", "=", "edx_user_id", ",", "password", "=", "edx_password", ",", "email", "=", "edx_email", ")", "edx_user_profile", "=", "UserProfile", "(", "user", "=", "edx_user", ")", "edx_user_profile", ".", "save", "(", ")", "created", "=", "True", "except", "IntegrityError", ":", "pass", "lti_user", "=", "LtiUser", "(", "lti_consumer", "=", "lti_consumer", ",", "lti_user_id", "=", "lti_user_id", ",", "edx_user", "=", "edx_user", ")", "lti_user", ".", "save", "(", ")", "return", "lti_user" ]
generate a new user on the edx platform with a random username and password .
train
false
42,618
def _locate_repo_files(repo, rewrite=False): ret_val = [] files = [] conf_dirs = ['/etc/xbps.d/', '/usr/share/xbps.d/'] name_glob = '*.conf' regex = re.compile((('\\s*repository\\s*=\\s*' + repo) + '/?\\s*(#.*)?$')) for cur_dir in conf_dirs: files.extend(glob.glob((cur_dir + name_glob))) for filename in files: write_buff = [] with salt.utils.fopen(filename, 'r') as cur_file: for line in cur_file: if regex.match(line): ret_val.append(filename) else: write_buff.append(line) if (rewrite and (filename in ret_val)): if (len(write_buff) > 0): with salt.utils.fopen(filename, 'w') as rewrite_file: rewrite_file.write(''.join(write_buff)) else: os.remove(filename) return ret_val
[ "def", "_locate_repo_files", "(", "repo", ",", "rewrite", "=", "False", ")", ":", "ret_val", "=", "[", "]", "files", "=", "[", "]", "conf_dirs", "=", "[", "'/etc/xbps.d/'", ",", "'/usr/share/xbps.d/'", "]", "name_glob", "=", "'*.conf'", "regex", "=", "re", ".", "compile", "(", "(", "(", "'\\\\s*repository\\\\s*=\\\\s*'", "+", "repo", ")", "+", "'/?\\\\s*(#.*)?$'", ")", ")", "for", "cur_dir", "in", "conf_dirs", ":", "files", ".", "extend", "(", "glob", ".", "glob", "(", "(", "cur_dir", "+", "name_glob", ")", ")", ")", "for", "filename", "in", "files", ":", "write_buff", "=", "[", "]", "with", "salt", ".", "utils", ".", "fopen", "(", "filename", ",", "'r'", ")", "as", "cur_file", ":", "for", "line", "in", "cur_file", ":", "if", "regex", ".", "match", "(", "line", ")", ":", "ret_val", ".", "append", "(", "filename", ")", "else", ":", "write_buff", ".", "append", "(", "line", ")", "if", "(", "rewrite", "and", "(", "filename", "in", "ret_val", ")", ")", ":", "if", "(", "len", "(", "write_buff", ")", ">", "0", ")", ":", "with", "salt", ".", "utils", ".", "fopen", "(", "filename", ",", "'w'", ")", "as", "rewrite_file", ":", "rewrite_file", ".", "write", "(", "''", ".", "join", "(", "write_buff", ")", ")", "else", ":", "os", ".", "remove", "(", "filename", ")", "return", "ret_val" ]
find what file a repo is called in .
train
false
42,620
def _parseClientSSL(**kwargs): from twisted.internet import ssl kwargs = _parseClientTCP(**kwargs) certKey = kwargs.pop('certKey', None) privateKey = kwargs.pop('privateKey', None) caCertsDir = kwargs.pop('caCertsDir', None) if (certKey is not None): certx509 = ssl.Certificate.loadPEM(FilePath(certKey).getContent()).original else: certx509 = None if (privateKey is not None): privateKey = ssl.PrivateCertificate.loadPEM(FilePath(privateKey).getContent()).privateKey.original else: privateKey = None if (caCertsDir is not None): verify = True caCerts = _loadCAsFromDir(FilePath(caCertsDir)) else: verify = False caCerts = None kwargs['sslContextFactory'] = ssl.CertificateOptions(method=ssl.SSL.SSLv23_METHOD, certificate=certx509, privateKey=privateKey, verify=verify, caCerts=caCerts) return kwargs
[ "def", "_parseClientSSL", "(", "**", "kwargs", ")", ":", "from", "twisted", ".", "internet", "import", "ssl", "kwargs", "=", "_parseClientTCP", "(", "**", "kwargs", ")", "certKey", "=", "kwargs", ".", "pop", "(", "'certKey'", ",", "None", ")", "privateKey", "=", "kwargs", ".", "pop", "(", "'privateKey'", ",", "None", ")", "caCertsDir", "=", "kwargs", ".", "pop", "(", "'caCertsDir'", ",", "None", ")", "if", "(", "certKey", "is", "not", "None", ")", ":", "certx509", "=", "ssl", ".", "Certificate", ".", "loadPEM", "(", "FilePath", "(", "certKey", ")", ".", "getContent", "(", ")", ")", ".", "original", "else", ":", "certx509", "=", "None", "if", "(", "privateKey", "is", "not", "None", ")", ":", "privateKey", "=", "ssl", ".", "PrivateCertificate", ".", "loadPEM", "(", "FilePath", "(", "privateKey", ")", ".", "getContent", "(", ")", ")", ".", "privateKey", ".", "original", "else", ":", "privateKey", "=", "None", "if", "(", "caCertsDir", "is", "not", "None", ")", ":", "verify", "=", "True", "caCerts", "=", "_loadCAsFromDir", "(", "FilePath", "(", "caCertsDir", ")", ")", "else", ":", "verify", "=", "False", "caCerts", "=", "None", "kwargs", "[", "'sslContextFactory'", "]", "=", "ssl", ".", "CertificateOptions", "(", "method", "=", "ssl", ".", "SSL", ".", "SSLv23_METHOD", ",", "certificate", "=", "certx509", ",", "privateKey", "=", "privateKey", ",", "verify", "=", "verify", ",", "caCerts", "=", "caCerts", ")", "return", "kwargs" ]
perform any argument value coercion necessary for ssl client parameters .
train
false
42,622
def get_limited_to(headers): policy_dict = dict() policy_dict['roles'] = headers.get('X-Roles', '').split(',') policy_dict['user_id'] = headers.get('X-User-Id') policy_dict['project_id'] = headers.get('X-Project-Id') rule_name = ('segregation' if _has_rule('segregation') else 'context_is_admin') if (not pecan.request.enforcer.enforce(rule_name, {}, policy_dict)): return (headers.get('X-User-Id'), headers.get('X-Project-Id')) return (None, None)
[ "def", "get_limited_to", "(", "headers", ")", ":", "policy_dict", "=", "dict", "(", ")", "policy_dict", "[", "'roles'", "]", "=", "headers", ".", "get", "(", "'X-Roles'", ",", "''", ")", ".", "split", "(", "','", ")", "policy_dict", "[", "'user_id'", "]", "=", "headers", ".", "get", "(", "'X-User-Id'", ")", "policy_dict", "[", "'project_id'", "]", "=", "headers", ".", "get", "(", "'X-Project-Id'", ")", "rule_name", "=", "(", "'segregation'", "if", "_has_rule", "(", "'segregation'", ")", "else", "'context_is_admin'", ")", "if", "(", "not", "pecan", ".", "request", ".", "enforcer", ".", "enforce", "(", "rule_name", ",", "{", "}", ",", "policy_dict", ")", ")", ":", "return", "(", "headers", ".", "get", "(", "'X-User-Id'", ")", ",", "headers", ".", "get", "(", "'X-Project-Id'", ")", ")", "return", "(", "None", ",", "None", ")" ]
return the user and project the request should be limited to .
train
false
42,623
def dup_degree(f): if (not f): return (- oo) return (len(f) - 1)
[ "def", "dup_degree", "(", "f", ")", ":", "if", "(", "not", "f", ")", ":", "return", "(", "-", "oo", ")", "return", "(", "len", "(", "f", ")", "-", "1", ")" ]
return the leading degree of f in k[x] .
train
false
42,624
def numCores(): if hasattr(numCores, 'ncores'): return numCores.ncores try: numCores.ncores = int(quietRun('grep -c processor /proc/cpuinfo')) except ValueError: return 0 return numCores.ncores
[ "def", "numCores", "(", ")", ":", "if", "hasattr", "(", "numCores", ",", "'ncores'", ")", ":", "return", "numCores", ".", "ncores", "try", ":", "numCores", ".", "ncores", "=", "int", "(", "quietRun", "(", "'grep -c processor /proc/cpuinfo'", ")", ")", "except", "ValueError", ":", "return", "0", "return", "numCores", ".", "ncores" ]
returns number of cpu cores based on /proc/cpuinfo .
train
false
42,625
def split_escape(string, sep, maxsplit=None, escape_char='\\'): assert (len(sep) == 1) assert (len(escape_char) == 1) if isinstance(string, bytes): if isinstance(escape_char, text_type): escape_char = escape_char.encode('ascii') iter_ = iterbytes else: iter_ = iter if (maxsplit is None): maxsplit = len(string) empty = string[:0] result = [] current = empty escaped = False for char in iter_(string): if escaped: if ((char != escape_char) and (char != sep)): current += escape_char current += char escaped = False elif (char == escape_char): escaped = True elif ((char == sep) and (len(result) < maxsplit)): result.append(current) current = empty else: current += char result.append(current) return result
[ "def", "split_escape", "(", "string", ",", "sep", ",", "maxsplit", "=", "None", ",", "escape_char", "=", "'\\\\'", ")", ":", "assert", "(", "len", "(", "sep", ")", "==", "1", ")", "assert", "(", "len", "(", "escape_char", ")", "==", "1", ")", "if", "isinstance", "(", "string", ",", "bytes", ")", ":", "if", "isinstance", "(", "escape_char", ",", "text_type", ")", ":", "escape_char", "=", "escape_char", ".", "encode", "(", "'ascii'", ")", "iter_", "=", "iterbytes", "else", ":", "iter_", "=", "iter", "if", "(", "maxsplit", "is", "None", ")", ":", "maxsplit", "=", "len", "(", "string", ")", "empty", "=", "string", "[", ":", "0", "]", "result", "=", "[", "]", "current", "=", "empty", "escaped", "=", "False", "for", "char", "in", "iter_", "(", "string", ")", ":", "if", "escaped", ":", "if", "(", "(", "char", "!=", "escape_char", ")", "and", "(", "char", "!=", "sep", ")", ")", ":", "current", "+=", "escape_char", "current", "+=", "char", "escaped", "=", "False", "elif", "(", "char", "==", "escape_char", ")", ":", "escaped", "=", "True", "elif", "(", "(", "char", "==", "sep", ")", "and", "(", "len", "(", "result", ")", "<", "maxsplit", ")", ")", ":", "result", ".", "append", "(", "current", ")", "current", "=", "empty", "else", ":", "current", "+=", "char", "result", ".", "append", "(", "current", ")", "return", "result" ]
like unicode/str/bytes .
train
true
42,628
def print_usage(actions): actions = actions.items() actions.sort() print ('usage: %s <action> [<options>]' % basename(sys.argv[0])) print (' %s --help' % basename(sys.argv[0])) print print 'actions:' for (name, (func, doc, arguments)) in actions: print (' %s:' % name) for line in doc.splitlines(): print (' %s' % line) if arguments: print for (arg, shortcut, default, argtype) in arguments: if isinstance(default, bool): print (' %s' % ((((shortcut and ('-%s, ' % shortcut)) or '') + '--') + arg)) else: print (' %-30s%-10s%s' % (((((shortcut and ('-%s, ' % shortcut)) or '') + '--') + arg), argtype, default)) print
[ "def", "print_usage", "(", "actions", ")", ":", "actions", "=", "actions", ".", "items", "(", ")", "actions", ".", "sort", "(", ")", "print", "(", "'usage: %s <action> [<options>]'", "%", "basename", "(", "sys", ".", "argv", "[", "0", "]", ")", ")", "print", "(", "' %s --help'", "%", "basename", "(", "sys", ".", "argv", "[", "0", "]", ")", ")", "print", "print", "'actions:'", "for", "(", "name", ",", "(", "func", ",", "doc", ",", "arguments", ")", ")", "in", "actions", ":", "print", "(", "' %s:'", "%", "name", ")", "for", "line", "in", "doc", ".", "splitlines", "(", ")", ":", "print", "(", "' %s'", "%", "line", ")", "if", "arguments", ":", "print", "for", "(", "arg", ",", "shortcut", ",", "default", ",", "argtype", ")", "in", "arguments", ":", "if", "isinstance", "(", "default", ",", "bool", ")", ":", "print", "(", "' %s'", "%", "(", "(", "(", "(", "shortcut", "and", "(", "'-%s, '", "%", "shortcut", ")", ")", "or", "''", ")", "+", "'--'", ")", "+", "arg", ")", ")", "else", ":", "print", "(", "' %-30s%-10s%s'", "%", "(", "(", "(", "(", "(", "shortcut", "and", "(", "'-%s, '", "%", "shortcut", ")", ")", "or", "''", ")", "+", "'--'", ")", "+", "arg", ")", ",", "argtype", ",", "default", ")", ")", "print" ]
print the usage information .
train
true
42,629
def list_to_dict(a_list): new_dict = {} for (i, v) in enumerate(a_list): new_dict[i] = v return new_dict
[ "def", "list_to_dict", "(", "a_list", ")", ":", "new_dict", "=", "{", "}", "for", "(", "i", ",", "v", ")", "in", "enumerate", "(", "a_list", ")", ":", "new_dict", "[", "i", "]", "=", "v", "return", "new_dict" ]
convert a list to a dictionary .
train
false
42,630
def _comp_sum_eeg(beta, ctheta, lut_fun, n_fact): n_chunk = (50000000 // ((8 * max(n_fact.shape)) * 2)) lims = np.concatenate([np.arange(0, beta.size, n_chunk), [beta.size]]) s0 = np.empty(beta.shape) for (start, stop) in zip(lims[:(-1)], lims[1:]): coeffs = lut_fun(ctheta[start:stop]) betans = np.tile(beta[start:stop][:, np.newaxis], (1, n_fact.shape[0])) np.cumprod(betans, axis=1, out=betans) coeffs *= betans s0[start:stop] = np.dot(coeffs, n_fact) return s0
[ "def", "_comp_sum_eeg", "(", "beta", ",", "ctheta", ",", "lut_fun", ",", "n_fact", ")", ":", "n_chunk", "=", "(", "50000000", "//", "(", "(", "8", "*", "max", "(", "n_fact", ".", "shape", ")", ")", "*", "2", ")", ")", "lims", "=", "np", ".", "concatenate", "(", "[", "np", ".", "arange", "(", "0", ",", "beta", ".", "size", ",", "n_chunk", ")", ",", "[", "beta", ".", "size", "]", "]", ")", "s0", "=", "np", ".", "empty", "(", "beta", ".", "shape", ")", "for", "(", "start", ",", "stop", ")", "in", "zip", "(", "lims", "[", ":", "(", "-", "1", ")", "]", ",", "lims", "[", "1", ":", "]", ")", ":", "coeffs", "=", "lut_fun", "(", "ctheta", "[", "start", ":", "stop", "]", ")", "betans", "=", "np", ".", "tile", "(", "beta", "[", "start", ":", "stop", "]", "[", ":", ",", "np", ".", "newaxis", "]", ",", "(", "1", ",", "n_fact", ".", "shape", "[", "0", "]", ")", ")", "np", ".", "cumprod", "(", "betans", ",", "axis", "=", "1", ",", "out", "=", "betans", ")", "coeffs", "*=", "betans", "s0", "[", "start", ":", "stop", "]", "=", "np", ".", "dot", "(", "coeffs", ",", "n_fact", ")", "return", "s0" ]
lead field dot products using legendre polynomial series .
train
false
42,631
def scale_matrix(factor, origin=None, direction=None): if (direction is None): M = numpy.diag([factor, factor, factor, 1.0]) if (origin is not None): M[:3, 3] = origin[:3] M[:3, 3] *= (1.0 - factor) else: direction = unit_vector(direction[:3]) factor = (1.0 - factor) M = numpy.identity(4) M[:3, :3] -= (factor * numpy.outer(direction, direction)) if (origin is not None): M[:3, 3] = ((factor * numpy.dot(origin[:3], direction)) * direction) return M
[ "def", "scale_matrix", "(", "factor", ",", "origin", "=", "None", ",", "direction", "=", "None", ")", ":", "if", "(", "direction", "is", "None", ")", ":", "M", "=", "numpy", ".", "diag", "(", "[", "factor", ",", "factor", ",", "factor", ",", "1.0", "]", ")", "if", "(", "origin", "is", "not", "None", ")", ":", "M", "[", ":", "3", ",", "3", "]", "=", "origin", "[", ":", "3", "]", "M", "[", ":", "3", ",", "3", "]", "*=", "(", "1.0", "-", "factor", ")", "else", ":", "direction", "=", "unit_vector", "(", "direction", "[", ":", "3", "]", ")", "factor", "=", "(", "1.0", "-", "factor", ")", "M", "=", "numpy", ".", "identity", "(", "4", ")", "M", "[", ":", "3", ",", ":", "3", "]", "-=", "(", "factor", "*", "numpy", ".", "outer", "(", "direction", ",", "direction", ")", ")", "if", "(", "origin", "is", "not", "None", ")", ":", "M", "[", ":", "3", ",", "3", "]", "=", "(", "(", "factor", "*", "numpy", ".", "dot", "(", "origin", "[", ":", "3", "]", ",", "direction", ")", ")", "*", "direction", ")", "return", "M" ]
return matrix to scale by factor around origin in direction .
train
true
42,632
def NodeName(node): if (node.type < 256): return token.tok_name[node.type] else: return pygram.python_grammar.number2symbol[node.type]
[ "def", "NodeName", "(", "node", ")", ":", "if", "(", "node", ".", "type", "<", "256", ")", ":", "return", "token", ".", "tok_name", "[", "node", ".", "type", "]", "else", ":", "return", "pygram", ".", "python_grammar", ".", "number2symbol", "[", "node", ".", "type", "]" ]
produce a string name for a given node .
train
false
42,633
def _blockdevicevolume_from_ebs_volume(ebs_volume): if ebs_volume.attachments: attached_to = unicode(ebs_volume.attachments[0]['InstanceId']) else: attached_to = None volume_dataset_id = _get_volume_tag(ebs_volume, DATASET_ID_LABEL) return BlockDeviceVolume(blockdevice_id=unicode(ebs_volume.id), size=int(GiB(ebs_volume.size).to_Byte().value), attached_to=attached_to, dataset_id=UUID(volume_dataset_id))
[ "def", "_blockdevicevolume_from_ebs_volume", "(", "ebs_volume", ")", ":", "if", "ebs_volume", ".", "attachments", ":", "attached_to", "=", "unicode", "(", "ebs_volume", ".", "attachments", "[", "0", "]", "[", "'InstanceId'", "]", ")", "else", ":", "attached_to", "=", "None", "volume_dataset_id", "=", "_get_volume_tag", "(", "ebs_volume", ",", "DATASET_ID_LABEL", ")", "return", "BlockDeviceVolume", "(", "blockdevice_id", "=", "unicode", "(", "ebs_volume", ".", "id", ")", ",", "size", "=", "int", "(", "GiB", "(", "ebs_volume", ".", "size", ")", ".", "to_Byte", "(", ")", ".", "value", ")", ",", "attached_to", "=", "attached_to", ",", "dataset_id", "=", "UUID", "(", "volume_dataset_id", ")", ")" ]
helper function to convert volume information from ebs format to flocker block device format .
train
false
42,634
def _estimate_gaussian_covariances_full(resp, X, nk, means, reg_covar): (n_components, n_features) = means.shape covariances = np.empty((n_components, n_features, n_features)) for k in range(n_components): diff = (X - means[k]) covariances[k] = (np.dot((resp[:, k] * diff.T), diff) / nk[k]) covariances[k].flat[::(n_features + 1)] += reg_covar return covariances
[ "def", "_estimate_gaussian_covariances_full", "(", "resp", ",", "X", ",", "nk", ",", "means", ",", "reg_covar", ")", ":", "(", "n_components", ",", "n_features", ")", "=", "means", ".", "shape", "covariances", "=", "np", ".", "empty", "(", "(", "n_components", ",", "n_features", ",", "n_features", ")", ")", "for", "k", "in", "range", "(", "n_components", ")", ":", "diff", "=", "(", "X", "-", "means", "[", "k", "]", ")", "covariances", "[", "k", "]", "=", "(", "np", ".", "dot", "(", "(", "resp", "[", ":", ",", "k", "]", "*", "diff", ".", "T", ")", ",", "diff", ")", "/", "nk", "[", "k", "]", ")", "covariances", "[", "k", "]", ".", "flat", "[", ":", ":", "(", "n_features", "+", "1", ")", "]", "+=", "reg_covar", "return", "covariances" ]
estimate the full covariance matrices .
train
false
42,636
@utils.auth.requires_login @blueprint.route('/upload_archive', methods=['POST']) def upload_archive(): files = flask.request.files archive_file = get_tempfile(files['archive'], '.archive') if tarfile.is_tarfile(archive_file): archive = tarfile.open(archive_file, 'r') names = archive.getnames() elif zipfile.is_zipfile(archive_file): archive = zipfile.ZipFile(archive_file, 'r') names = archive.namelist() else: return (flask.jsonify({'status': 'Incorrect Archive Type'}), 500) if ('info.json' in names): tempdir = tempfile.mkdtemp() labels_file = None archive.extractall(path=tempdir) with open(os.path.join(tempdir, 'info.json')) as data_file: info = json.load(data_file) (valid, key) = validate_archive_keys(info) if (valid is False): return (flask.jsonify({'status': (("Missing Key '" + key) + "' in info.json")}), 500) weights_file = os.path.join(tempdir, info['snapshot file']) if ('model file' in info): model_file = os.path.join(tempdir, info['model file']) elif ('network file' in info): model_file = os.path.join(tempdir, info['network file']) else: return (flask.jsonify({'status': 'Missing model definition in info.json'}), 500) if ('labels file' in info): labels_file = os.path.join(tempdir, info['labels file']) job = PretrainedModelJob(weights_file, model_file, labels_file, info['framework'], username=utils.auth.get_username(), name=info['name']) scheduler.add_job(job) job.wait_completion() shutil.rmtree(tempdir, ignore_errors=True) return (flask.jsonify({'status': 'success'}), 200) else: return (flask.jsonify({'status': 'Missing or Incorrect json file'}), 500)
[ "@", "utils", ".", "auth", ".", "requires_login", "@", "blueprint", ".", "route", "(", "'/upload_archive'", ",", "methods", "=", "[", "'POST'", "]", ")", "def", "upload_archive", "(", ")", ":", "files", "=", "flask", ".", "request", ".", "files", "archive_file", "=", "get_tempfile", "(", "files", "[", "'archive'", "]", ",", "'.archive'", ")", "if", "tarfile", ".", "is_tarfile", "(", "archive_file", ")", ":", "archive", "=", "tarfile", ".", "open", "(", "archive_file", ",", "'r'", ")", "names", "=", "archive", ".", "getnames", "(", ")", "elif", "zipfile", ".", "is_zipfile", "(", "archive_file", ")", ":", "archive", "=", "zipfile", ".", "ZipFile", "(", "archive_file", ",", "'r'", ")", "names", "=", "archive", ".", "namelist", "(", ")", "else", ":", "return", "(", "flask", ".", "jsonify", "(", "{", "'status'", ":", "'Incorrect Archive Type'", "}", ")", ",", "500", ")", "if", "(", "'info.json'", "in", "names", ")", ":", "tempdir", "=", "tempfile", ".", "mkdtemp", "(", ")", "labels_file", "=", "None", "archive", ".", "extractall", "(", "path", "=", "tempdir", ")", "with", "open", "(", "os", ".", "path", ".", "join", "(", "tempdir", ",", "'info.json'", ")", ")", "as", "data_file", ":", "info", "=", "json", ".", "load", "(", "data_file", ")", "(", "valid", ",", "key", ")", "=", "validate_archive_keys", "(", "info", ")", "if", "(", "valid", "is", "False", ")", ":", "return", "(", "flask", ".", "jsonify", "(", "{", "'status'", ":", "(", "(", "\"Missing Key '\"", "+", "key", ")", "+", "\"' in info.json\"", ")", "}", ")", ",", "500", ")", "weights_file", "=", "os", ".", "path", ".", "join", "(", "tempdir", ",", "info", "[", "'snapshot file'", "]", ")", "if", "(", "'model file'", "in", "info", ")", ":", "model_file", "=", "os", ".", "path", ".", "join", "(", "tempdir", ",", "info", "[", "'model file'", "]", ")", "elif", "(", "'network file'", "in", "info", ")", ":", "model_file", "=", "os", ".", "path", ".", "join", "(", "tempdir", ",", "info", "[", "'network file'", "]", ")", "else", ":", "return", "(", "flask", ".", "jsonify", "(", "{", "'status'", ":", "'Missing model definition in info.json'", "}", ")", ",", "500", ")", "if", "(", "'labels file'", "in", "info", ")", ":", "labels_file", "=", "os", ".", "path", ".", "join", "(", "tempdir", ",", "info", "[", "'labels file'", "]", ")", "job", "=", "PretrainedModelJob", "(", "weights_file", ",", "model_file", ",", "labels_file", ",", "info", "[", "'framework'", "]", ",", "username", "=", "utils", ".", "auth", ".", "get_username", "(", ")", ",", "name", "=", "info", "[", "'name'", "]", ")", "scheduler", ".", "add_job", "(", "job", ")", "job", ".", "wait_completion", "(", ")", "shutil", ".", "rmtree", "(", "tempdir", ",", "ignore_errors", "=", "True", ")", "return", "(", "flask", ".", "jsonify", "(", "{", "'status'", ":", "'success'", "}", ")", ",", "200", ")", "else", ":", "return", "(", "flask", ".", "jsonify", "(", "{", "'status'", ":", "'Missing or Incorrect json file'", "}", ")", ",", "500", ")" ]
a wrapper around the actual upload view function to clean up the temporary file afterwards .
train
false
42,637
def float_compare(value1, value2, precision_digits=None, precision_rounding=None): rounding_factor = _float_check_precision(precision_digits=precision_digits, precision_rounding=precision_rounding) value1 = float_round(value1, precision_rounding=rounding_factor) value2 = float_round(value2, precision_rounding=rounding_factor) delta = (value1 - value2) if float_is_zero(delta, precision_rounding=rounding_factor): return 0 return ((-1) if (delta < 0.0) else 1)
[ "def", "float_compare", "(", "value1", ",", "value2", ",", "precision_digits", "=", "None", ",", "precision_rounding", "=", "None", ")", ":", "rounding_factor", "=", "_float_check_precision", "(", "precision_digits", "=", "precision_digits", ",", "precision_rounding", "=", "precision_rounding", ")", "value1", "=", "float_round", "(", "value1", ",", "precision_rounding", "=", "rounding_factor", ")", "value2", "=", "float_round", "(", "value2", ",", "precision_rounding", "=", "rounding_factor", ")", "delta", "=", "(", "value1", "-", "value2", ")", "if", "float_is_zero", "(", "delta", ",", "precision_rounding", "=", "rounding_factor", ")", ":", "return", "0", "return", "(", "(", "-", "1", ")", "if", "(", "delta", "<", "0.0", ")", "else", "1", ")" ]
compare value1 and value2 after rounding them according to the given precision .
train
false
42,639
def _switch_package_license_ids(old_ids, old_license_titles, map): new_ids = {} for (package_id, old_license_id) in old_ids.items(): if (old_license_id != None): old_license_title = old_license_titles[old_license_id] new_license_id = map[old_license_title] new_ids[package_id] = new_license_id print ('Switched license_id %s to %s' % (old_license_id, new_license_id)) return new_ids
[ "def", "_switch_package_license_ids", "(", "old_ids", ",", "old_license_titles", ",", "map", ")", ":", "new_ids", "=", "{", "}", "for", "(", "package_id", ",", "old_license_id", ")", "in", "old_ids", ".", "items", "(", ")", ":", "if", "(", "old_license_id", "!=", "None", ")", ":", "old_license_title", "=", "old_license_titles", "[", "old_license_id", "]", "new_license_id", "=", "map", "[", "old_license_title", "]", "new_ids", "[", "package_id", "]", "=", "new_license_id", "print", "(", "'Switched license_id %s to %s'", "%", "(", "old_license_id", ",", "new_license_id", ")", ")", "return", "new_ids" ]
returns a dict of new license ids .
train
false
42,642
def abs(a): return _abs(a)
[ "def", "abs", "(", "a", ")", ":", "return", "_abs", "(", "a", ")" ]
element-wise absolute value .
train
false
42,643
@register.simple_tag(takes_context=True) def avatar_urls(context, user, size, service_id=None): service = avatar_services.for_user(user, service_id) if (service is None): logging.error(u'Could not get a suitable avatar service for user %s.', user) urls = {} else: urls = {resolution: url for (resolution, url) in six.iteritems(service.get_avatar_urls(request=context[u'request'], user=user, size=size))} return mark_safe(json.dumps(urls))
[ "@", "register", ".", "simple_tag", "(", "takes_context", "=", "True", ")", "def", "avatar_urls", "(", "context", ",", "user", ",", "size", ",", "service_id", "=", "None", ")", ":", "service", "=", "avatar_services", ".", "for_user", "(", "user", ",", "service_id", ")", "if", "(", "service", "is", "None", ")", ":", "logging", ".", "error", "(", "u'Could not get a suitable avatar service for user %s.'", ",", "user", ")", "urls", "=", "{", "}", "else", ":", "urls", "=", "{", "resolution", ":", "url", "for", "(", "resolution", ",", "url", ")", "in", "six", ".", "iteritems", "(", "service", ".", "get_avatar_urls", "(", "request", "=", "context", "[", "u'request'", "]", ",", "user", "=", "user", ",", "size", "=", "size", ")", ")", "}", "return", "mark_safe", "(", "json", ".", "dumps", "(", "urls", ")", ")" ]
serialize the users avatar urls into a javascript object .
train
false
42,644
def quaternion_multiply(quaternion1, quaternion0): (w0, x0, y0, z0) = quaternion0 (w1, x1, y1, z1) = quaternion1 return numpy.array([(((((- x1) * x0) - (y1 * y0)) - (z1 * z0)) + (w1 * w0)), ((((x1 * w0) + (y1 * z0)) - (z1 * y0)) + (w1 * x0)), (((((- x1) * z0) + (y1 * w0)) + (z1 * x0)) + (w1 * y0)), ((((x1 * y0) - (y1 * x0)) + (z1 * w0)) + (w1 * z0))], dtype=numpy.float64)
[ "def", "quaternion_multiply", "(", "quaternion1", ",", "quaternion0", ")", ":", "(", "w0", ",", "x0", ",", "y0", ",", "z0", ")", "=", "quaternion0", "(", "w1", ",", "x1", ",", "y1", ",", "z1", ")", "=", "quaternion1", "return", "numpy", ".", "array", "(", "[", "(", "(", "(", "(", "(", "-", "x1", ")", "*", "x0", ")", "-", "(", "y1", "*", "y0", ")", ")", "-", "(", "z1", "*", "z0", ")", ")", "+", "(", "w1", "*", "w0", ")", ")", ",", "(", "(", "(", "(", "x1", "*", "w0", ")", "+", "(", "y1", "*", "z0", ")", ")", "-", "(", "z1", "*", "y0", ")", ")", "+", "(", "w1", "*", "x0", ")", ")", ",", "(", "(", "(", "(", "(", "-", "x1", ")", "*", "z0", ")", "+", "(", "y1", "*", "w0", ")", ")", "+", "(", "z1", "*", "x0", ")", ")", "+", "(", "w1", "*", "y0", ")", ")", ",", "(", "(", "(", "(", "x1", "*", "y0", ")", "-", "(", "y1", "*", "x0", ")", ")", "+", "(", "z1", "*", "w0", ")", ")", "+", "(", "w1", "*", "z0", ")", ")", "]", ",", "dtype", "=", "numpy", ".", "float64", ")" ]
return multiplication of two quaternions .
train
true
42,645
def equateSphericalDotElevation(point, returnValue): radius = abs(point) if (radius <= 0.0): return azimuthComplex = point.dropAxis() azimuthRadius = abs(azimuthComplex) if (azimuthRadius <= 0.0): return elevationComplex = euclidean.getWiddershinsUnitPolar(math.radians(returnValue)) azimuthComplex *= ((radius / azimuthRadius) * elevationComplex.real) point.x = azimuthComplex.real point.y = azimuthComplex.imag point.z = (elevationComplex.imag * radius)
[ "def", "equateSphericalDotElevation", "(", "point", ",", "returnValue", ")", ":", "radius", "=", "abs", "(", "point", ")", "if", "(", "radius", "<=", "0.0", ")", ":", "return", "azimuthComplex", "=", "point", ".", "dropAxis", "(", ")", "azimuthRadius", "=", "abs", "(", "azimuthComplex", ")", "if", "(", "azimuthRadius", "<=", "0.0", ")", ":", "return", "elevationComplex", "=", "euclidean", ".", "getWiddershinsUnitPolar", "(", "math", ".", "radians", "(", "returnValue", ")", ")", "azimuthComplex", "*=", "(", "(", "radius", "/", "azimuthRadius", ")", "*", "elevationComplex", ".", "real", ")", "point", ".", "x", "=", "azimuthComplex", ".", "real", "point", ".", "y", "=", "azimuthComplex", ".", "imag", "point", ".", "z", "=", "(", "elevationComplex", ".", "imag", "*", "radius", ")" ]
get equation for spherical elevation .
train
false
42,647
def get_filterdate(filter_date, date_time): returnvalue = '' date_year = strftime('%Y', gmtime(date_time)) date_month = strftime('%m', gmtime(date_time)) date_day = strftime('%d', gmtime(date_time)) if ((filter_date == 'today') and (int(date_year) == int(localtime()[0])) and (int(date_month) == int(localtime()[1])) and (int(date_day) == int(localtime()[2]))): returnvalue = 'true' elif ((filter_date == 'thismonth') and (date_time >= (time() - 2592000))): returnvalue = 'true' elif ((filter_date == 'thisyear') and (int(date_year) == int(localtime()[0]))): returnvalue = 'true' elif ((filter_date == 'past7days') and (date_time >= (time() - 604800))): returnvalue = 'true' elif (filter_date == ''): returnvalue = 'true' return returnvalue
[ "def", "get_filterdate", "(", "filter_date", ",", "date_time", ")", ":", "returnvalue", "=", "''", "date_year", "=", "strftime", "(", "'%Y'", ",", "gmtime", "(", "date_time", ")", ")", "date_month", "=", "strftime", "(", "'%m'", ",", "gmtime", "(", "date_time", ")", ")", "date_day", "=", "strftime", "(", "'%d'", ",", "gmtime", "(", "date_time", ")", ")", "if", "(", "(", "filter_date", "==", "'today'", ")", "and", "(", "int", "(", "date_year", ")", "==", "int", "(", "localtime", "(", ")", "[", "0", "]", ")", ")", "and", "(", "int", "(", "date_month", ")", "==", "int", "(", "localtime", "(", ")", "[", "1", "]", ")", ")", "and", "(", "int", "(", "date_day", ")", "==", "int", "(", "localtime", "(", ")", "[", "2", "]", ")", ")", ")", ":", "returnvalue", "=", "'true'", "elif", "(", "(", "filter_date", "==", "'thismonth'", ")", "and", "(", "date_time", ">=", "(", "time", "(", ")", "-", "2592000", ")", ")", ")", ":", "returnvalue", "=", "'true'", "elif", "(", "(", "filter_date", "==", "'thisyear'", ")", "and", "(", "int", "(", "date_year", ")", "==", "int", "(", "localtime", "(", ")", "[", "0", "]", ")", ")", ")", ":", "returnvalue", "=", "'true'", "elif", "(", "(", "filter_date", "==", "'past7days'", ")", "and", "(", "date_time", ">=", "(", "time", "(", ")", "-", "604800", ")", ")", ")", ":", "returnvalue", "=", "'true'", "elif", "(", "filter_date", "==", "''", ")", ":", "returnvalue", "=", "'true'", "return", "returnvalue" ]
get filterdate .
train
false
42,648
def graph_menus(): ret = MPMenuSubMenu('Graphs', []) for i in range(len(mestate.graphs)): g = mestate.graphs[i] path = g.name.split('/') name = path[(-1)] path = path[:(-1)] ret.add_to_submenu(path, MPMenuItem(name, name, ('# graph :%u' % i))) return ret
[ "def", "graph_menus", "(", ")", ":", "ret", "=", "MPMenuSubMenu", "(", "'Graphs'", ",", "[", "]", ")", "for", "i", "in", "range", "(", "len", "(", "mestate", ".", "graphs", ")", ")", ":", "g", "=", "mestate", ".", "graphs", "[", "i", "]", "path", "=", "g", ".", "name", ".", "split", "(", "'/'", ")", "name", "=", "path", "[", "(", "-", "1", ")", "]", "path", "=", "path", "[", ":", "(", "-", "1", ")", "]", "ret", ".", "add_to_submenu", "(", "path", ",", "MPMenuItem", "(", "name", ",", "name", ",", "(", "'# graph :%u'", "%", "i", ")", ")", ")", "return", "ret" ]
return menu tree for graphs .
train
true
42,649
def _GetMSBuildPropertyGroup(spec, label, properties): group = ['PropertyGroup'] if label: group.append({'Label': label}) num_configurations = len(spec['configurations']) def GetEdges(node): edges = set() for value in sorted(properties[node].keys()): edges.update(set([v for v in MSVS_VARIABLE_REFERENCE.findall(value) if ((v in properties) and (v != node))])) return edges properties_ordered = gyp.common.TopologicallySorted(properties.keys(), GetEdges) for name in reversed(properties_ordered): values = properties[name] for (value, conditions) in sorted(values.iteritems()): if (len(conditions) == num_configurations): group.append([name, value]) else: for condition in conditions: group.append([name, {'Condition': condition}, value]) return [group]
[ "def", "_GetMSBuildPropertyGroup", "(", "spec", ",", "label", ",", "properties", ")", ":", "group", "=", "[", "'PropertyGroup'", "]", "if", "label", ":", "group", ".", "append", "(", "{", "'Label'", ":", "label", "}", ")", "num_configurations", "=", "len", "(", "spec", "[", "'configurations'", "]", ")", "def", "GetEdges", "(", "node", ")", ":", "edges", "=", "set", "(", ")", "for", "value", "in", "sorted", "(", "properties", "[", "node", "]", ".", "keys", "(", ")", ")", ":", "edges", ".", "update", "(", "set", "(", "[", "v", "for", "v", "in", "MSVS_VARIABLE_REFERENCE", ".", "findall", "(", "value", ")", "if", "(", "(", "v", "in", "properties", ")", "and", "(", "v", "!=", "node", ")", ")", "]", ")", ")", "return", "edges", "properties_ordered", "=", "gyp", ".", "common", ".", "TopologicallySorted", "(", "properties", ".", "keys", "(", ")", ",", "GetEdges", ")", "for", "name", "in", "reversed", "(", "properties_ordered", ")", ":", "values", "=", "properties", "[", "name", "]", "for", "(", "value", ",", "conditions", ")", "in", "sorted", "(", "values", ".", "iteritems", "(", ")", ")", ":", "if", "(", "len", "(", "conditions", ")", "==", "num_configurations", ")", ":", "group", ".", "append", "(", "[", "name", ",", "value", "]", ")", "else", ":", "for", "condition", "in", "conditions", ":", "group", ".", "append", "(", "[", "name", ",", "{", "'Condition'", ":", "condition", "}", ",", "value", "]", ")", "return", "[", "group", "]" ]
returns a propertygroup definition for the specified properties .
train
false
42,652
def add_section(data, label, icon, items): if (not items): return data.append({u'label': label, u'icon': icon, u'items': items})
[ "def", "add_section", "(", "data", ",", "label", ",", "icon", ",", "items", ")", ":", "if", "(", "not", "items", ")", ":", "return", "data", ".", "append", "(", "{", "u'label'", ":", "label", ",", "u'icon'", ":", "icon", ",", "u'items'", ":", "items", "}", ")" ]
adds a section to the module data .
train
false
42,653
def acc_check(expected, got, rel_err=2e-15, abs_err=5e-323): if (math.isinf(expected) and (got == expected)): return None error = (got - expected) permitted_error = max(abs_err, (rel_err * abs(expected))) if (abs(error) < permitted_error): return None return 'error = {}; permitted error = {}'.format(error, permitted_error)
[ "def", "acc_check", "(", "expected", ",", "got", ",", "rel_err", "=", "2e-15", ",", "abs_err", "=", "5e-323", ")", ":", "if", "(", "math", ".", "isinf", "(", "expected", ")", "and", "(", "got", "==", "expected", ")", ")", ":", "return", "None", "error", "=", "(", "got", "-", "expected", ")", "permitted_error", "=", "max", "(", "abs_err", ",", "(", "rel_err", "*", "abs", "(", "expected", ")", ")", ")", "if", "(", "abs", "(", "error", ")", "<", "permitted_error", ")", ":", "return", "None", "return", "'error = {}; permitted error = {}'", ".", "format", "(", "error", ",", "permitted_error", ")" ]
determine whether non-nan floats a and b are equal to within a rounding error .
train
false
42,654
def is_scalar_neg(builder, value): return _scalar_pred_against_zero(builder, value, functools.partial(builder.fcmp_ordered, '<'), '<')
[ "def", "is_scalar_neg", "(", "builder", ",", "value", ")", ":", "return", "_scalar_pred_against_zero", "(", "builder", ",", "value", ",", "functools", ".", "partial", "(", "builder", ".", "fcmp_ordered", ",", "'<'", ")", ",", "'<'", ")" ]
is *value* negative? assumes *value* is signed .
train
false
42,655
def decode_header(header): if hasattr(header, '_chunks'): return [(_charset._encode(string, str(charset)), str(charset)) for (string, charset) in header._chunks] if (not ecre.search(header)): return [(header, None)] words = [] for line in header.splitlines(): parts = ecre.split(line) first = True while parts: unencoded = parts.pop(0) if first: unencoded = unencoded.lstrip() first = False if unencoded: words.append((unencoded, None, None)) if parts: charset = parts.pop(0).lower() encoding = parts.pop(0).lower() encoded = parts.pop(0) words.append((encoded, encoding, charset)) droplist = [] for (n, w) in enumerate(words): if ((n > 1) and w[1] and words[(n - 2)][1] and words[(n - 1)][0].isspace()): droplist.append((n - 1)) for d in reversed(droplist): del words[d] decoded_words = [] for (encoded_string, encoding, charset) in words: if (encoding is None): decoded_words.append((encoded_string, charset)) elif (encoding == 'q'): word = email.quoprimime.header_decode(encoded_string) decoded_words.append((word, charset)) elif (encoding == 'b'): paderr = (len(encoded_string) % 4) if paderr: encoded_string += '==='[:(4 - paderr)] try: word = email.base64mime.decode(encoded_string) except binascii.Error: raise HeaderParseError('Base64 decoding error') else: decoded_words.append((word, charset)) else: raise AssertionError(('Unexpected encoding: ' + encoding)) collapsed = [] last_word = last_charset = None for (word, charset) in decoded_words: if isinstance(word, unicode): word = bytes(word, 'raw-unicode-escape') if (last_word is None): last_word = word last_charset = charset elif (charset != last_charset): collapsed.append((last_word, last_charset)) last_word = word last_charset = charset elif (last_charset is None): last_word += (' ' + word) else: last_word += word collapsed.append((last_word, last_charset)) return collapsed
[ "def", "decode_header", "(", "header", ")", ":", "if", "hasattr", "(", "header", ",", "'_chunks'", ")", ":", "return", "[", "(", "_charset", ".", "_encode", "(", "string", ",", "str", "(", "charset", ")", ")", ",", "str", "(", "charset", ")", ")", "for", "(", "string", ",", "charset", ")", "in", "header", ".", "_chunks", "]", "if", "(", "not", "ecre", ".", "search", "(", "header", ")", ")", ":", "return", "[", "(", "header", ",", "None", ")", "]", "words", "=", "[", "]", "for", "line", "in", "header", ".", "splitlines", "(", ")", ":", "parts", "=", "ecre", ".", "split", "(", "line", ")", "first", "=", "True", "while", "parts", ":", "unencoded", "=", "parts", ".", "pop", "(", "0", ")", "if", "first", ":", "unencoded", "=", "unencoded", ".", "lstrip", "(", ")", "first", "=", "False", "if", "unencoded", ":", "words", ".", "append", "(", "(", "unencoded", ",", "None", ",", "None", ")", ")", "if", "parts", ":", "charset", "=", "parts", ".", "pop", "(", "0", ")", ".", "lower", "(", ")", "encoding", "=", "parts", ".", "pop", "(", "0", ")", ".", "lower", "(", ")", "encoded", "=", "parts", ".", "pop", "(", "0", ")", "words", ".", "append", "(", "(", "encoded", ",", "encoding", ",", "charset", ")", ")", "droplist", "=", "[", "]", "for", "(", "n", ",", "w", ")", "in", "enumerate", "(", "words", ")", ":", "if", "(", "(", "n", ">", "1", ")", "and", "w", "[", "1", "]", "and", "words", "[", "(", "n", "-", "2", ")", "]", "[", "1", "]", "and", "words", "[", "(", "n", "-", "1", ")", "]", "[", "0", "]", ".", "isspace", "(", ")", ")", ":", "droplist", ".", "append", "(", "(", "n", "-", "1", ")", ")", "for", "d", "in", "reversed", "(", "droplist", ")", ":", "del", "words", "[", "d", "]", "decoded_words", "=", "[", "]", "for", "(", "encoded_string", ",", "encoding", ",", "charset", ")", "in", "words", ":", "if", "(", "encoding", "is", "None", ")", ":", "decoded_words", ".", "append", "(", "(", "encoded_string", ",", "charset", ")", ")", "elif", "(", "encoding", "==", "'q'", ")", ":", "word", "=", "email", ".", "quoprimime", ".", "header_decode", "(", "encoded_string", ")", "decoded_words", ".", "append", "(", "(", "word", ",", "charset", ")", ")", "elif", "(", "encoding", "==", "'b'", ")", ":", "paderr", "=", "(", "len", "(", "encoded_string", ")", "%", "4", ")", "if", "paderr", ":", "encoded_string", "+=", "'==='", "[", ":", "(", "4", "-", "paderr", ")", "]", "try", ":", "word", "=", "email", ".", "base64mime", ".", "decode", "(", "encoded_string", ")", "except", "binascii", ".", "Error", ":", "raise", "HeaderParseError", "(", "'Base64 decoding error'", ")", "else", ":", "decoded_words", ".", "append", "(", "(", "word", ",", "charset", ")", ")", "else", ":", "raise", "AssertionError", "(", "(", "'Unexpected encoding: '", "+", "encoding", ")", ")", "collapsed", "=", "[", "]", "last_word", "=", "last_charset", "=", "None", "for", "(", "word", ",", "charset", ")", "in", "decoded_words", ":", "if", "isinstance", "(", "word", ",", "unicode", ")", ":", "word", "=", "bytes", "(", "word", ",", "'raw-unicode-escape'", ")", "if", "(", "last_word", "is", "None", ")", ":", "last_word", "=", "word", "last_charset", "=", "charset", "elif", "(", "charset", "!=", "last_charset", ")", ":", "collapsed", ".", "append", "(", "(", "last_word", ",", "last_charset", ")", ")", "last_word", "=", "word", "last_charset", "=", "charset", "elif", "(", "last_charset", "is", "None", ")", ":", "last_word", "+=", "(", "' '", "+", "word", ")", "else", ":", "last_word", "+=", "word", "collapsed", ".", "append", "(", "(", "last_word", ",", "last_charset", ")", ")", "return", "collapsed" ]
make sure the header is an unicode string .
train
true
42,656
def connected_masters(): port = 4505 config_port = __salt__['config.get']('publish_port') if config_port: port = config_port connected_masters_ips = _remote_port_tcp(port) return connected_masters_ips
[ "def", "connected_masters", "(", ")", ":", "port", "=", "4505", "config_port", "=", "__salt__", "[", "'config.get'", "]", "(", "'publish_port'", ")", "if", "config_port", ":", "port", "=", "config_port", "connected_masters_ips", "=", "_remote_port_tcp", "(", "port", ")", "return", "connected_masters_ips" ]
return current connected masters .
train
false
42,657
def freemem(): return utils.run('sync && echo 3 > /proc/sys/vm/drop_caches')
[ "def", "freemem", "(", ")", ":", "return", "utils", ".", "run", "(", "'sync && echo 3 > /proc/sys/vm/drop_caches'", ")" ]
free useless memoery .
train
false
42,658
def format_index(index, start, end): start -= start.utcoffset() end -= end.utcoffset() indexes = [] while (start.date() <= end.date()): indexes.append(start.strftime(index)) start += datetime.timedelta(days=1) return ','.join(indexes)
[ "def", "format_index", "(", "index", ",", "start", ",", "end", ")", ":", "start", "-=", "start", ".", "utcoffset", "(", ")", "end", "-=", "end", ".", "utcoffset", "(", ")", "indexes", "=", "[", "]", "while", "(", "start", ".", "date", "(", ")", "<=", "end", ".", "date", "(", ")", ")", ":", "indexes", ".", "append", "(", "start", ".", "strftime", "(", "index", ")", ")", "start", "+=", "datetime", ".", "timedelta", "(", "days", "=", "1", ")", "return", "','", ".", "join", "(", "indexes", ")" ]
takes an index .
train
false
42,659
def type_coerce(expression, type_): type_ = type_api.to_instance(type_) if hasattr(expression, u'__clause_element__'): return type_coerce(expression.__clause_element__(), type_) elif isinstance(expression, BindParameter): bp = expression._clone() bp.type = type_ return bp elif (not isinstance(expression, Visitable)): if (expression is None): return Null() else: return literal(expression, type_=type_) else: return Label(None, expression, type_=type_)
[ "def", "type_coerce", "(", "expression", ",", "type_", ")", ":", "type_", "=", "type_api", ".", "to_instance", "(", "type_", ")", "if", "hasattr", "(", "expression", ",", "u'__clause_element__'", ")", ":", "return", "type_coerce", "(", "expression", ".", "__clause_element__", "(", ")", ",", "type_", ")", "elif", "isinstance", "(", "expression", ",", "BindParameter", ")", ":", "bp", "=", "expression", ".", "_clone", "(", ")", "bp", ".", "type", "=", "type_", "return", "bp", "elif", "(", "not", "isinstance", "(", "expression", ",", "Visitable", ")", ")", ":", "if", "(", "expression", "is", "None", ")", ":", "return", "Null", "(", ")", "else", ":", "return", "literal", "(", "expression", ",", "type_", "=", "type_", ")", "else", ":", "return", "Label", "(", "None", ",", "expression", ",", "type_", "=", "type_", ")" ]
associate a sql expression with a particular type .
train
false
42,661
def render_page(request, page, current_language, slug): template_name = get_template_from_request(request, page, no_current_page=True) context = {} context['lang'] = current_language context['current_page'] = page context['has_change_permissions'] = user_can_change_page(request.user, page) context['has_view_permissions'] = user_can_view_page(request.user, page) if (not context['has_view_permissions']): return _handle_no_page(request, slug) response = TemplateResponse(request, template_name, context) response.add_post_render_callback(set_page_cache) xframe_options = page.get_xframe_options() if ((xframe_options == Page.X_FRAME_OPTIONS_INHERIT) or (xframe_options is None)): return response response.xframe_options_exempt = True if (xframe_options == Page.X_FRAME_OPTIONS_ALLOW): return response elif (xframe_options == Page.X_FRAME_OPTIONS_SAMEORIGIN): response['X-Frame-Options'] = 'SAMEORIGIN' elif (xframe_options == Page.X_FRAME_OPTIONS_DENY): response['X-Frame-Options'] = 'DENY' return response
[ "def", "render_page", "(", "request", ",", "page", ",", "current_language", ",", "slug", ")", ":", "template_name", "=", "get_template_from_request", "(", "request", ",", "page", ",", "no_current_page", "=", "True", ")", "context", "=", "{", "}", "context", "[", "'lang'", "]", "=", "current_language", "context", "[", "'current_page'", "]", "=", "page", "context", "[", "'has_change_permissions'", "]", "=", "user_can_change_page", "(", "request", ".", "user", ",", "page", ")", "context", "[", "'has_view_permissions'", "]", "=", "user_can_view_page", "(", "request", ".", "user", ",", "page", ")", "if", "(", "not", "context", "[", "'has_view_permissions'", "]", ")", ":", "return", "_handle_no_page", "(", "request", ",", "slug", ")", "response", "=", "TemplateResponse", "(", "request", ",", "template_name", ",", "context", ")", "response", ".", "add_post_render_callback", "(", "set_page_cache", ")", "xframe_options", "=", "page", ".", "get_xframe_options", "(", ")", "if", "(", "(", "xframe_options", "==", "Page", ".", "X_FRAME_OPTIONS_INHERIT", ")", "or", "(", "xframe_options", "is", "None", ")", ")", ":", "return", "response", "response", ".", "xframe_options_exempt", "=", "True", "if", "(", "xframe_options", "==", "Page", ".", "X_FRAME_OPTIONS_ALLOW", ")", ":", "return", "response", "elif", "(", "xframe_options", "==", "Page", ".", "X_FRAME_OPTIONS_SAMEORIGIN", ")", ":", "response", "[", "'X-Frame-Options'", "]", "=", "'SAMEORIGIN'", "elif", "(", "xframe_options", "==", "Page", ".", "X_FRAME_OPTIONS_DENY", ")", ":", "response", "[", "'X-Frame-Options'", "]", "=", "'DENY'", "return", "response" ]
renders a page .
train
false
42,662
def sew_messages_and_reactions(messages, reactions): for message in messages: message['reactions'] = [] converted_messages = {message['id']: message for message in messages} for reaction in reactions: converted_messages[reaction['message_id']]['reactions'].append(reaction) return list(converted_messages.values())
[ "def", "sew_messages_and_reactions", "(", "messages", ",", "reactions", ")", ":", "for", "message", "in", "messages", ":", "message", "[", "'reactions'", "]", "=", "[", "]", "converted_messages", "=", "{", "message", "[", "'id'", "]", ":", "message", "for", "message", "in", "messages", "}", "for", "reaction", "in", "reactions", ":", "converted_messages", "[", "reaction", "[", "'message_id'", "]", "]", "[", "'reactions'", "]", ".", "append", "(", "reaction", ")", "return", "list", "(", "converted_messages", ".", "values", "(", ")", ")" ]
given a iterable of messages and reactions stitch reactions into messages .
train
false
42,663
def describe_file_set(modules): descriptor = FileSet() file_descriptors = [] for module in modules: file_descriptors.append(describe_file(module)) if file_descriptors: descriptor.files = file_descriptors return descriptor
[ "def", "describe_file_set", "(", "modules", ")", ":", "descriptor", "=", "FileSet", "(", ")", "file_descriptors", "=", "[", "]", "for", "module", "in", "modules", ":", "file_descriptors", ".", "append", "(", "describe_file", "(", "module", ")", ")", "if", "file_descriptors", ":", "descriptor", ".", "files", "=", "file_descriptors", "return", "descriptor" ]
build a file set from a specified python modules .
train
true