id_within_dataset
int64
1
55.5k
snippet
stringlengths
19
14.2k
tokens
listlengths
6
1.63k
nl
stringlengths
6
352
split_within_dataset
stringclasses
1 value
is_duplicated
bool
2 classes
16,300
def descriptor_global_local_resource_url(block, uri): raise NotImplementedError('Applications must monkey-patch this function before using local_resource_url for studio_view')
[ "def", "descriptor_global_local_resource_url", "(", "block", ",", "uri", ")", ":", "raise", "NotImplementedError", "(", "'Applications must monkey-patch this function before using local_resource_url for studio_view'", ")" ]
see :meth:xblock .
train
false
16,301
def block_device_mapping_get_all_by_volume_id(context, volume_id, columns_to_join=None): return IMPL.block_device_mapping_get_all_by_volume_id(context, volume_id, columns_to_join)
[ "def", "block_device_mapping_get_all_by_volume_id", "(", "context", ",", "volume_id", ",", "columns_to_join", "=", "None", ")", ":", "return", "IMPL", ".", "block_device_mapping_get_all_by_volume_id", "(", "context", ",", "volume_id", ",", "columns_to_join", ")" ]
get block device mapping for a given volume .
train
false
16,302
def has_exec(cmd): return (which(cmd) is not None)
[ "def", "has_exec", "(", "cmd", ")", ":", "return", "(", "which", "(", "cmd", ")", "is", "not", "None", ")" ]
returns true if the executable is available on the minion .
train
false
16,303
def get_meta_entry(dist, name): meta = get_dist_meta(dist) return meta.get(name)
[ "def", "get_meta_entry", "(", "dist", ",", "name", ")", ":", "meta", "=", "get_dist_meta", "(", "dist", ")", "return", "meta", ".", "get", "(", "name", ")" ]
get the contents of the named entry from the distributions pkg-info file .
train
false
16,305
@utils.positional(1) def transaction_async(callback, **ctx_options): from . import tasklets return tasklets.get_context().transaction(callback, **ctx_options)
[ "@", "utils", ".", "positional", "(", "1", ")", "def", "transaction_async", "(", "callback", ",", "**", "ctx_options", ")", ":", "from", ".", "import", "tasklets", "return", "tasklets", ".", "get_context", "(", ")", ".", "transaction", "(", "callback", ",", "**", "ctx_options", ")" ]
run a callback in a transaction .
train
false
16,306
def htmldiff_tokens(html1_tokens, html2_tokens): s = InsensitiveSequenceMatcher(a=html1_tokens, b=html2_tokens) commands = s.get_opcodes() result = [] for (command, i1, i2, j1, j2) in commands: if (command == 'equal'): result.extend(expand_tokens(html2_tokens[j1:j2], equal=True)) continue if ((command == 'insert') or (command == 'replace')): ins_tokens = expand_tokens(html2_tokens[j1:j2]) merge_insert(ins_tokens, result) if ((command == 'delete') or (command == 'replace')): del_tokens = expand_tokens(html1_tokens[i1:i2]) merge_delete(del_tokens, result) result = cleanup_delete(result) return result
[ "def", "htmldiff_tokens", "(", "html1_tokens", ",", "html2_tokens", ")", ":", "s", "=", "InsensitiveSequenceMatcher", "(", "a", "=", "html1_tokens", ",", "b", "=", "html2_tokens", ")", "commands", "=", "s", ".", "get_opcodes", "(", ")", "result", "=", "[", "]", "for", "(", "command", ",", "i1", ",", "i2", ",", "j1", ",", "j2", ")", "in", "commands", ":", "if", "(", "command", "==", "'equal'", ")", ":", "result", ".", "extend", "(", "expand_tokens", "(", "html2_tokens", "[", "j1", ":", "j2", "]", ",", "equal", "=", "True", ")", ")", "continue", "if", "(", "(", "command", "==", "'insert'", ")", "or", "(", "command", "==", "'replace'", ")", ")", ":", "ins_tokens", "=", "expand_tokens", "(", "html2_tokens", "[", "j1", ":", "j2", "]", ")", "merge_insert", "(", "ins_tokens", ",", "result", ")", "if", "(", "(", "command", "==", "'delete'", ")", "or", "(", "command", "==", "'replace'", ")", ")", ":", "del_tokens", "=", "expand_tokens", "(", "html1_tokens", "[", "i1", ":", "i2", "]", ")", "merge_delete", "(", "del_tokens", ",", "result", ")", "result", "=", "cleanup_delete", "(", "result", ")", "return", "result" ]
does a diff on the tokens themselves .
train
true
16,307
def __get_size(conn, vm_): size = config.get_cloud_config_value('size', vm_, __opts__, default='n1-standard-1', search_global=False) return conn.ex_get_size(size, __get_location(conn, vm_))
[ "def", "__get_size", "(", "conn", ",", "vm_", ")", ":", "size", "=", "config", ".", "get_cloud_config_value", "(", "'size'", ",", "vm_", ",", "__opts__", ",", "default", "=", "'n1-standard-1'", ",", "search_global", "=", "False", ")", "return", "conn", ".", "ex_get_size", "(", "size", ",", "__get_location", "(", "conn", ",", "vm_", ")", ")" ]
need to override libcloud to find the machine type in the proper zone .
train
true
16,308
def _get_admin_info(command, host=None, core_name=None): url = _format_url('admin/{0}'.format(command), host, core_name=core_name) resp = _http_request(url) return resp
[ "def", "_get_admin_info", "(", "command", ",", "host", "=", "None", ",", "core_name", "=", "None", ")", ":", "url", "=", "_format_url", "(", "'admin/{0}'", ".", "format", "(", "command", ")", ",", "host", ",", "core_name", "=", "core_name", ")", "resp", "=", "_http_request", "(", "url", ")", "return", "resp" ]
private method calls the _http_request method and passes the admin command to execute and stores the data .
train
true
16,309
def Deserializer(stream_or_string, **options): if isinstance(stream_or_string, basestring): stream = StringIO(stream_or_string) else: stream = stream_or_string for obj in PythonDeserializer(yaml.load(stream), **options): (yield obj)
[ "def", "Deserializer", "(", "stream_or_string", ",", "**", "options", ")", ":", "if", "isinstance", "(", "stream_or_string", ",", "basestring", ")", ":", "stream", "=", "StringIO", "(", "stream_or_string", ")", "else", ":", "stream", "=", "stream_or_string", "for", "obj", "in", "PythonDeserializer", "(", "yaml", ".", "load", "(", "stream", ")", ",", "**", "options", ")", ":", "(", "yield", "obj", ")" ]
deserialize a stream or string of yaml data .
train
false
16,310
def generateCookieSecret(): return base64.b64encode((uuid.uuid4().bytes + uuid.uuid4().bytes))
[ "def", "generateCookieSecret", "(", ")", ":", "return", "base64", ".", "b64encode", "(", "(", "uuid", ".", "uuid4", "(", ")", ".", "bytes", "+", "uuid", ".", "uuid4", "(", ")", ".", "bytes", ")", ")" ]
generate a new cookie secret .
train
false
16,311
def check_state_result(running, recurse=False): if (not isinstance(running, dict)): return False if (not running): return False ret = True for state_result in six.itervalues(running): if ((not recurse) and (not isinstance(state_result, dict))): ret = False if (ret and isinstance(state_result, dict)): result = state_result.get('result', _empty) if (result is False): ret = False elif ((result is _empty) and isinstance(state_result, dict) and ret): ret = check_state_result(state_result, recurse=True) if (not ret): break return ret
[ "def", "check_state_result", "(", "running", ",", "recurse", "=", "False", ")", ":", "if", "(", "not", "isinstance", "(", "running", ",", "dict", ")", ")", ":", "return", "False", "if", "(", "not", "running", ")", ":", "return", "False", "ret", "=", "True", "for", "state_result", "in", "six", ".", "itervalues", "(", "running", ")", ":", "if", "(", "(", "not", "recurse", ")", "and", "(", "not", "isinstance", "(", "state_result", ",", "dict", ")", ")", ")", ":", "ret", "=", "False", "if", "(", "ret", "and", "isinstance", "(", "state_result", ",", "dict", ")", ")", ":", "result", "=", "state_result", ".", "get", "(", "'result'", ",", "_empty", ")", "if", "(", "result", "is", "False", ")", ":", "ret", "=", "False", "elif", "(", "(", "result", "is", "_empty", ")", "and", "isinstance", "(", "state_result", ",", "dict", ")", "and", "ret", ")", ":", "ret", "=", "check_state_result", "(", "state_result", ",", "recurse", "=", "True", ")", "if", "(", "not", "ret", ")", ":", "break", "return", "ret" ]
check the total return value of the run and determine if the running dict has any issues .
train
false
16,312
def _set_msg_reply(msg_reply): def _set_cls_msg_reply(cls): cls.cls_msg_reply = msg_reply return cls return _set_cls_msg_reply
[ "def", "_set_msg_reply", "(", "msg_reply", ")", ":", "def", "_set_cls_msg_reply", "(", "cls", ")", ":", "cls", ".", "cls_msg_reply", "=", "msg_reply", "return", "cls", "return", "_set_cls_msg_reply" ]
annotate ofp reply message class .
train
false
16,313
def osf_storage_root(addon_config, node_settings, auth, **kwargs): node = node_settings.owner root = rubeus.build_addon_root(node_settings=node_settings, name=u'', permissions=auth, user=auth.user, nodeUrl=node.url, nodeApiUrl=node.api_url) return [root]
[ "def", "osf_storage_root", "(", "addon_config", ",", "node_settings", ",", "auth", ",", "**", "kwargs", ")", ":", "node", "=", "node_settings", ".", "owner", "root", "=", "rubeus", ".", "build_addon_root", "(", "node_settings", "=", "node_settings", ",", "name", "=", "u''", ",", "permissions", "=", "auth", ",", "user", "=", "auth", ".", "user", ",", "nodeUrl", "=", "node", ".", "url", ",", "nodeApiUrl", "=", "node", ".", "api_url", ")", "return", "[", "root", "]" ]
build hgrid json for root node .
train
false
16,314
def p_item_string_expr(p): p[0] = (p[1][1:(-1)], p[2])
[ "def", "p_item_string_expr", "(", "p", ")", ":", "p", "[", "0", "]", "=", "(", "p", "[", "1", "]", "[", "1", ":", "(", "-", "1", ")", "]", ",", "p", "[", "2", "]", ")" ]
pitem : string expr .
train
false
16,316
def random_variables(): return tf.get_collection(RANDOM_VARIABLE_COLLECTION)
[ "def", "random_variables", "(", ")", ":", "return", "tf", ".", "get_collection", "(", "RANDOM_VARIABLE_COLLECTION", ")" ]
return all random variables in the tensorflow graph .
train
false
16,317
def OSVersion(flavor): urlbase = path.basename(isoURLs.get(flavor, 'unknown')) return path.splitext(urlbase)[0]
[ "def", "OSVersion", "(", "flavor", ")", ":", "urlbase", "=", "path", ".", "basename", "(", "isoURLs", ".", "get", "(", "flavor", ",", "'unknown'", ")", ")", "return", "path", ".", "splitext", "(", "urlbase", ")", "[", "0", "]" ]
return full os version string for build flavor .
train
false
16,319
def IndexXmlForQuery(kind, ancestor, props): serialized_xml = [] serialized_xml.append((' <datastore-index kind="%s" ancestor="%s">' % (kind, ('true' if ancestor else 'false')))) for (name, direction) in props: serialized_xml.append((' <property name="%s" direction="%s" />' % (name, ('asc' if (direction == ASCENDING) else 'desc')))) serialized_xml.append(' </datastore-index>') return '\n'.join(serialized_xml)
[ "def", "IndexXmlForQuery", "(", "kind", ",", "ancestor", ",", "props", ")", ":", "serialized_xml", "=", "[", "]", "serialized_xml", ".", "append", "(", "(", "' <datastore-index kind=\"%s\" ancestor=\"%s\">'", "%", "(", "kind", ",", "(", "'true'", "if", "ancestor", "else", "'false'", ")", ")", ")", ")", "for", "(", "name", ",", "direction", ")", "in", "props", ":", "serialized_xml", ".", "append", "(", "(", "' <property name=\"%s\" direction=\"%s\" />'", "%", "(", "name", ",", "(", "'asc'", "if", "(", "direction", "==", "ASCENDING", ")", "else", "'desc'", ")", ")", ")", ")", "serialized_xml", ".", "append", "(", "' </datastore-index>'", ")", "return", "'\\n'", ".", "join", "(", "serialized_xml", ")" ]
return the composite index definition xml needed for a query .
train
false
16,320
def xframe_allow_whitelisted(view_func): def wrapped_view(request, *args, **kwargs): ' Modify the response with the correct X-Frame-Options. ' resp = view_func(request, *args, **kwargs) x_frame_option = 'DENY' if settings.FEATURES['ENABLE_THIRD_PARTY_AUTH']: referer = request.META.get('HTTP_REFERER') if (referer is not None): parsed_url = urlparse(referer) hostname = parsed_url.hostname if LTIProviderConfig.objects.current_set().filter(lti_hostname=hostname, enabled=True).exists(): x_frame_option = 'ALLOW' resp['X-Frame-Options'] = x_frame_option return resp return wraps(view_func, assigned=available_attrs(view_func))(wrapped_view)
[ "def", "xframe_allow_whitelisted", "(", "view_func", ")", ":", "def", "wrapped_view", "(", "request", ",", "*", "args", ",", "**", "kwargs", ")", ":", "resp", "=", "view_func", "(", "request", ",", "*", "args", ",", "**", "kwargs", ")", "x_frame_option", "=", "'DENY'", "if", "settings", ".", "FEATURES", "[", "'ENABLE_THIRD_PARTY_AUTH'", "]", ":", "referer", "=", "request", ".", "META", ".", "get", "(", "'HTTP_REFERER'", ")", "if", "(", "referer", "is", "not", "None", ")", ":", "parsed_url", "=", "urlparse", "(", "referer", ")", "hostname", "=", "parsed_url", ".", "hostname", "if", "LTIProviderConfig", ".", "objects", ".", "current_set", "(", ")", ".", "filter", "(", "lti_hostname", "=", "hostname", ",", "enabled", "=", "True", ")", ".", "exists", "(", ")", ":", "x_frame_option", "=", "'ALLOW'", "resp", "[", "'X-Frame-Options'", "]", "=", "x_frame_option", "return", "resp", "return", "wraps", "(", "view_func", ",", "assigned", "=", "available_attrs", "(", "view_func", ")", ")", "(", "wrapped_view", ")" ]
modifies a view function so that its response has the x-frame-options http header set to deny if the request http referrer is not from a whitelisted hostname .
train
false
16,322
def test_no_data_with_no_values_with_include_x_axis(Chart): chart = Chart(include_x_axis=True) q = chart.render_pyquery() assert (q('.text-overlay text').text() == 'No data')
[ "def", "test_no_data_with_no_values_with_include_x_axis", "(", "Chart", ")", ":", "chart", "=", "Chart", "(", "include_x_axis", "=", "True", ")", "q", "=", "chart", ".", "render_pyquery", "(", ")", "assert", "(", "q", "(", "'.text-overlay text'", ")", ".", "text", "(", ")", "==", "'No data'", ")" ]
test no data and include_x_axis .
train
false
16,323
@register_stabilize @register_canonicalize @local_optimizer([Solve]) def tag_solve_triangular(node): if (node.op == solve): if (node.op.A_structure == 'general'): (A, b) = node.inputs if (A.owner and isinstance(A.owner.op, type(cholesky))): if A.owner.op.lower: return [Solve('lower_triangular')(A, b)] else: return [Solve('upper_triangular')(A, b)] if (A.owner and isinstance(A.owner.op, DimShuffle) and (A.owner.op.new_order == (1, 0))): (A_T,) = A.owner.inputs if (A_T.owner and isinstance(A_T.owner.op, type(cholesky))): if A_T.owner.op.lower: return [Solve('upper_triangular')(A, b)] else: return [Solve('lower_triangular')(A, b)]
[ "@", "register_stabilize", "@", "register_canonicalize", "@", "local_optimizer", "(", "[", "Solve", "]", ")", "def", "tag_solve_triangular", "(", "node", ")", ":", "if", "(", "node", ".", "op", "==", "solve", ")", ":", "if", "(", "node", ".", "op", ".", "A_structure", "==", "'general'", ")", ":", "(", "A", ",", "b", ")", "=", "node", ".", "inputs", "if", "(", "A", ".", "owner", "and", "isinstance", "(", "A", ".", "owner", ".", "op", ",", "type", "(", "cholesky", ")", ")", ")", ":", "if", "A", ".", "owner", ".", "op", ".", "lower", ":", "return", "[", "Solve", "(", "'lower_triangular'", ")", "(", "A", ",", "b", ")", "]", "else", ":", "return", "[", "Solve", "(", "'upper_triangular'", ")", "(", "A", ",", "b", ")", "]", "if", "(", "A", ".", "owner", "and", "isinstance", "(", "A", ".", "owner", ".", "op", ",", "DimShuffle", ")", "and", "(", "A", ".", "owner", ".", "op", ".", "new_order", "==", "(", "1", ",", "0", ")", ")", ")", ":", "(", "A_T", ",", ")", "=", "A", ".", "owner", ".", "inputs", "if", "(", "A_T", ".", "owner", "and", "isinstance", "(", "A_T", ".", "owner", ".", "op", ",", "type", "(", "cholesky", ")", ")", ")", ":", "if", "A_T", ".", "owner", ".", "op", ".", "lower", ":", "return", "[", "Solve", "(", "'upper_triangular'", ")", "(", "A", ",", "b", ")", "]", "else", ":", "return", "[", "Solve", "(", "'lower_triangular'", ")", "(", "A", ",", "b", ")", "]" ]
if a general solve() is applied to the output of a cholesky op .
train
false
16,324
def evaluation_data(): return s3_rest_controller()
[ "def", "evaluation_data", "(", ")", ":", "return", "s3_rest_controller", "(", ")" ]
restful crud controller .
train
false
16,326
def vserver_servicegroup_delete(v_name, sg_name, **connection_args): ret = True if (not vserver_servicegroup_exists(v_name, sg_name, **connection_args)): return False nitro = _connect(**connection_args) if (nitro is None): return False vsg = NSLBVServerServiceGroupBinding() vsg.set_name(v_name) vsg.set_servicegroupname(sg_name) try: NSLBVServerServiceGroupBinding.delete(nitro, vsg) except NSNitroError as error: log.debug('netscaler module error - NSLBVServerServiceGroupBinding.delete() failed: {0}'.format(error)) ret = False _disconnect(nitro) return ret
[ "def", "vserver_servicegroup_delete", "(", "v_name", ",", "sg_name", ",", "**", "connection_args", ")", ":", "ret", "=", "True", "if", "(", "not", "vserver_servicegroup_exists", "(", "v_name", ",", "sg_name", ",", "**", "connection_args", ")", ")", ":", "return", "False", "nitro", "=", "_connect", "(", "**", "connection_args", ")", "if", "(", "nitro", "is", "None", ")", ":", "return", "False", "vsg", "=", "NSLBVServerServiceGroupBinding", "(", ")", "vsg", ".", "set_name", "(", "v_name", ")", "vsg", ".", "set_servicegroupname", "(", "sg_name", ")", "try", ":", "NSLBVServerServiceGroupBinding", ".", "delete", "(", "nitro", ",", "vsg", ")", "except", "NSNitroError", "as", "error", ":", "log", ".", "debug", "(", "'netscaler module error - NSLBVServerServiceGroupBinding.delete() failed: {0}'", ".", "format", "(", "error", ")", ")", "ret", "=", "False", "_disconnect", "(", "nitro", ")", "return", "ret" ]
unbind a servicegroup from a vserver cli example: .
train
true
16,328
def matchFirst(string, *args): for patternlist in args: for pattern in patternlist: r = pattern.search(string) if (r is not None): name = r.group(1) return name return string
[ "def", "matchFirst", "(", "string", ",", "*", "args", ")", ":", "for", "patternlist", "in", "args", ":", "for", "pattern", "in", "patternlist", ":", "r", "=", "pattern", ".", "search", "(", "string", ")", "if", "(", "r", "is", "not", "None", ")", ":", "name", "=", "r", ".", "group", "(", "1", ")", "return", "name", "return", "string" ]
matches against list of regexp and returns first match .
train
false
16,329
def write_bem_surfaces(fname, surfs): if isinstance(surfs, dict): surfs = [surfs] with start_file(fname) as fid: start_block(fid, FIFF.FIFFB_BEM) write_int(fid, FIFF.FIFF_BEM_COORD_FRAME, surfs[0]['coord_frame']) _write_bem_surfaces_block(fid, surfs) end_block(fid, FIFF.FIFFB_BEM) end_file(fid)
[ "def", "write_bem_surfaces", "(", "fname", ",", "surfs", ")", ":", "if", "isinstance", "(", "surfs", ",", "dict", ")", ":", "surfs", "=", "[", "surfs", "]", "with", "start_file", "(", "fname", ")", "as", "fid", ":", "start_block", "(", "fid", ",", "FIFF", ".", "FIFFB_BEM", ")", "write_int", "(", "fid", ",", "FIFF", ".", "FIFF_BEM_COORD_FRAME", ",", "surfs", "[", "0", "]", "[", "'coord_frame'", "]", ")", "_write_bem_surfaces_block", "(", "fid", ",", "surfs", ")", "end_block", "(", "fid", ",", "FIFF", ".", "FIFFB_BEM", ")", "end_file", "(", "fid", ")" ]
write bem surfaces to a fiff file .
train
false
16,330
def disconnected_grad(x): return disconnected_grad_(x)
[ "def", "disconnected_grad", "(", "x", ")", ":", "return", "disconnected_grad_", "(", "x", ")" ]
consider an expression constant when computing gradients .
train
false
16,332
def makedirs_(path, user=None, group=None, mode=None): path = os.path.expanduser(path) dirname = os.path.normpath(os.path.dirname(path)) if os.path.isdir(dirname): msg = "Directory '{0}' already exists".format(dirname) log.debug(msg) return msg if os.path.exists(dirname): msg = "The path '{0}' already exists and is not a directory".format(dirname) log.debug(msg) return msg directories_to_create = [] while True: if os.path.isdir(dirname): break directories_to_create.append(dirname) current_dirname = dirname dirname = os.path.dirname(dirname) if (current_dirname == dirname): raise SaltInvocationError("Recursive creation for path '{0}' would result in an infinite loop. Please use an absolute path.".format(dirname)) directories_to_create.reverse() for directory_to_create in directories_to_create: log.debug('Creating directory: %s', directory_to_create) mkdir(directory_to_create, user=user, group=group, mode=mode)
[ "def", "makedirs_", "(", "path", ",", "user", "=", "None", ",", "group", "=", "None", ",", "mode", "=", "None", ")", ":", "path", "=", "os", ".", "path", ".", "expanduser", "(", "path", ")", "dirname", "=", "os", ".", "path", ".", "normpath", "(", "os", ".", "path", ".", "dirname", "(", "path", ")", ")", "if", "os", ".", "path", ".", "isdir", "(", "dirname", ")", ":", "msg", "=", "\"Directory '{0}' already exists\"", ".", "format", "(", "dirname", ")", "log", ".", "debug", "(", "msg", ")", "return", "msg", "if", "os", ".", "path", ".", "exists", "(", "dirname", ")", ":", "msg", "=", "\"The path '{0}' already exists and is not a directory\"", ".", "format", "(", "dirname", ")", "log", ".", "debug", "(", "msg", ")", "return", "msg", "directories_to_create", "=", "[", "]", "while", "True", ":", "if", "os", ".", "path", ".", "isdir", "(", "dirname", ")", ":", "break", "directories_to_create", ".", "append", "(", "dirname", ")", "current_dirname", "=", "dirname", "dirname", "=", "os", ".", "path", ".", "dirname", "(", "dirname", ")", "if", "(", "current_dirname", "==", "dirname", ")", ":", "raise", "SaltInvocationError", "(", "\"Recursive creation for path '{0}' would result in an infinite loop. Please use an absolute path.\"", ".", "format", "(", "dirname", ")", ")", "directories_to_create", ".", "reverse", "(", ")", "for", "directory_to_create", "in", "directories_to_create", ":", "log", ".", "debug", "(", "'Creating directory: %s'", ",", "directory_to_create", ")", "mkdir", "(", "directory_to_create", ",", "user", "=", "user", ",", "group", "=", "group", ",", "mode", "=", "mode", ")" ]
ensure that the directory containing this path is available .
train
true
16,333
def missing_node_cache(prov_dir, node_list, provider, opts): cached_nodes = [] for node in os.listdir(prov_dir): cached_nodes.append(os.path.splitext(node)[0]) for node in cached_nodes: if (node not in node_list): delete_minion_cachedir(node, provider, opts) if (('diff_cache_events' in opts) and opts['diff_cache_events']): fire_event('event', 'cached node missing from provider', 'salt/cloud/{0}/cache_node_missing'.format(node), args={'missing node': node}, sock_dir=opts.get('sock_dir', os.path.join(__opts__['sock_dir'], 'master')), transport=opts.get('transport', 'zeromq'))
[ "def", "missing_node_cache", "(", "prov_dir", ",", "node_list", ",", "provider", ",", "opts", ")", ":", "cached_nodes", "=", "[", "]", "for", "node", "in", "os", ".", "listdir", "(", "prov_dir", ")", ":", "cached_nodes", ".", "append", "(", "os", ".", "path", ".", "splitext", "(", "node", ")", "[", "0", "]", ")", "for", "node", "in", "cached_nodes", ":", "if", "(", "node", "not", "in", "node_list", ")", ":", "delete_minion_cachedir", "(", "node", ",", "provider", ",", "opts", ")", "if", "(", "(", "'diff_cache_events'", "in", "opts", ")", "and", "opts", "[", "'diff_cache_events'", "]", ")", ":", "fire_event", "(", "'event'", ",", "'cached node missing from provider'", ",", "'salt/cloud/{0}/cache_node_missing'", ".", "format", "(", "node", ")", ",", "args", "=", "{", "'missing node'", ":", "node", "}", ",", "sock_dir", "=", "opts", ".", "get", "(", "'sock_dir'", ",", "os", ".", "path", ".", "join", "(", "__opts__", "[", "'sock_dir'", "]", ",", "'master'", ")", ")", ",", "transport", "=", "opts", ".", "get", "(", "'transport'", ",", "'zeromq'", ")", ")" ]
check list of nodes to see if any nodes which were previously known about in the cache have been removed from the node list .
train
true
16,335
def checked(response): status_code = int(response['status'][:3]) if (status_code == 401): raise Unauthorized if (status_code == 406): raise NoSession if (status_code == 407): raise DownloadLimitReached if (status_code == 413): raise InvalidImdbid if (status_code == 414): raise UnknownUserAgent if (status_code == 415): raise DisabledUserAgent if (status_code == 503): raise ServiceUnavailable if (status_code != 200): raise OpenSubtitlesError(response['status']) return response
[ "def", "checked", "(", "response", ")", ":", "status_code", "=", "int", "(", "response", "[", "'status'", "]", "[", ":", "3", "]", ")", "if", "(", "status_code", "==", "401", ")", ":", "raise", "Unauthorized", "if", "(", "status_code", "==", "406", ")", ":", "raise", "NoSession", "if", "(", "status_code", "==", "407", ")", ":", "raise", "DownloadLimitReached", "if", "(", "status_code", "==", "413", ")", ":", "raise", "InvalidImdbid", "if", "(", "status_code", "==", "414", ")", ":", "raise", "UnknownUserAgent", "if", "(", "status_code", "==", "415", ")", ":", "raise", "DisabledUserAgent", "if", "(", "status_code", "==", "503", ")", ":", "raise", "ServiceUnavailable", "if", "(", "status_code", "!=", "200", ")", ":", "raise", "OpenSubtitlesError", "(", "response", "[", "'status'", "]", ")", "return", "response" ]
check a response status before returning it .
train
true
16,336
def queries(): out = {} for plugin in find_plugins(): out.update(plugin.queries()) return out
[ "def", "queries", "(", ")", ":", "out", "=", "{", "}", "for", "plugin", "in", "find_plugins", "(", ")", ":", "out", ".", "update", "(", "plugin", ".", "queries", "(", ")", ")", "return", "out" ]
returns a dict mapping prefix strings to query subclasses all loaded plugins .
train
false
16,337
def format_display_date(date): return date.strftime(DISPLAY_DATE_FORMAT)
[ "def", "format_display_date", "(", "date", ")", ":", "return", "date", ".", "strftime", "(", "DISPLAY_DATE_FORMAT", ")" ]
returns a formatted date string meant for cli output .
train
false
16,339
def get_wms(version='1.1.1', type_name=None): url = (GEOSERVER_URL + ('%s/wms?request=getcapabilities' % type_name.replace(':', '/'))) return WebMapService(url, version=version, username=GEOSERVER_USER, password=GEOSERVER_PASSWD)
[ "def", "get_wms", "(", "version", "=", "'1.1.1'", ",", "type_name", "=", "None", ")", ":", "url", "=", "(", "GEOSERVER_URL", "+", "(", "'%s/wms?request=getcapabilities'", "%", "type_name", ".", "replace", "(", "':'", ",", "'/'", ")", ")", ")", "return", "WebMapService", "(", "url", ",", "version", "=", "version", ",", "username", "=", "GEOSERVER_USER", ",", "password", "=", "GEOSERVER_PASSWD", ")" ]
function to return an owslib wms object .
train
false
16,340
def get_extension(extension_id): for extension in get_extensions(): if (extension.get_id() == extension_id): return extension return None
[ "def", "get_extension", "(", "extension_id", ")", ":", "for", "extension", "in", "get_extensions", "(", ")", ":", "if", "(", "extension", ".", "get_id", "(", ")", "==", "extension_id", ")", ":", "return", "extension", "return", "None" ]
get the extension of a given url .
train
false
16,341
@handle_response_format @treeio_login_required def index_assets(request, response_format='html'): if request.GET: query = _get_filter_query(Asset, request.GET) else: query = Q() filters = AssetFilterForm(request.user.profile, 'title', request.GET) assets = Object.filter_by_request(request, Asset.objects.filter(query), mode='r') return render_to_response('finance/index_assets', {'assets': assets, 'filters': filters}, context_instance=RequestContext(request), response_format=response_format)
[ "@", "handle_response_format", "@", "treeio_login_required", "def", "index_assets", "(", "request", ",", "response_format", "=", "'html'", ")", ":", "if", "request", ".", "GET", ":", "query", "=", "_get_filter_query", "(", "Asset", ",", "request", ".", "GET", ")", "else", ":", "query", "=", "Q", "(", ")", "filters", "=", "AssetFilterForm", "(", "request", ".", "user", ".", "profile", ",", "'title'", ",", "request", ".", "GET", ")", "assets", "=", "Object", ".", "filter_by_request", "(", "request", ",", "Asset", ".", "objects", ".", "filter", "(", "query", ")", ",", "mode", "=", "'r'", ")", "return", "render_to_response", "(", "'finance/index_assets'", ",", "{", "'assets'", ":", "assets", ",", "'filters'", ":", "filters", "}", ",", "context_instance", "=", "RequestContext", "(", "request", ")", ",", "response_format", "=", "response_format", ")" ]
index_assets page: displays all assets .
train
false
16,342
def _init_stylesheet(profile): old_script = profile.scripts().findScript('_qute_stylesheet') if (not old_script.isNull()): profile.scripts().remove(old_script) css = shared.get_user_stylesheet() source = "\n (function() {{\n var css = document.createElement('style');\n css.setAttribute('type', 'text/css');\n css.appendChild(document.createTextNode('{}'));\n document.getElementsByTagName('head')[0].appendChild(css);\n }})()\n ".format(javascript.string_escape(css)) script = QWebEngineScript() script.setName('_qute_stylesheet') script.setInjectionPoint(QWebEngineScript.DocumentReady) script.setWorldId(QWebEngineScript.ApplicationWorld) script.setRunsOnSubFrames(True) script.setSourceCode(source) profile.scripts().insert(script)
[ "def", "_init_stylesheet", "(", "profile", ")", ":", "old_script", "=", "profile", ".", "scripts", "(", ")", ".", "findScript", "(", "'_qute_stylesheet'", ")", "if", "(", "not", "old_script", ".", "isNull", "(", ")", ")", ":", "profile", ".", "scripts", "(", ")", ".", "remove", "(", "old_script", ")", "css", "=", "shared", ".", "get_user_stylesheet", "(", ")", "source", "=", "\"\\n (function() {{\\n var css = document.createElement('style');\\n css.setAttribute('type', 'text/css');\\n css.appendChild(document.createTextNode('{}'));\\n document.getElementsByTagName('head')[0].appendChild(css);\\n }})()\\n \"", ".", "format", "(", "javascript", ".", "string_escape", "(", "css", ")", ")", "script", "=", "QWebEngineScript", "(", ")", "script", ".", "setName", "(", "'_qute_stylesheet'", ")", "script", ".", "setInjectionPoint", "(", "QWebEngineScript", ".", "DocumentReady", ")", "script", ".", "setWorldId", "(", "QWebEngineScript", ".", "ApplicationWorld", ")", "script", ".", "setRunsOnSubFrames", "(", "True", ")", "script", ".", "setSourceCode", "(", "source", ")", "profile", ".", "scripts", "(", ")", ".", "insert", "(", "script", ")" ]
initialize custom stylesheets .
train
false
16,343
def RemoveLinkDependenciesFromNoneTargets(targets): for (target_name, target_dict) in targets.iteritems(): for dependency_key in dependency_sections: dependencies = target_dict.get(dependency_key, []) if dependencies: for t in dependencies: if (target_dict.get('type', None) == 'none'): if targets[t].get('variables', {}).get('link_dependency', 0): target_dict[dependency_key] = Filter(target_dict[dependency_key], t)
[ "def", "RemoveLinkDependenciesFromNoneTargets", "(", "targets", ")", ":", "for", "(", "target_name", ",", "target_dict", ")", "in", "targets", ".", "iteritems", "(", ")", ":", "for", "dependency_key", "in", "dependency_sections", ":", "dependencies", "=", "target_dict", ".", "get", "(", "dependency_key", ",", "[", "]", ")", "if", "dependencies", ":", "for", "t", "in", "dependencies", ":", "if", "(", "target_dict", ".", "get", "(", "'type'", ",", "None", ")", "==", "'none'", ")", ":", "if", "targets", "[", "t", "]", ".", "get", "(", "'variables'", ",", "{", "}", ")", ".", "get", "(", "'link_dependency'", ",", "0", ")", ":", "target_dict", "[", "dependency_key", "]", "=", "Filter", "(", "target_dict", "[", "dependency_key", "]", ",", "t", ")" ]
remove dependencies having the link_dependency attribute from the none targets .
train
false
16,344
def spring(): rc(u'image', cmap=u'spring') im = gci() if (im is not None): im.set_cmap(cm.spring)
[ "def", "spring", "(", ")", ":", "rc", "(", "u'image'", ",", "cmap", "=", "u'spring'", ")", "im", "=", "gci", "(", ")", "if", "(", "im", "is", "not", "None", ")", ":", "im", ".", "set_cmap", "(", "cm", ".", "spring", ")" ]
set the default colormap to spring and apply to current image if any .
train
false
16,345
def detect_landmarks(path): vision_client = vision.Client() with io.open(path, 'rb') as image_file: content = image_file.read() image = vision_client.image(content=content) landmarks = image.detect_landmarks() print 'Landmarks:' for landmark in landmarks: print landmark.description
[ "def", "detect_landmarks", "(", "path", ")", ":", "vision_client", "=", "vision", ".", "Client", "(", ")", "with", "io", ".", "open", "(", "path", ",", "'rb'", ")", "as", "image_file", ":", "content", "=", "image_file", ".", "read", "(", ")", "image", "=", "vision_client", ".", "image", "(", "content", "=", "content", ")", "landmarks", "=", "image", ".", "detect_landmarks", "(", ")", "print", "'Landmarks:'", "for", "landmark", "in", "landmarks", ":", "print", "landmark", ".", "description" ]
detects landmarks in the file .
train
false
16,346
def Win32RawInput(prompt=None): try: sys.stdout.flush() sys.stderr.flush() except: pass if (prompt is None): prompt = '' ret = dialog.GetSimpleInput(prompt) if (ret == None): raise KeyboardInterrupt('operation cancelled') return ret
[ "def", "Win32RawInput", "(", "prompt", "=", "None", ")", ":", "try", ":", "sys", ".", "stdout", ".", "flush", "(", ")", "sys", ".", "stderr", ".", "flush", "(", ")", "except", ":", "pass", "if", "(", "prompt", "is", "None", ")", ":", "prompt", "=", "''", "ret", "=", "dialog", ".", "GetSimpleInput", "(", "prompt", ")", "if", "(", "ret", "==", "None", ")", ":", "raise", "KeyboardInterrupt", "(", "'operation cancelled'", ")", "return", "ret" ]
provide raw_input() for gui apps .
train
false
16,348
@profiler.trace def network_create(request, **kwargs): LOG.debug(('network_create(): kwargs = %s' % kwargs)) if ('net_profile_id' in kwargs): kwargs['n1kv:profile'] = kwargs.pop('net_profile_id') if ('tenant_id' not in kwargs): kwargs['tenant_id'] = request.user.project_id body = {'network': kwargs} network = neutronclient(request).create_network(body=body).get('network') return Network(network)
[ "@", "profiler", ".", "trace", "def", "network_create", "(", "request", ",", "**", "kwargs", ")", ":", "LOG", ".", "debug", "(", "(", "'network_create(): kwargs = %s'", "%", "kwargs", ")", ")", "if", "(", "'net_profile_id'", "in", "kwargs", ")", ":", "kwargs", "[", "'n1kv:profile'", "]", "=", "kwargs", ".", "pop", "(", "'net_profile_id'", ")", "if", "(", "'tenant_id'", "not", "in", "kwargs", ")", ":", "kwargs", "[", "'tenant_id'", "]", "=", "request", ".", "user", ".", "project_id", "body", "=", "{", "'network'", ":", "kwargs", "}", "network", "=", "neutronclient", "(", "request", ")", ".", "create_network", "(", "body", "=", "body", ")", ".", "get", "(", "'network'", ")", "return", "Network", "(", "network", ")" ]
create private network cli example: .
train
true
16,349
def wheels(opts, whitelist=None): return LazyLoader(_module_dirs(opts, 'wheel'), opts, tag='wheel', whitelist=whitelist)
[ "def", "wheels", "(", "opts", ",", "whitelist", "=", "None", ")", ":", "return", "LazyLoader", "(", "_module_dirs", "(", "opts", ",", "'wheel'", ")", ",", "opts", ",", "tag", "=", "'wheel'", ",", "whitelist", "=", "whitelist", ")" ]
returns the wheels modules .
train
false
16,350
def _is_fromfile_compatible(stream): if (sys.version_info[0] < 3): return True bad_cls = [] try: import gzip bad_cls.append(gzip.GzipFile) except ImportError: pass try: import bz2 bad_cls.append(bz2.BZ2File) except ImportError: pass bad_cls = tuple(bad_cls) return (not isinstance(stream, bad_cls))
[ "def", "_is_fromfile_compatible", "(", "stream", ")", ":", "if", "(", "sys", ".", "version_info", "[", "0", "]", "<", "3", ")", ":", "return", "True", "bad_cls", "=", "[", "]", "try", ":", "import", "gzip", "bad_cls", ".", "append", "(", "gzip", ".", "GzipFile", ")", "except", "ImportError", ":", "pass", "try", ":", "import", "bz2", "bad_cls", ".", "append", "(", "bz2", ".", "BZ2File", ")", "except", "ImportError", ":", "pass", "bad_cls", "=", "tuple", "(", "bad_cls", ")", "return", "(", "not", "isinstance", "(", "stream", ",", "bad_cls", ")", ")" ]
check whether stream is compatible with numpy .
train
false
16,351
def test_http_header_encoding(): mocked_socket = mock.MagicMock() mocked_socket.sendall = mock.MagicMock() mocked_request = mock.MagicMock() response = Response(mocked_request, mocked_socket, None) response.headers.append(('foo', u'h\xe4der')) with pytest.raises(UnicodeEncodeError): response.send_headers() tosend = response.default_headers() tosend.extend([('%s: %s\r\n' % (k, v)) for (k, v) in response.headers]) header_str = ('%s\r\n' % ''.join(tosend)) with pytest.raises(UnicodeEncodeError): mocked_socket.sendall(util.to_bytestring(header_str, 'ascii'))
[ "def", "test_http_header_encoding", "(", ")", ":", "mocked_socket", "=", "mock", ".", "MagicMock", "(", ")", "mocked_socket", ".", "sendall", "=", "mock", ".", "MagicMock", "(", ")", "mocked_request", "=", "mock", ".", "MagicMock", "(", ")", "response", "=", "Response", "(", "mocked_request", ",", "mocked_socket", ",", "None", ")", "response", ".", "headers", ".", "append", "(", "(", "'foo'", ",", "u'h\\xe4der'", ")", ")", "with", "pytest", ".", "raises", "(", "UnicodeEncodeError", ")", ":", "response", ".", "send_headers", "(", ")", "tosend", "=", "response", ".", "default_headers", "(", ")", "tosend", ".", "extend", "(", "[", "(", "'%s: %s\\r\\n'", "%", "(", "k", ",", "v", ")", ")", "for", "(", "k", ",", "v", ")", "in", "response", ".", "headers", "]", ")", "header_str", "=", "(", "'%s\\r\\n'", "%", "''", ".", "join", "(", "tosend", ")", ")", "with", "pytest", ".", "raises", "(", "UnicodeEncodeError", ")", ":", "mocked_socket", ".", "sendall", "(", "util", ".", "to_bytestring", "(", "header_str", ",", "'ascii'", ")", ")" ]
tests whether http response headers are usascii encoded .
train
false
16,352
def multiple_file_normalize_CSS(input_dir, output_dir, output_CSS_statistics): if (not exists(output_dir)): makedirs(output_dir) file_names = [fname for fname in listdir(input_dir) if (not (fname.startswith('.') or isdir(fname)))] for fname in file_names: (base_fname, ext) = splitext(fname) original_fname = (base_fname + '.biom') hdf5_infile = join(input_dir, original_fname) tmp_bt = load_table(hdf5_infile) outfile = join(output_dir, (('CSS_' + base_fname) + '.biom')) if output_CSS_statistics: output_CSS_statistics = join(output_dir, (('CSS_statistics_' + base_fname) + '.txt')) with tempfile.NamedTemporaryFile(dir=get_qiime_temp_dir(), prefix='QIIME-normalize-table-temp-table-', suffix='.biom') as temp_fh: temp_fh.write(tmp_bt.to_json('forR')) temp_fh.flush() run_CSS(temp_fh.name, outfile, output_CSS_statistics=output_CSS_statistics)
[ "def", "multiple_file_normalize_CSS", "(", "input_dir", ",", "output_dir", ",", "output_CSS_statistics", ")", ":", "if", "(", "not", "exists", "(", "output_dir", ")", ")", ":", "makedirs", "(", "output_dir", ")", "file_names", "=", "[", "fname", "for", "fname", "in", "listdir", "(", "input_dir", ")", "if", "(", "not", "(", "fname", ".", "startswith", "(", "'.'", ")", "or", "isdir", "(", "fname", ")", ")", ")", "]", "for", "fname", "in", "file_names", ":", "(", "base_fname", ",", "ext", ")", "=", "splitext", "(", "fname", ")", "original_fname", "=", "(", "base_fname", "+", "'.biom'", ")", "hdf5_infile", "=", "join", "(", "input_dir", ",", "original_fname", ")", "tmp_bt", "=", "load_table", "(", "hdf5_infile", ")", "outfile", "=", "join", "(", "output_dir", ",", "(", "(", "'CSS_'", "+", "base_fname", ")", "+", "'.biom'", ")", ")", "if", "output_CSS_statistics", ":", "output_CSS_statistics", "=", "join", "(", "output_dir", ",", "(", "(", "'CSS_statistics_'", "+", "base_fname", ")", "+", "'.txt'", ")", ")", "with", "tempfile", ".", "NamedTemporaryFile", "(", "dir", "=", "get_qiime_temp_dir", "(", ")", ",", "prefix", "=", "'QIIME-normalize-table-temp-table-'", ",", "suffix", "=", "'.biom'", ")", "as", "temp_fh", ":", "temp_fh", ".", "write", "(", "tmp_bt", ".", "to_json", "(", "'forR'", ")", ")", "temp_fh", ".", "flush", "(", ")", "run_CSS", "(", "temp_fh", ".", "name", ",", "outfile", ",", "output_CSS_statistics", "=", "output_CSS_statistics", ")" ]
performs metagenomeseqs css normalization on a directory of raw abundance otu matrices .
train
false
16,353
def kodi_to_config(MASTER_SETTINGS, config, new_settings): for (setting, new_value) in new_settings.iteritems(): setting_protocols = MASTER_SETTINGS.get(setting, None) if (setting_protocols == None): continue config = general_config_set(config, new_settings, new_value, **setting_protocols) return config
[ "def", "kodi_to_config", "(", "MASTER_SETTINGS", ",", "config", ",", "new_settings", ")", ":", "for", "(", "setting", ",", "new_value", ")", "in", "new_settings", ".", "iteritems", "(", ")", ":", "setting_protocols", "=", "MASTER_SETTINGS", ".", "get", "(", "setting", ",", "None", ")", "if", "(", "setting_protocols", "==", "None", ")", ":", "continue", "config", "=", "general_config_set", "(", "config", ",", "new_settings", ",", "new_value", ",", "**", "setting_protocols", ")", "return", "config" ]
takes the existing config .
train
false
16,356
def directLoop(isWiddershins, loop): if (euclidean.isWiddershins(loop) != isWiddershins): loop.reverse()
[ "def", "directLoop", "(", "isWiddershins", ",", "loop", ")", ":", "if", "(", "euclidean", ".", "isWiddershins", "(", "loop", ")", "!=", "isWiddershins", ")", ":", "loop", ".", "reverse", "(", ")" ]
direct the loop .
train
false
16,357
def remove_reqs_readonly(): if (not os.path.isdir(REQS_DIR)): return os.chmod(REQS_DIR, 493) for (root, dirs, files) in os.walk(REQS_DIR): for d in dirs: os.chmod(os.path.join(root, d), 493) for f in files: os.chmod(os.path.join(root, f), 493)
[ "def", "remove_reqs_readonly", "(", ")", ":", "if", "(", "not", "os", ".", "path", ".", "isdir", "(", "REQS_DIR", ")", ")", ":", "return", "os", ".", "chmod", "(", "REQS_DIR", ",", "493", ")", "for", "(", "root", ",", "dirs", ",", "files", ")", "in", "os", ".", "walk", "(", "REQS_DIR", ")", ":", "for", "d", "in", "dirs", ":", "os", ".", "chmod", "(", "os", ".", "path", ".", "join", "(", "root", ",", "d", ")", ",", "493", ")", "for", "f", "in", "files", ":", "os", ".", "chmod", "(", "os", ".", "path", ".", "join", "(", "root", ",", "f", ")", ",", "493", ")" ]
workaround for issue #569 .
train
false
16,358
def _summary_judgment(rec): if config['import']['quiet']: if (rec == Recommendation.strong): return importer.action.APPLY else: action = config['import']['quiet_fallback'].as_choice({'skip': importer.action.SKIP, 'asis': importer.action.ASIS}) elif (rec == Recommendation.none): action = config['import']['none_rec_action'].as_choice({'skip': importer.action.SKIP, 'asis': importer.action.ASIS, 'ask': None}) else: return None if (action == importer.action.SKIP): print_(u'Skipping.') elif (action == importer.action.ASIS): print_(u'Importing as-is.') return action
[ "def", "_summary_judgment", "(", "rec", ")", ":", "if", "config", "[", "'import'", "]", "[", "'quiet'", "]", ":", "if", "(", "rec", "==", "Recommendation", ".", "strong", ")", ":", "return", "importer", ".", "action", ".", "APPLY", "else", ":", "action", "=", "config", "[", "'import'", "]", "[", "'quiet_fallback'", "]", ".", "as_choice", "(", "{", "'skip'", ":", "importer", ".", "action", ".", "SKIP", ",", "'asis'", ":", "importer", ".", "action", ".", "ASIS", "}", ")", "elif", "(", "rec", "==", "Recommendation", ".", "none", ")", ":", "action", "=", "config", "[", "'import'", "]", "[", "'none_rec_action'", "]", ".", "as_choice", "(", "{", "'skip'", ":", "importer", ".", "action", ".", "SKIP", ",", "'asis'", ":", "importer", ".", "action", ".", "ASIS", ",", "'ask'", ":", "None", "}", ")", "else", ":", "return", "None", "if", "(", "action", "==", "importer", ".", "action", ".", "SKIP", ")", ":", "print_", "(", "u'Skipping.'", ")", "elif", "(", "action", "==", "importer", ".", "action", ".", "ASIS", ")", ":", "print_", "(", "u'Importing as-is.'", ")", "return", "action" ]
determines whether a decision should be made without even asking the user .
train
false
16,359
def center_text(text, length=80, left_edge='|', right_edge='|', text_length=None): if (text_length is None): text_length = get_text_length(text) output = [] char_start = (((length // 2) - (text_length // 2)) - 1) output.append(((left_edge + (' ' * char_start)) + text)) length_so_far = ((get_text_length(left_edge) + char_start) + text_length) right_side_spaces = ((length - get_text_length(right_edge)) - length_so_far) output.append((' ' * right_side_spaces)) output.append(right_edge) final = ''.join(output) return final
[ "def", "center_text", "(", "text", ",", "length", "=", "80", ",", "left_edge", "=", "'|'", ",", "right_edge", "=", "'|'", ",", "text_length", "=", "None", ")", ":", "if", "(", "text_length", "is", "None", ")", ":", "text_length", "=", "get_text_length", "(", "text", ")", "output", "=", "[", "]", "char_start", "=", "(", "(", "(", "length", "//", "2", ")", "-", "(", "text_length", "//", "2", ")", ")", "-", "1", ")", "output", ".", "append", "(", "(", "(", "left_edge", "+", "(", "' '", "*", "char_start", ")", ")", "+", "text", ")", ")", "length_so_far", "=", "(", "(", "get_text_length", "(", "left_edge", ")", "+", "char_start", ")", "+", "text_length", ")", "right_side_spaces", "=", "(", "(", "length", "-", "get_text_length", "(", "right_edge", ")", ")", "-", "length_so_far", ")", "output", ".", "append", "(", "(", "' '", "*", "right_side_spaces", ")", ")", "output", ".", "append", "(", "right_edge", ")", "final", "=", "''", ".", "join", "(", "output", ")", "return", "final" ]
center text with specified edge chars .
train
false
16,360
def xblock_local_resource_url(block, uri): xblock_class = getattr(block.__class__, 'unmixed_class', block.__class__) if (settings.PIPELINE_ENABLED or (not settings.REQUIRE_DEBUG)): return staticfiles_storage.url('xblock/resources/{package_name}/{path}'.format(package_name=xblock_class.__module__, path=uri)) else: return reverse('xblock_resource_url', kwargs={'block_type': block.scope_ids.block_type, 'uri': uri})
[ "def", "xblock_local_resource_url", "(", "block", ",", "uri", ")", ":", "xblock_class", "=", "getattr", "(", "block", ".", "__class__", ",", "'unmixed_class'", ",", "block", ".", "__class__", ")", "if", "(", "settings", ".", "PIPELINE_ENABLED", "or", "(", "not", "settings", ".", "REQUIRE_DEBUG", ")", ")", ":", "return", "staticfiles_storage", ".", "url", "(", "'xblock/resources/{package_name}/{path}'", ".", "format", "(", "package_name", "=", "xblock_class", ".", "__module__", ",", "path", "=", "uri", ")", ")", "else", ":", "return", "reverse", "(", "'xblock_resource_url'", ",", "kwargs", "=", "{", "'block_type'", ":", "block", ".", "scope_ids", ".", "block_type", ",", "'uri'", ":", "uri", "}", ")" ]
returns the url for an xblocks local resource .
train
false
16,361
def is_scalar_nonsparse_matrix(circuit, nqubits, identity_only): matrix = represent(Mul(*circuit), nqubits=nqubits) if isinstance(matrix, Number): return ((matrix == 1) if identity_only else True) else: matrix_trace = matrix.trace() adjusted_matrix_trace = ((matrix_trace / matrix[0]) if (not identity_only) else matrix_trace) is_identity = ((matrix[0] == 1.0) if identity_only else True) has_correct_trace = (adjusted_matrix_trace == pow(2, nqubits)) return bool((matrix.is_diagonal() and has_correct_trace and is_identity))
[ "def", "is_scalar_nonsparse_matrix", "(", "circuit", ",", "nqubits", ",", "identity_only", ")", ":", "matrix", "=", "represent", "(", "Mul", "(", "*", "circuit", ")", ",", "nqubits", "=", "nqubits", ")", "if", "isinstance", "(", "matrix", ",", "Number", ")", ":", "return", "(", "(", "matrix", "==", "1", ")", "if", "identity_only", "else", "True", ")", "else", ":", "matrix_trace", "=", "matrix", ".", "trace", "(", ")", "adjusted_matrix_trace", "=", "(", "(", "matrix_trace", "/", "matrix", "[", "0", "]", ")", "if", "(", "not", "identity_only", ")", "else", "matrix_trace", ")", "is_identity", "=", "(", "(", "matrix", "[", "0", "]", "==", "1.0", ")", "if", "identity_only", "else", "True", ")", "has_correct_trace", "=", "(", "adjusted_matrix_trace", "==", "pow", "(", "2", ",", "nqubits", ")", ")", "return", "bool", "(", "(", "matrix", ".", "is_diagonal", "(", ")", "and", "has_correct_trace", "and", "is_identity", ")", ")" ]
checks if a given circuit .
train
false
16,363
@requires_sklearn def test_spatio_temporal_tris_connectivity(): tris = np.array([[0, 1, 2], [3, 4, 5]]) connectivity = spatio_temporal_tris_connectivity(tris, 2) x = [1, 1, 0, 0, 0, 0, 1, 0, 0, 0, 0, 1] components = stats.cluster_level._get_components(np.array(x), connectivity) old_fmt = [0, 0, (-2), (-2), (-2), (-2), 0, (-2), (-2), (-2), (-2), 1] new_fmt = np.array(old_fmt) new_fmt = [np.nonzero((new_fmt == v))[0] for v in np.unique(new_fmt[(new_fmt >= 0)])] assert_true(len(new_fmt), len(components)) for (c, n) in zip(components, new_fmt): assert_array_equal(c, n)
[ "@", "requires_sklearn", "def", "test_spatio_temporal_tris_connectivity", "(", ")", ":", "tris", "=", "np", ".", "array", "(", "[", "[", "0", ",", "1", ",", "2", "]", ",", "[", "3", ",", "4", ",", "5", "]", "]", ")", "connectivity", "=", "spatio_temporal_tris_connectivity", "(", "tris", ",", "2", ")", "x", "=", "[", "1", ",", "1", ",", "0", ",", "0", ",", "0", ",", "0", ",", "1", ",", "0", ",", "0", ",", "0", ",", "0", ",", "1", "]", "components", "=", "stats", ".", "cluster_level", ".", "_get_components", "(", "np", ".", "array", "(", "x", ")", ",", "connectivity", ")", "old_fmt", "=", "[", "0", ",", "0", ",", "(", "-", "2", ")", ",", "(", "-", "2", ")", ",", "(", "-", "2", ")", ",", "(", "-", "2", ")", ",", "0", ",", "(", "-", "2", ")", ",", "(", "-", "2", ")", ",", "(", "-", "2", ")", ",", "(", "-", "2", ")", ",", "1", "]", "new_fmt", "=", "np", ".", "array", "(", "old_fmt", ")", "new_fmt", "=", "[", "np", ".", "nonzero", "(", "(", "new_fmt", "==", "v", ")", ")", "[", "0", "]", "for", "v", "in", "np", ".", "unique", "(", "new_fmt", "[", "(", "new_fmt", ">=", "0", ")", "]", ")", "]", "assert_true", "(", "len", "(", "new_fmt", ")", ",", "len", "(", "components", ")", ")", "for", "(", "c", ",", "n", ")", "in", "zip", "(", "components", ",", "new_fmt", ")", ":", "assert_array_equal", "(", "c", ",", "n", ")" ]
test spatio-temporal connectivity from triangles .
train
false
16,366
def upload(): os.system('cd build/html; rsync -avz . pandas@pandas.pydata.org:/usr/share/nginx/pandas/pandas-docs/vbench/ -essh')
[ "def", "upload", "(", ")", ":", "os", ".", "system", "(", "'cd build/html; rsync -avz . pandas@pandas.pydata.org:/usr/share/nginx/pandas/pandas-docs/vbench/ -essh'", ")" ]
upload handler .
train
false
16,367
def getProcessOutputAndValue(executable, args=(), env={}, path=None, reactor=None): return _callProtocolWithDeferred(_EverythingGetter, executable, args, env, path, reactor)
[ "def", "getProcessOutputAndValue", "(", "executable", ",", "args", "=", "(", ")", ",", "env", "=", "{", "}", ",", "path", "=", "None", ",", "reactor", "=", "None", ")", ":", "return", "_callProtocolWithDeferred", "(", "_EverythingGetter", ",", "executable", ",", "args", ",", "env", ",", "path", ",", "reactor", ")" ]
spawn a process and returns a deferred that will be called back with its output and its exit code as if a signal is raised .
train
false
16,368
def corr_ar(k_vars, ar): from scipy.linalg import toeplitz if (len(ar) < k_vars): ar_ = np.zeros(k_vars) ar_[:len(ar)] = ar ar = ar_ return toeplitz(ar)
[ "def", "corr_ar", "(", "k_vars", ",", "ar", ")", ":", "from", "scipy", ".", "linalg", "import", "toeplitz", "if", "(", "len", "(", "ar", ")", "<", "k_vars", ")", ":", "ar_", "=", "np", ".", "zeros", "(", "k_vars", ")", "ar_", "[", ":", "len", "(", "ar", ")", "]", "=", "ar", "ar", "=", "ar_", "return", "toeplitz", "(", "ar", ")" ]
create autoregressive correlation matrix this might be ma .
train
false
16,369
def RecordFromLine(line): try: (created, level, unused_source_location, message) = _StrictParseLogEntry(line, clean_message=False) message = Stripnl(message) return LoggingRecord(level, created, message, None) except ValueError: return StderrRecord(line)
[ "def", "RecordFromLine", "(", "line", ")", ":", "try", ":", "(", "created", ",", "level", ",", "unused_source_location", ",", "message", ")", "=", "_StrictParseLogEntry", "(", "line", ",", "clean_message", "=", "False", ")", "message", "=", "Stripnl", "(", "message", ")", "return", "LoggingRecord", "(", "level", ",", "created", ",", "message", ",", "None", ")", "except", "ValueError", ":", "return", "StderrRecord", "(", "line", ")" ]
create the correct type of record based on what the line looks like .
train
false
16,370
def GetSigner(secret): return GetSecretsManagerForSecret(secret).GetSigner(secret)
[ "def", "GetSigner", "(", "secret", ")", ":", "return", "GetSecretsManagerForSecret", "(", "secret", ")", ".", "GetSigner", "(", "secret", ")" ]
returns the keyczar signer object returned by the secrets manager instance getsigner method .
train
false
16,371
def _portsnap(): ret = ['portsnap'] if (float(__grains__['osrelease']) >= 10): ret.append('--interactive') return ret
[ "def", "_portsnap", "(", ")", ":", "ret", "=", "[", "'portsnap'", "]", "if", "(", "float", "(", "__grains__", "[", "'osrelease'", "]", ")", ">=", "10", ")", ":", "ret", ".", "append", "(", "'--interactive'", ")", "return", "ret" ]
return portsnap --interactive for freebsd 10 .
train
false
16,372
def isscalarlike(x): return (np.isscalar(x) or (isdense(x) and (x.ndim == 0)))
[ "def", "isscalarlike", "(", "x", ")", ":", "return", "(", "np", ".", "isscalar", "(", "x", ")", "or", "(", "isdense", "(", "x", ")", "and", "(", "x", ".", "ndim", "==", "0", ")", ")", ")" ]
is x either a scalar .
train
false
16,373
def xfs_mkfs_options(tune2fs_dict, mkfs_option): xfs_mapping = {'meta-data: isize': '-i size', 'meta-data: agcount': '-d agcount', 'meta-data: sectsz': '-s size', 'meta-data: attr': '-i attr', 'data: bsize': '-b size', 'data: imaxpct': '-i maxpct', 'data: sunit': '-d sunit', 'data: swidth': '-d swidth', 'data: unwritten': '-d unwritten', 'naming: version': '-n version', 'naming: bsize': '-n size', 'log: version': '-l version', 'log: sectsz': '-l sectsize', 'log: sunit': '-l sunit', 'log: lazy-count': '-l lazy-count', 'realtime: extsz': '-r extsize', 'realtime: blocks': '-r size', 'realtime: rtextents': '-r rtdev'} mkfs_option['-l size'] = (tune2fs_dict['log: bsize'] * tune2fs_dict['log: blocks']) for (key, value) in xfs_mapping.iteritems(): mkfs_option[value] = tune2fs_dict[key]
[ "def", "xfs_mkfs_options", "(", "tune2fs_dict", ",", "mkfs_option", ")", ":", "xfs_mapping", "=", "{", "'meta-data: isize'", ":", "'-i size'", ",", "'meta-data: agcount'", ":", "'-d agcount'", ",", "'meta-data: sectsz'", ":", "'-s size'", ",", "'meta-data: attr'", ":", "'-i attr'", ",", "'data: bsize'", ":", "'-b size'", ",", "'data: imaxpct'", ":", "'-i maxpct'", ",", "'data: sunit'", ":", "'-d sunit'", ",", "'data: swidth'", ":", "'-d swidth'", ",", "'data: unwritten'", ":", "'-d unwritten'", ",", "'naming: version'", ":", "'-n version'", ",", "'naming: bsize'", ":", "'-n size'", ",", "'log: version'", ":", "'-l version'", ",", "'log: sectsz'", ":", "'-l sectsize'", ",", "'log: sunit'", ":", "'-l sunit'", ",", "'log: lazy-count'", ":", "'-l lazy-count'", ",", "'realtime: extsz'", ":", "'-r extsize'", ",", "'realtime: blocks'", ":", "'-r size'", ",", "'realtime: rtextents'", ":", "'-r rtdev'", "}", "mkfs_option", "[", "'-l size'", "]", "=", "(", "tune2fs_dict", "[", "'log: bsize'", "]", "*", "tune2fs_dict", "[", "'log: blocks'", "]", ")", "for", "(", "key", ",", "value", ")", "in", "xfs_mapping", ".", "iteritems", "(", ")", ":", "mkfs_option", "[", "value", "]", "=", "tune2fs_dict", "[", "key", "]" ]
maps filesystem tunables to their corresponding mkfs options .
train
false
16,374
def format_add_taxa_summary_mapping(summary, tax_order, mapping, header, delimiter=';'): tax_order = [delimiter.join(tax) for tax in tax_order] header.extend(tax_order) (yield ('#%s\n' % ' DCTB '.join(header))) for row in mapping: sample_id = row[0] if (sample_id not in summary): continue row.extend(map(str, summary[sample_id])) (yield ('%s\n' % ' DCTB '.join(row)))
[ "def", "format_add_taxa_summary_mapping", "(", "summary", ",", "tax_order", ",", "mapping", ",", "header", ",", "delimiter", "=", "';'", ")", ":", "tax_order", "=", "[", "delimiter", ".", "join", "(", "tax", ")", "for", "tax", "in", "tax_order", "]", "header", ".", "extend", "(", "tax_order", ")", "(", "yield", "(", "'#%s\\n'", "%", "' DCTB '", ".", "join", "(", "header", ")", ")", ")", "for", "row", "in", "mapping", ":", "sample_id", "=", "row", "[", "0", "]", "if", "(", "sample_id", "not", "in", "summary", ")", ":", "continue", "row", ".", "extend", "(", "map", "(", "str", ",", "summary", "[", "sample_id", "]", ")", ")", "(", "yield", "(", "'%s\\n'", "%", "' DCTB '", ".", "join", "(", "row", ")", ")", ")" ]
formats a summarized taxonomy with mapping information .
train
false
16,375
@aborts def test_aborts_on_nonexistent_roles(): merge([], ['badrole'], [], {})
[ "@", "aborts", "def", "test_aborts_on_nonexistent_roles", "(", ")", ":", "merge", "(", "[", "]", ",", "[", "'badrole'", "]", ",", "[", "]", ",", "{", "}", ")" ]
aborts if any given roles arent found .
train
false
16,376
def link_fd_to_path(fd, target_path, dirs_created=0, retries=2, fsync=True): dirpath = os.path.dirname(target_path) for _junk in range(0, retries): try: linkat(linkat.AT_FDCWD, ('/proc/self/fd/%d' % fd), linkat.AT_FDCWD, target_path, linkat.AT_SYMLINK_FOLLOW) break except IOError as err: if (err.errno == errno.ENOENT): dirs_created = makedirs_count(dirpath) elif (err.errno == errno.EEXIST): try: os.unlink(target_path) except OSError as e: if (e.errno != errno.ENOENT): raise else: raise if fsync: for i in range(0, (dirs_created + 1)): fsync_dir(dirpath) dirpath = os.path.dirname(dirpath)
[ "def", "link_fd_to_path", "(", "fd", ",", "target_path", ",", "dirs_created", "=", "0", ",", "retries", "=", "2", ",", "fsync", "=", "True", ")", ":", "dirpath", "=", "os", ".", "path", ".", "dirname", "(", "target_path", ")", "for", "_junk", "in", "range", "(", "0", ",", "retries", ")", ":", "try", ":", "linkat", "(", "linkat", ".", "AT_FDCWD", ",", "(", "'/proc/self/fd/%d'", "%", "fd", ")", ",", "linkat", ".", "AT_FDCWD", ",", "target_path", ",", "linkat", ".", "AT_SYMLINK_FOLLOW", ")", "break", "except", "IOError", "as", "err", ":", "if", "(", "err", ".", "errno", "==", "errno", ".", "ENOENT", ")", ":", "dirs_created", "=", "makedirs_count", "(", "dirpath", ")", "elif", "(", "err", ".", "errno", "==", "errno", ".", "EEXIST", ")", ":", "try", ":", "os", ".", "unlink", "(", "target_path", ")", "except", "OSError", "as", "e", ":", "if", "(", "e", ".", "errno", "!=", "errno", ".", "ENOENT", ")", ":", "raise", "else", ":", "raise", "if", "fsync", ":", "for", "i", "in", "range", "(", "0", ",", "(", "dirs_created", "+", "1", ")", ")", ":", "fsync_dir", "(", "dirpath", ")", "dirpath", "=", "os", ".", "path", ".", "dirname", "(", "dirpath", ")" ]
creates a link to file descriptor at target_path specified .
train
false
16,380
def reprogress(): if (_last_progress and _last_progress.endswith('\r')): progress(_last_progress)
[ "def", "reprogress", "(", ")", ":", "if", "(", "_last_progress", "and", "_last_progress", ".", "endswith", "(", "'\\r'", ")", ")", ":", "progress", "(", "_last_progress", ")" ]
calls progress() to redisplay the most recent progress message .
train
false
16,381
def with_timeout(seconds, function, *args, **kwds): timeout_value = kwds.pop('timeout_value', _NONE) timeout = Timeout.start_new(seconds) try: return function(*args, **kwds) except Timeout: if ((sys.exc_info()[1] is timeout) and (timeout_value is not _NONE)): return timeout_value raise finally: timeout.cancel()
[ "def", "with_timeout", "(", "seconds", ",", "function", ",", "*", "args", ",", "**", "kwds", ")", ":", "timeout_value", "=", "kwds", ".", "pop", "(", "'timeout_value'", ",", "_NONE", ")", "timeout", "=", "Timeout", ".", "start_new", "(", "seconds", ")", "try", ":", "return", "function", "(", "*", "args", ",", "**", "kwds", ")", "except", "Timeout", ":", "if", "(", "(", "sys", ".", "exc_info", "(", ")", "[", "1", "]", "is", "timeout", ")", "and", "(", "timeout_value", "is", "not", "_NONE", ")", ")", ":", "return", "timeout_value", "raise", "finally", ":", "timeout", ".", "cancel", "(", ")" ]
wraps a .
train
false
16,382
def validate_fields_spec(cls, model, opts, flds, label): for fields in flds: if (type(fields) != tuple): fields = (fields,) for field in fields: if (field in cls.readonly_fields): continue check_formfield(cls, model, opts, label, field) try: f = opts.get_field(field) except models.FieldDoesNotExist: continue if (isinstance(f, models.ManyToManyField) and (not f.rel.through._meta.auto_created)): raise ImproperlyConfigured(("'%s.%s' can't include the ManyToManyField field '%s' because '%s' manually specifies a 'through' model." % (cls.__name__, label, field, field)))
[ "def", "validate_fields_spec", "(", "cls", ",", "model", ",", "opts", ",", "flds", ",", "label", ")", ":", "for", "fields", "in", "flds", ":", "if", "(", "type", "(", "fields", ")", "!=", "tuple", ")", ":", "fields", "=", "(", "fields", ",", ")", "for", "field", "in", "fields", ":", "if", "(", "field", "in", "cls", ".", "readonly_fields", ")", ":", "continue", "check_formfield", "(", "cls", ",", "model", ",", "opts", ",", "label", ",", "field", ")", "try", ":", "f", "=", "opts", ".", "get_field", "(", "field", ")", "except", "models", ".", "FieldDoesNotExist", ":", "continue", "if", "(", "isinstance", "(", "f", ",", "models", ".", "ManyToManyField", ")", "and", "(", "not", "f", ".", "rel", ".", "through", ".", "_meta", ".", "auto_created", ")", ")", ":", "raise", "ImproperlyConfigured", "(", "(", "\"'%s.%s' can't include the ManyToManyField field '%s' because '%s' manually specifies a 'through' model.\"", "%", "(", "cls", ".", "__name__", ",", "label", ",", "field", ",", "field", ")", ")", ")" ]
validate the fields specification in flds from a modeladmin subclass cls for the model model .
train
false
16,383
@pytest.fixture def install_egg(modules_tmpdir, monkeypatch): def inner(name, base=modules_tmpdir): if (not isinstance(name, str)): raise ValueError(name) base.join(name).ensure_dir() base.join(name).join('__init__.py').ensure() egg_setup = base.join('setup.py') egg_setup.write(textwrap.dedent("\n from setuptools import setup\n setup(name='{0}',\n version='1.0',\n packages=['site_egg'],\n zip_safe=True)\n ".format(name))) import subprocess subprocess.check_call([sys.executable, 'setup.py', 'bdist_egg'], cwd=str(modules_tmpdir)) (egg_path,) = modules_tmpdir.join('dist/').listdir() monkeypatch.syspath_prepend(str(egg_path)) return egg_path return inner
[ "@", "pytest", ".", "fixture", "def", "install_egg", "(", "modules_tmpdir", ",", "monkeypatch", ")", ":", "def", "inner", "(", "name", ",", "base", "=", "modules_tmpdir", ")", ":", "if", "(", "not", "isinstance", "(", "name", ",", "str", ")", ")", ":", "raise", "ValueError", "(", "name", ")", "base", ".", "join", "(", "name", ")", ".", "ensure_dir", "(", ")", "base", ".", "join", "(", "name", ")", ".", "join", "(", "'__init__.py'", ")", ".", "ensure", "(", ")", "egg_setup", "=", "base", ".", "join", "(", "'setup.py'", ")", "egg_setup", ".", "write", "(", "textwrap", ".", "dedent", "(", "\"\\n from setuptools import setup\\n setup(name='{0}',\\n version='1.0',\\n packages=['site_egg'],\\n zip_safe=True)\\n \"", ".", "format", "(", "name", ")", ")", ")", "import", "subprocess", "subprocess", ".", "check_call", "(", "[", "sys", ".", "executable", ",", "'setup.py'", ",", "'bdist_egg'", "]", ",", "cwd", "=", "str", "(", "modules_tmpdir", ")", ")", "(", "egg_path", ",", ")", "=", "modules_tmpdir", ".", "join", "(", "'dist/'", ")", ".", "listdir", "(", ")", "monkeypatch", ".", "syspath_prepend", "(", "str", "(", "egg_path", ")", ")", "return", "egg_path", "return", "inner" ]
generate egg from package name inside base and put the egg into sys .
train
false
16,384
def system_script_extension(system=None): exts = {'windows': '.bat', 'darwin': '.command', 'linux': '.sh'} system = (system or platform.system()) return exts.get(system.lower(), '.sh')
[ "def", "system_script_extension", "(", "system", "=", "None", ")", ":", "exts", "=", "{", "'windows'", ":", "'.bat'", ",", "'darwin'", ":", "'.command'", ",", "'linux'", ":", "'.sh'", "}", "system", "=", "(", "system", "or", "platform", ".", "system", "(", ")", ")", "return", "exts", ".", "get", "(", "system", ".", "lower", "(", ")", ",", "'.sh'", ")" ]
the extension for the one script that could be considered "the os script" for the given system .
train
false
16,388
def test_seeg_ecog(): (n_epochs, n_channels, n_times, sfreq) = (5, 10, 20, 1000.0) data = np.ones((n_epochs, n_channels, n_times)) events = np.array([np.arange(n_epochs), ([0] * n_epochs), ([1] * n_epochs)]).T pick_dict = dict(meg=False, exclude=[]) for key in ('seeg', 'ecog'): info = create_info(n_channels, sfreq, key) epochs = EpochsArray(data, info, events) pick_dict.update({key: True}) picks = pick_types(epochs.info, **pick_dict) del pick_dict[key] assert_equal(len(picks), n_channels)
[ "def", "test_seeg_ecog", "(", ")", ":", "(", "n_epochs", ",", "n_channels", ",", "n_times", ",", "sfreq", ")", "=", "(", "5", ",", "10", ",", "20", ",", "1000.0", ")", "data", "=", "np", ".", "ones", "(", "(", "n_epochs", ",", "n_channels", ",", "n_times", ")", ")", "events", "=", "np", ".", "array", "(", "[", "np", ".", "arange", "(", "n_epochs", ")", ",", "(", "[", "0", "]", "*", "n_epochs", ")", ",", "(", "[", "1", "]", "*", "n_epochs", ")", "]", ")", ".", "T", "pick_dict", "=", "dict", "(", "meg", "=", "False", ",", "exclude", "=", "[", "]", ")", "for", "key", "in", "(", "'seeg'", ",", "'ecog'", ")", ":", "info", "=", "create_info", "(", "n_channels", ",", "sfreq", ",", "key", ")", "epochs", "=", "EpochsArray", "(", "data", ",", "info", ",", "events", ")", "pick_dict", ".", "update", "(", "{", "key", ":", "True", "}", ")", "picks", "=", "pick_types", "(", "epochs", ".", "info", ",", "**", "pick_dict", ")", "del", "pick_dict", "[", "key", "]", "assert_equal", "(", "len", "(", "picks", ")", ",", "n_channels", ")" ]
test the compatibility of the epoch object with seeg and ecog data .
train
false
16,391
def pillar(tgt, delimiter=DEFAULT_TARGET_DELIM): matcher = salt.minion.Matcher({'pillar': __pillar__}, __salt__) try: return matcher.pillar_match(tgt, delimiter=delimiter) except Exception as exc: log.exception(exc) return False
[ "def", "pillar", "(", "tgt", ",", "delimiter", "=", "DEFAULT_TARGET_DELIM", ")", ":", "matcher", "=", "salt", ".", "minion", ".", "Matcher", "(", "{", "'pillar'", ":", "__pillar__", "}", ",", "__salt__", ")", "try", ":", "return", "matcher", ".", "pillar_match", "(", "tgt", ",", "delimiter", "=", "delimiter", ")", "except", "Exception", "as", "exc", ":", "log", ".", "exception", "(", "exc", ")", "return", "False" ]
return true if the minion matches the given pillar target .
train
false
16,392
def get_pk_columns(cls): retval = [] for (k, v) in cls.get_flat_type_info(cls).items(): if ((v.Attributes.sqla_column_args is not None) and v.Attributes.sqla_column_args[(-1)].get('primary_key', False)): retval.append((k, v)) return (tuple(retval) if (len(retval) > 0) else None)
[ "def", "get_pk_columns", "(", "cls", ")", ":", "retval", "=", "[", "]", "for", "(", "k", ",", "v", ")", "in", "cls", ".", "get_flat_type_info", "(", "cls", ")", ".", "items", "(", ")", ":", "if", "(", "(", "v", ".", "Attributes", ".", "sqla_column_args", "is", "not", "None", ")", "and", "v", ".", "Attributes", ".", "sqla_column_args", "[", "(", "-", "1", ")", "]", ".", "get", "(", "'primary_key'", ",", "False", ")", ")", ":", "retval", ".", "append", "(", "(", "k", ",", "v", ")", ")", "return", "(", "tuple", "(", "retval", ")", "if", "(", "len", "(", "retval", ")", ">", "0", ")", "else", "None", ")" ]
return primary key fields of a spyne object .
train
false
16,393
def display_upstream_changes(registry, xml_parent, data): XML.SubElement(xml_parent, 'jenkins.plugins.displayupstreamchanges.DisplayUpstreamChangesRecorder')
[ "def", "display_upstream_changes", "(", "registry", ",", "xml_parent", ",", "data", ")", ":", "XML", ".", "SubElement", "(", "xml_parent", ",", "'jenkins.plugins.displayupstreamchanges.DisplayUpstreamChangesRecorder'", ")" ]
yaml: display-upstream-changes display scm changes of upstream jobs .
train
false
16,395
def read_error_handler(func): @wraps(func) def error_handler_wrapper(*args, **kwargs): try: return func(*args, **kwargs) except Exception: return (args[0], '') return error_handler_wrapper
[ "def", "read_error_handler", "(", "func", ")", ":", "@", "wraps", "(", "func", ")", "def", "error_handler_wrapper", "(", "*", "args", ",", "**", "kwargs", ")", ":", "try", ":", "return", "func", "(", "*", "args", ",", "**", "kwargs", ")", "except", "Exception", ":", "return", "(", "args", "[", "0", "]", ",", "''", ")", "return", "error_handler_wrapper" ]
ignores exceptions by always returning the filename and an empty string as read file content .
train
false
16,396
def load_square_sprite_image(img_path, n_sprites): (tile_rows, tile_cols) = get_tiles_height_width(n_sprites) return load_sprite_image(img_path, (tile_rows, tile_cols), n_sprites=n_sprites)
[ "def", "load_square_sprite_image", "(", "img_path", ",", "n_sprites", ")", ":", "(", "tile_rows", ",", "tile_cols", ")", "=", "get_tiles_height_width", "(", "n_sprites", ")", "return", "load_sprite_image", "(", "img_path", ",", "(", "tile_rows", ",", "tile_cols", ")", ",", "n_sprites", "=", "n_sprites", ")" ]
just like load_sprite_image but assumes tiled image is square .
train
false
16,397
def max_none(data): if (not data): return max(data) non_none_data = [] for d in data: if (d is not None): non_none_data.append(d) return (max(non_none_data) if non_none_data else None)
[ "def", "max_none", "(", "data", ")", ":", "if", "(", "not", "data", ")", ":", "return", "max", "(", "data", ")", "non_none_data", "=", "[", "]", "for", "d", "in", "data", ":", "if", "(", "d", "is", "not", "None", ")", ":", "non_none_data", ".", "append", "(", "d", ")", "return", "(", "max", "(", "non_none_data", ")", "if", "non_none_data", "else", "None", ")" ]
given a list of data .
train
false
16,399
def getargtxt(obj, one_arg_per_line=True): args = getargs(obj) if args: sep = ', ' textlist = None for (i_arg, arg) in enumerate(args): if (textlist is None): textlist = [''] textlist[(-1)] += arg if (i_arg < (len(args) - 1)): textlist[(-1)] += sep if ((len(textlist[(-1)]) >= 32) or one_arg_per_line): textlist.append('') if (inspect.isclass(obj) or inspect.ismethod(obj)): if (len(textlist) == 1): return None if (('self' + sep) in textlist): textlist.remove(('self' + sep)) return textlist
[ "def", "getargtxt", "(", "obj", ",", "one_arg_per_line", "=", "True", ")", ":", "args", "=", "getargs", "(", "obj", ")", "if", "args", ":", "sep", "=", "', '", "textlist", "=", "None", "for", "(", "i_arg", ",", "arg", ")", "in", "enumerate", "(", "args", ")", ":", "if", "(", "textlist", "is", "None", ")", ":", "textlist", "=", "[", "''", "]", "textlist", "[", "(", "-", "1", ")", "]", "+=", "arg", "if", "(", "i_arg", "<", "(", "len", "(", "args", ")", "-", "1", ")", ")", ":", "textlist", "[", "(", "-", "1", ")", "]", "+=", "sep", "if", "(", "(", "len", "(", "textlist", "[", "(", "-", "1", ")", "]", ")", ">=", "32", ")", "or", "one_arg_per_line", ")", ":", "textlist", ".", "append", "(", "''", ")", "if", "(", "inspect", ".", "isclass", "(", "obj", ")", "or", "inspect", ".", "ismethod", "(", "obj", ")", ")", ":", "if", "(", "len", "(", "textlist", ")", "==", "1", ")", ":", "return", "None", "if", "(", "(", "'self'", "+", "sep", ")", "in", "textlist", ")", ":", "textlist", ".", "remove", "(", "(", "'self'", "+", "sep", ")", ")", "return", "textlist" ]
get the names and default values of a functions arguments return list with separators formatted for calltips .
train
true
16,400
def get_role(trans, id): id = trans.security.decode_id(id) role = trans.sa_session.query(trans.model.Role).get(id) if (not role): return trans.show_error_message(('Role not found for id (%s)' % str(id))) return role
[ "def", "get_role", "(", "trans", ",", "id", ")", ":", "id", "=", "trans", ".", "security", ".", "decode_id", "(", "id", ")", "role", "=", "trans", ".", "sa_session", ".", "query", "(", "trans", ".", "model", ".", "Role", ")", ".", "get", "(", "id", ")", "if", "(", "not", "role", ")", ":", "return", "trans", ".", "show_error_message", "(", "(", "'Role not found for id (%s)'", "%", "str", "(", "id", ")", ")", ")", "return", "role" ]
retrieve a role by name .
train
false
16,401
@bdd.given('I have a fresh instance') def fresh_instance(quteproc): quteproc.terminate() quteproc.start()
[ "@", "bdd", ".", "given", "(", "'I have a fresh instance'", ")", "def", "fresh_instance", "(", "quteproc", ")", ":", "quteproc", ".", "terminate", "(", ")", "quteproc", ".", "start", "(", ")" ]
restart qutebrowser instance for tests needing a fresh state .
train
false
16,402
def to_categorical(y, nb_classes): y = np.asarray(y, dtype='int32') if (not nb_classes): nb_classes = (np.max(y) + 1) Y = np.zeros((len(y), nb_classes)) for i in range(len(y)): Y[(i, y[i])] = 1.0 return Y
[ "def", "to_categorical", "(", "y", ",", "nb_classes", ")", ":", "y", "=", "np", ".", "asarray", "(", "y", ",", "dtype", "=", "'int32'", ")", "if", "(", "not", "nb_classes", ")", ":", "nb_classes", "=", "(", "np", ".", "max", "(", "y", ")", "+", "1", ")", "Y", "=", "np", ".", "zeros", "(", "(", "len", "(", "y", ")", ",", "nb_classes", ")", ")", "for", "i", "in", "range", "(", "len", "(", "y", ")", ")", ":", "Y", "[", "(", "i", ",", "y", "[", "i", "]", ")", "]", "=", "1.0", "return", "Y" ]
converts a class vector to binary class matrix .
train
false
16,403
def perform_reset(request, obj): return execute_locked(request, obj, _('All repositories have been reset.'), obj.do_reset, request)
[ "def", "perform_reset", "(", "request", ",", "obj", ")", ":", "return", "execute_locked", "(", "request", ",", "obj", ",", "_", "(", "'All repositories have been reset.'", ")", ",", "obj", ".", "do_reset", ",", "request", ")" ]
helper function to do the repository reset .
train
false
16,404
def check_access(action, context, data_dict=None): try: audit = context.get('__auth_audit', [])[(-1)] except IndexError: audit = '' if (audit and (audit[0] == action)): context['__auth_audit'].pop() user = context.get('user') try: if ('auth_user_obj' not in context): context['auth_user_obj'] = None if (not context.get('ignore_auth')): if (not context.get('__auth_user_obj_checked')): if (context.get('user') and (not context.get('auth_user_obj'))): context['auth_user_obj'] = model.User.by_name(context['user']) context['__auth_user_obj_checked'] = True context = _prepopulate_context(context) logic_authorization = authz.is_authorized(action, context, data_dict) if (not logic_authorization['success']): msg = logic_authorization.get('msg', '') raise NotAuthorized(msg) except NotAuthorized as e: log.debug(u'check access NotAuthorized - %s user=%s "%s"', action, user, unicode(e)) raise log.debug('check access OK - %s user=%s', action, user) return True
[ "def", "check_access", "(", "action", ",", "context", ",", "data_dict", "=", "None", ")", ":", "try", ":", "audit", "=", "context", ".", "get", "(", "'__auth_audit'", ",", "[", "]", ")", "[", "(", "-", "1", ")", "]", "except", "IndexError", ":", "audit", "=", "''", "if", "(", "audit", "and", "(", "audit", "[", "0", "]", "==", "action", ")", ")", ":", "context", "[", "'__auth_audit'", "]", ".", "pop", "(", ")", "user", "=", "context", ".", "get", "(", "'user'", ")", "try", ":", "if", "(", "'auth_user_obj'", "not", "in", "context", ")", ":", "context", "[", "'auth_user_obj'", "]", "=", "None", "if", "(", "not", "context", ".", "get", "(", "'ignore_auth'", ")", ")", ":", "if", "(", "not", "context", ".", "get", "(", "'__auth_user_obj_checked'", ")", ")", ":", "if", "(", "context", ".", "get", "(", "'user'", ")", "and", "(", "not", "context", ".", "get", "(", "'auth_user_obj'", ")", ")", ")", ":", "context", "[", "'auth_user_obj'", "]", "=", "model", ".", "User", ".", "by_name", "(", "context", "[", "'user'", "]", ")", "context", "[", "'__auth_user_obj_checked'", "]", "=", "True", "context", "=", "_prepopulate_context", "(", "context", ")", "logic_authorization", "=", "authz", ".", "is_authorized", "(", "action", ",", "context", ",", "data_dict", ")", "if", "(", "not", "logic_authorization", "[", "'success'", "]", ")", ":", "msg", "=", "logic_authorization", ".", "get", "(", "'msg'", ",", "''", ")", "raise", "NotAuthorized", "(", "msg", ")", "except", "NotAuthorized", "as", "e", ":", "log", ".", "debug", "(", "u'check access NotAuthorized - %s user=%s \"%s\"'", ",", "action", ",", "user", ",", "unicode", "(", "e", ")", ")", "raise", "log", ".", "debug", "(", "'check access OK - %s user=%s'", ",", "action", ",", "user", ")", "return", "True" ]
check if external address is allowed given access_type access_type: 1=nzb .
train
false
16,405
def waterbutler_url_for(request_type, provider, path, node_id, token, obj_args=None, **query): url = furl.furl(website_settings.WATERBUTLER_URL) url.path.segments.append(request_type) url.args.update({'path': path, 'nid': node_id, 'provider': provider}) if (token is not None): url.args['cookie'] = token if ('view_only' in obj_args): url.args['view_only'] = obj_args['view_only'] url.args.update(query) return url.url
[ "def", "waterbutler_url_for", "(", "request_type", ",", "provider", ",", "path", ",", "node_id", ",", "token", ",", "obj_args", "=", "None", ",", "**", "query", ")", ":", "url", "=", "furl", ".", "furl", "(", "website_settings", ".", "WATERBUTLER_URL", ")", "url", ".", "path", ".", "segments", ".", "append", "(", "request_type", ")", "url", ".", "args", ".", "update", "(", "{", "'path'", ":", "path", ",", "'nid'", ":", "node_id", ",", "'provider'", ":", "provider", "}", ")", "if", "(", "token", "is", "not", "None", ")", ":", "url", ".", "args", "[", "'cookie'", "]", "=", "token", "if", "(", "'view_only'", "in", "obj_args", ")", ":", "url", ".", "args", "[", "'view_only'", "]", "=", "obj_args", "[", "'view_only'", "]", "url", ".", "args", ".", "update", "(", "query", ")", "return", "url", ".", "url" ]
reverse url lookup for waterbutler routes .
train
false
16,406
def monitorFiles(outfiles, seconds, timeoutms): devnull = open('/dev/null', 'w') (tails, fdToFile, fdToHost) = ({}, {}, {}) for (h, outfile) in outfiles.iteritems(): tail = Popen(['tail', '-f', outfile], stdout=PIPE, stderr=devnull) fd = tail.stdout.fileno() tails[h] = tail fdToFile[fd] = tail.stdout fdToHost[fd] = h readable = poll() for t in tails.values(): readable.register(t.stdout.fileno(), POLLIN) endTime = (time() + seconds) while (time() < endTime): fdlist = readable.poll(timeoutms) if fdlist: for (fd, _flags) in fdlist: f = fdToFile[fd] host = fdToHost[fd] line = f.readline().strip() (yield (host, line)) else: (yield (None, '')) for t in tails.values(): t.terminate() devnull.close()
[ "def", "monitorFiles", "(", "outfiles", ",", "seconds", ",", "timeoutms", ")", ":", "devnull", "=", "open", "(", "'/dev/null'", ",", "'w'", ")", "(", "tails", ",", "fdToFile", ",", "fdToHost", ")", "=", "(", "{", "}", ",", "{", "}", ",", "{", "}", ")", "for", "(", "h", ",", "outfile", ")", "in", "outfiles", ".", "iteritems", "(", ")", ":", "tail", "=", "Popen", "(", "[", "'tail'", ",", "'-f'", ",", "outfile", "]", ",", "stdout", "=", "PIPE", ",", "stderr", "=", "devnull", ")", "fd", "=", "tail", ".", "stdout", ".", "fileno", "(", ")", "tails", "[", "h", "]", "=", "tail", "fdToFile", "[", "fd", "]", "=", "tail", ".", "stdout", "fdToHost", "[", "fd", "]", "=", "h", "readable", "=", "poll", "(", ")", "for", "t", "in", "tails", ".", "values", "(", ")", ":", "readable", ".", "register", "(", "t", ".", "stdout", ".", "fileno", "(", ")", ",", "POLLIN", ")", "endTime", "=", "(", "time", "(", ")", "+", "seconds", ")", "while", "(", "time", "(", ")", "<", "endTime", ")", ":", "fdlist", "=", "readable", ".", "poll", "(", "timeoutms", ")", "if", "fdlist", ":", "for", "(", "fd", ",", "_flags", ")", "in", "fdlist", ":", "f", "=", "fdToFile", "[", "fd", "]", "host", "=", "fdToHost", "[", "fd", "]", "line", "=", "f", ".", "readline", "(", ")", ".", "strip", "(", ")", "(", "yield", "(", "host", ",", "line", ")", ")", "else", ":", "(", "yield", "(", "None", ",", "''", ")", ")", "for", "t", "in", "tails", ".", "values", "(", ")", ":", "t", ".", "terminate", "(", ")", "devnull", ".", "close", "(", ")" ]
monitor set of files and return [ .
train
false
16,407
def get_debug_count(environ): if ('paste.evalexception.debug_count' in environ): return environ['paste.evalexception.debug_count'] else: environ['paste.evalexception.debug_count'] = next = six.next(debug_counter) return next
[ "def", "get_debug_count", "(", "environ", ")", ":", "if", "(", "'paste.evalexception.debug_count'", "in", "environ", ")", ":", "return", "environ", "[", "'paste.evalexception.debug_count'", "]", "else", ":", "environ", "[", "'paste.evalexception.debug_count'", "]", "=", "next", "=", "six", ".", "next", "(", "debug_counter", ")", "return", "next" ]
return the unique debug count for the current request .
train
false
16,409
def _primitive_root_prime_iter(p): p = as_int(p) v = [((p - 1) // i) for i in factorint((p - 1)).keys()] a = 2 while (a < p): for pw in v: if (pow(a, pw, p) == 1): break else: (yield a) a += 1
[ "def", "_primitive_root_prime_iter", "(", "p", ")", ":", "p", "=", "as_int", "(", "p", ")", "v", "=", "[", "(", "(", "p", "-", "1", ")", "//", "i", ")", "for", "i", "in", "factorint", "(", "(", "p", "-", "1", ")", ")", ".", "keys", "(", ")", "]", "a", "=", "2", "while", "(", "a", "<", "p", ")", ":", "for", "pw", "in", "v", ":", "if", "(", "pow", "(", "a", ",", "pw", ",", "p", ")", "==", "1", ")", ":", "break", "else", ":", "(", "yield", "a", ")", "a", "+=", "1" ]
generates the primitive roots for a prime p references .
train
false
16,412
def _parse_circ_path(path): if path: try: return [_parse_circ_entry(entry) for entry in path.split(',')] except stem.ProtocolError as exc: raise stem.ProtocolError(('%s: %s' % (exc, path))) else: return []
[ "def", "_parse_circ_path", "(", "path", ")", ":", "if", "path", ":", "try", ":", "return", "[", "_parse_circ_entry", "(", "entry", ")", "for", "entry", "in", "path", ".", "split", "(", "','", ")", "]", "except", "stem", ".", "ProtocolError", "as", "exc", ":", "raise", "stem", ".", "ProtocolError", "(", "(", "'%s: %s'", "%", "(", "exc", ",", "path", ")", ")", ")", "else", ":", "return", "[", "]" ]
parses a circuit path as a list of **** tuples .
train
false
16,413
def server_powerstatus(host=None, admin_username=None, admin_password=None, module=None): ret = __execute_ret('serveraction powerstatus', host=host, admin_username=admin_username, admin_password=admin_password, module=module) result = {'retcode': 0} if (ret['stdout'] == 'ON'): result['status'] = True result['comment'] = 'Power is on' if (ret['stdout'] == 'OFF'): result['status'] = False result['comment'] = 'Power is on' if ret['stdout'].startswith('ERROR'): result['status'] = False result['comment'] = ret['stdout'] return result
[ "def", "server_powerstatus", "(", "host", "=", "None", ",", "admin_username", "=", "None", ",", "admin_password", "=", "None", ",", "module", "=", "None", ")", ":", "ret", "=", "__execute_ret", "(", "'serveraction powerstatus'", ",", "host", "=", "host", ",", "admin_username", "=", "admin_username", ",", "admin_password", "=", "admin_password", ",", "module", "=", "module", ")", "result", "=", "{", "'retcode'", ":", "0", "}", "if", "(", "ret", "[", "'stdout'", "]", "==", "'ON'", ")", ":", "result", "[", "'status'", "]", "=", "True", "result", "[", "'comment'", "]", "=", "'Power is on'", "if", "(", "ret", "[", "'stdout'", "]", "==", "'OFF'", ")", ":", "result", "[", "'status'", "]", "=", "False", "result", "[", "'comment'", "]", "=", "'Power is on'", "if", "ret", "[", "'stdout'", "]", ".", "startswith", "(", "'ERROR'", ")", ":", "result", "[", "'status'", "]", "=", "False", "result", "[", "'comment'", "]", "=", "ret", "[", "'stdout'", "]", "return", "result" ]
return the power status for the passed module cli example: .
train
true
16,414
def libvlc_vprinterr(fmt, ap): f = (_Cfunctions.get('libvlc_vprinterr', None) or _Cfunction('libvlc_vprinterr', ((1,), (1,)), None, ctypes.c_char_p, ctypes.c_char_p, ctypes.c_void_p)) return f(fmt, ap)
[ "def", "libvlc_vprinterr", "(", "fmt", ",", "ap", ")", ":", "f", "=", "(", "_Cfunctions", ".", "get", "(", "'libvlc_vprinterr'", ",", "None", ")", "or", "_Cfunction", "(", "'libvlc_vprinterr'", ",", "(", "(", "1", ",", ")", ",", "(", "1", ",", ")", ")", ",", "None", ",", "ctypes", ".", "c_char_p", ",", "ctypes", ".", "c_char_p", ",", "ctypes", ".", "c_void_p", ")", ")", "return", "f", "(", "fmt", ",", "ap", ")" ]
sets the libvlc error status and message for the current thread .
train
true
16,415
def do_hypervisor_stats(cs, args): stats = cs.hypervisor_stats.statistics() utils.print_dict(stats.to_dict())
[ "def", "do_hypervisor_stats", "(", "cs", ",", "args", ")", ":", "stats", "=", "cs", ".", "hypervisor_stats", ".", "statistics", "(", ")", "utils", ".", "print_dict", "(", "stats", ".", "to_dict", "(", ")", ")" ]
get hypervisor statistics over all compute nodes .
train
false
16,416
def KAMA(ds, count, timeperiod=(- (2 ** 31))): return call_talib_with_ds(ds, count, talib.KAMA, timeperiod)
[ "def", "KAMA", "(", "ds", ",", "count", ",", "timeperiod", "=", "(", "-", "(", "2", "**", "31", ")", ")", ")", ":", "return", "call_talib_with_ds", "(", "ds", ",", "count", ",", "talib", ".", "KAMA", ",", "timeperiod", ")" ]
kaufman adaptive moving average .
train
false
16,418
def CheckRequirements(filename): from pip.req import parse_requirements errors = [] for req in parse_requirements(filename): req.check_if_exists() if (not req.satisfied_by): errors.append(req) if errors: raise RuntimeError(('Requirements not installed: %s' % [str(e) for e in errors]))
[ "def", "CheckRequirements", "(", "filename", ")", ":", "from", "pip", ".", "req", "import", "parse_requirements", "errors", "=", "[", "]", "for", "req", "in", "parse_requirements", "(", "filename", ")", ":", "req", ".", "check_if_exists", "(", ")", "if", "(", "not", "req", ".", "satisfied_by", ")", ":", "errors", ".", "append", "(", "req", ")", "if", "errors", ":", "raise", "RuntimeError", "(", "(", "'Requirements not installed: %s'", "%", "[", "str", "(", "e", ")", "for", "e", "in", "errors", "]", ")", ")" ]
parse a pip requirements .
train
false
16,419
def _mod_bufsize_linux(iface, *args, **kwargs): ret = {'result': False, 'comment': 'Requires rx=<val> tx==<val> rx-mini=<val> and/or rx-jumbo=<val>'} cmd = ('/sbin/ethtool -G ' + iface) if (not kwargs): return ret if args: ret['comment'] = ('Unknown arguments: ' + ' '.join([str(item) for item in args])) return ret eargs = '' for kw in ['rx', 'tx', 'rx-mini', 'rx-jumbo']: value = kwargs.get(kw) if (value is not None): eargs += (((' ' + kw) + ' ') + str(value)) if (not eargs): return ret cmd += eargs out = __salt__['cmd.run'](cmd) if out: ret['comment'] = out else: ret['comment'] = eargs.strip() ret['result'] = True return ret
[ "def", "_mod_bufsize_linux", "(", "iface", ",", "*", "args", ",", "**", "kwargs", ")", ":", "ret", "=", "{", "'result'", ":", "False", ",", "'comment'", ":", "'Requires rx=<val> tx==<val> rx-mini=<val> and/or rx-jumbo=<val>'", "}", "cmd", "=", "(", "'/sbin/ethtool -G '", "+", "iface", ")", "if", "(", "not", "kwargs", ")", ":", "return", "ret", "if", "args", ":", "ret", "[", "'comment'", "]", "=", "(", "'Unknown arguments: '", "+", "' '", ".", "join", "(", "[", "str", "(", "item", ")", "for", "item", "in", "args", "]", ")", ")", "return", "ret", "eargs", "=", "''", "for", "kw", "in", "[", "'rx'", ",", "'tx'", ",", "'rx-mini'", ",", "'rx-jumbo'", "]", ":", "value", "=", "kwargs", ".", "get", "(", "kw", ")", "if", "(", "value", "is", "not", "None", ")", ":", "eargs", "+=", "(", "(", "(", "' '", "+", "kw", ")", "+", "' '", ")", "+", "str", "(", "value", ")", ")", "if", "(", "not", "eargs", ")", ":", "return", "ret", "cmd", "+=", "eargs", "out", "=", "__salt__", "[", "'cmd.run'", "]", "(", "cmd", ")", "if", "out", ":", "ret", "[", "'comment'", "]", "=", "out", "else", ":", "ret", "[", "'comment'", "]", "=", "eargs", ".", "strip", "(", ")", "ret", "[", "'result'", "]", "=", "True", "return", "ret" ]
modify network interface buffer sizes using ethtool .
train
true
16,420
def reset(): _runtime.reset()
[ "def", "reset", "(", ")", ":", "_runtime", ".", "reset", "(", ")" ]
set the enrollments and courses arrays to be empty .
train
false
16,421
def build_authenticate_header(realm=''): return {'WWW-Authenticate': ('OAuth realm="%s"' % realm)}
[ "def", "build_authenticate_header", "(", "realm", "=", "''", ")", ":", "return", "{", "'WWW-Authenticate'", ":", "(", "'OAuth realm=\"%s\"'", "%", "realm", ")", "}" ]
optional www-authenticate header .
train
false
16,422
def harvest_lettuces(only_the_apps=None, avoid_apps=None, path='features'): apps = get_apps() if (isinstance(only_the_apps, (list, tuple)) and any(only_the_apps)): def _filter_only_specified(module): return (module.__name__ in only_the_apps) apps = filter(_filter_only_specified, apps) else: apps = filter(_filter_bultins, apps) apps = filter(_filter_configured_apps, apps) apps = filter(_filter_configured_avoids, apps) if (isinstance(avoid_apps, (list, tuple)) and any(avoid_apps)): def _filter_avoid(module): return (module.__name__ not in avoid_apps) apps = filter(_filter_avoid, apps) joinpath = (lambda app: (join(dirname(app.__file__), path), app)) return map(joinpath, apps)
[ "def", "harvest_lettuces", "(", "only_the_apps", "=", "None", ",", "avoid_apps", "=", "None", ",", "path", "=", "'features'", ")", ":", "apps", "=", "get_apps", "(", ")", "if", "(", "isinstance", "(", "only_the_apps", ",", "(", "list", ",", "tuple", ")", ")", "and", "any", "(", "only_the_apps", ")", ")", ":", "def", "_filter_only_specified", "(", "module", ")", ":", "return", "(", "module", ".", "__name__", "in", "only_the_apps", ")", "apps", "=", "filter", "(", "_filter_only_specified", ",", "apps", ")", "else", ":", "apps", "=", "filter", "(", "_filter_bultins", ",", "apps", ")", "apps", "=", "filter", "(", "_filter_configured_apps", ",", "apps", ")", "apps", "=", "filter", "(", "_filter_configured_avoids", ",", "apps", ")", "if", "(", "isinstance", "(", "avoid_apps", ",", "(", "list", ",", "tuple", ")", ")", "and", "any", "(", "avoid_apps", ")", ")", ":", "def", "_filter_avoid", "(", "module", ")", ":", "return", "(", "module", ".", "__name__", "not", "in", "avoid_apps", ")", "apps", "=", "filter", "(", "_filter_avoid", ",", "apps", ")", "joinpath", "=", "(", "lambda", "app", ":", "(", "join", "(", "dirname", "(", "app", ".", "__file__", ")", ",", "path", ")", ",", "app", ")", ")", "return", "map", "(", "joinpath", ",", "apps", ")" ]
gets all installed apps that are not from django .
train
false
16,423
def _get_ico_surface(grade, patch_stats=False): from .bem import read_bem_surfaces ico_file_name = op.join(op.dirname(__file__), 'data', 'icos.fif.gz') ico = read_bem_surfaces(ico_file_name, patch_stats, s_id=(9000 + grade), verbose=False) return ico
[ "def", "_get_ico_surface", "(", "grade", ",", "patch_stats", "=", "False", ")", ":", "from", ".", "bem", "import", "read_bem_surfaces", "ico_file_name", "=", "op", ".", "join", "(", "op", ".", "dirname", "(", "__file__", ")", ",", "'data'", ",", "'icos.fif.gz'", ")", "ico", "=", "read_bem_surfaces", "(", "ico_file_name", ",", "patch_stats", ",", "s_id", "=", "(", "9000", "+", "grade", ")", ",", "verbose", "=", "False", ")", "return", "ico" ]
return an icosahedral surface of the desired grade .
train
false
16,425
def _GetIncludeDirs(config): include_dirs = (config.get('include_dirs', []) + config.get('msvs_system_include_dirs', [])) midl_include_dirs = (config.get('midl_include_dirs', []) + config.get('msvs_system_include_dirs', [])) resource_include_dirs = config.get('resource_include_dirs', include_dirs) include_dirs = _FixPaths(include_dirs) midl_include_dirs = _FixPaths(midl_include_dirs) resource_include_dirs = _FixPaths(resource_include_dirs) return (include_dirs, midl_include_dirs, resource_include_dirs)
[ "def", "_GetIncludeDirs", "(", "config", ")", ":", "include_dirs", "=", "(", "config", ".", "get", "(", "'include_dirs'", ",", "[", "]", ")", "+", "config", ".", "get", "(", "'msvs_system_include_dirs'", ",", "[", "]", ")", ")", "midl_include_dirs", "=", "(", "config", ".", "get", "(", "'midl_include_dirs'", ",", "[", "]", ")", "+", "config", ".", "get", "(", "'msvs_system_include_dirs'", ",", "[", "]", ")", ")", "resource_include_dirs", "=", "config", ".", "get", "(", "'resource_include_dirs'", ",", "include_dirs", ")", "include_dirs", "=", "_FixPaths", "(", "include_dirs", ")", "midl_include_dirs", "=", "_FixPaths", "(", "midl_include_dirs", ")", "resource_include_dirs", "=", "_FixPaths", "(", "resource_include_dirs", ")", "return", "(", "include_dirs", ",", "midl_include_dirs", ",", "resource_include_dirs", ")" ]
returns the list of directories to be used for #include directives .
train
false
16,426
def make_interp_full_matr(x, y, t, k): assert (x.size == y.size) assert (t.size == ((x.size + k) + 1)) n = x.size A = np.zeros((n, n), dtype=np.float_) for j in range(n): xval = x[j] if (xval == t[k]): left = k else: left = (np.searchsorted(t, xval) - 1) bb = _bspl.evaluate_all_bspl(t, k, xval, left) A[j, (left - k):(left + 1)] = bb c = sl.solve(A, y) return c
[ "def", "make_interp_full_matr", "(", "x", ",", "y", ",", "t", ",", "k", ")", ":", "assert", "(", "x", ".", "size", "==", "y", ".", "size", ")", "assert", "(", "t", ".", "size", "==", "(", "(", "x", ".", "size", "+", "k", ")", "+", "1", ")", ")", "n", "=", "x", ".", "size", "A", "=", "np", ".", "zeros", "(", "(", "n", ",", "n", ")", ",", "dtype", "=", "np", ".", "float_", ")", "for", "j", "in", "range", "(", "n", ")", ":", "xval", "=", "x", "[", "j", "]", "if", "(", "xval", "==", "t", "[", "k", "]", ")", ":", "left", "=", "k", "else", ":", "left", "=", "(", "np", ".", "searchsorted", "(", "t", ",", "xval", ")", "-", "1", ")", "bb", "=", "_bspl", ".", "evaluate_all_bspl", "(", "t", ",", "k", ",", "xval", ",", "left", ")", "A", "[", "j", ",", "(", "left", "-", "k", ")", ":", "(", "left", "+", "1", ")", "]", "=", "bb", "c", "=", "sl", ".", "solve", "(", "A", ",", "y", ")", "return", "c" ]
assemble an spline order k with knots t to interpolate y(x) using full matrices .
train
false
16,428
def rol(value, count): for y in range(count): value *= 2 if (value > 18446744073709551615L): value -= 18446744073709551616L value += 1 return value
[ "def", "rol", "(", "value", ",", "count", ")", ":", "for", "y", "in", "range", "(", "count", ")", ":", "value", "*=", "2", "if", "(", "value", ">", "18446744073709551615", "L", ")", ":", "value", "-=", "18446744073709551616", "L", "value", "+=", "1", "return", "value" ]
returns a rotation by k of n .
train
false
16,430
def _convert_for_comparison(self, other, equality_op=False): if isinstance(other, Decimal): return (self, other) if isinstance(other, _numbers.Rational): if (not self._is_special): self = _dec_from_triple(self._sign, str((int(self._int) * other.denominator)), self._exp) return (self, Decimal(other.numerator)) if (equality_op and isinstance(other, _numbers.Complex) and (other.imag == 0)): other = other.real if isinstance(other, float): context = getcontext() if equality_op: context.flags[FloatOperation] = 1 else: context._raise_error(FloatOperation, 'strict semantics for mixing floats and Decimals are enabled') return (self, Decimal.from_float(other)) return (NotImplemented, NotImplemented)
[ "def", "_convert_for_comparison", "(", "self", ",", "other", ",", "equality_op", "=", "False", ")", ":", "if", "isinstance", "(", "other", ",", "Decimal", ")", ":", "return", "(", "self", ",", "other", ")", "if", "isinstance", "(", "other", ",", "_numbers", ".", "Rational", ")", ":", "if", "(", "not", "self", ".", "_is_special", ")", ":", "self", "=", "_dec_from_triple", "(", "self", ".", "_sign", ",", "str", "(", "(", "int", "(", "self", ".", "_int", ")", "*", "other", ".", "denominator", ")", ")", ",", "self", ".", "_exp", ")", "return", "(", "self", ",", "Decimal", "(", "other", ".", "numerator", ")", ")", "if", "(", "equality_op", "and", "isinstance", "(", "other", ",", "_numbers", ".", "Complex", ")", "and", "(", "other", ".", "imag", "==", "0", ")", ")", ":", "other", "=", "other", ".", "real", "if", "isinstance", "(", "other", ",", "float", ")", ":", "context", "=", "getcontext", "(", ")", "if", "equality_op", ":", "context", ".", "flags", "[", "FloatOperation", "]", "=", "1", "else", ":", "context", ".", "_raise_error", "(", "FloatOperation", ",", "'strict semantics for mixing floats and Decimals are enabled'", ")", "return", "(", "self", ",", "Decimal", ".", "from_float", "(", "other", ")", ")", "return", "(", "NotImplemented", ",", "NotImplemented", ")" ]
given a decimal instance self and a python object other .
train
false
16,431
def make_cache_table(metadata, table_name='beaker_cache'): return sa.Table(table_name, metadata, sa.Column('namespace', sa.String(255), primary_key=True), sa.Column('accessed', sa.DateTime, nullable=False), sa.Column('created', sa.DateTime, nullable=False), sa.Column('data', sa.PickleType, nullable=False))
[ "def", "make_cache_table", "(", "metadata", ",", "table_name", "=", "'beaker_cache'", ")", ":", "return", "sa", ".", "Table", "(", "table_name", ",", "metadata", ",", "sa", ".", "Column", "(", "'namespace'", ",", "sa", ".", "String", "(", "255", ")", ",", "primary_key", "=", "True", ")", ",", "sa", ".", "Column", "(", "'accessed'", ",", "sa", ".", "DateTime", ",", "nullable", "=", "False", ")", ",", "sa", ".", "Column", "(", "'created'", ",", "sa", ".", "DateTime", ",", "nullable", "=", "False", ")", ",", "sa", ".", "Column", "(", "'data'", ",", "sa", ".", "PickleType", ",", "nullable", "=", "False", ")", ")" ]
return a table object suitable for storing cached values for the namespace manager .
train
false