id_within_dataset
int64
1
55.5k
snippet
stringlengths
19
14.2k
tokens
listlengths
6
1.63k
nl
stringlengths
6
352
split_within_dataset
stringclasses
1 value
is_duplicated
bool
2 classes
46,807
def ensure_vpn_forward(public_ip, port, private_ip): iptables_manager.ipv4['filter'].add_rule('FORWARD', ('-d %s -p udp --dport 1194 -j ACCEPT' % private_ip)) iptables_manager.ipv4['nat'].add_rule('PREROUTING', ('-d %s -p udp --dport %s -j DNAT --to %s:1194' % (public_ip, port, private_ip))) iptables_manager.ipv4['nat'].add_rule('OUTPUT', ('-d %s -p udp --dport %s -j DNAT --to %s:1194' % (public_ip, port, private_ip))) iptables_manager.apply()
[ "def", "ensure_vpn_forward", "(", "public_ip", ",", "port", ",", "private_ip", ")", ":", "iptables_manager", ".", "ipv4", "[", "'filter'", "]", ".", "add_rule", "(", "'FORWARD'", ",", "(", "'-d %s -p udp --dport 1194 -j ACCEPT'", "%", "private_ip", ")", ")", "iptables_manager", ".", "ipv4", "[", "'nat'", "]", ".", "add_rule", "(", "'PREROUTING'", ",", "(", "'-d %s -p udp --dport %s -j DNAT --to %s:1194'", "%", "(", "public_ip", ",", "port", ",", "private_ip", ")", ")", ")", "iptables_manager", ".", "ipv4", "[", "'nat'", "]", ".", "add_rule", "(", "'OUTPUT'", ",", "(", "'-d %s -p udp --dport %s -j DNAT --to %s:1194'", "%", "(", "public_ip", ",", "port", ",", "private_ip", ")", ")", ")", "iptables_manager", ".", "apply", "(", ")" ]
sets up forwarding rules for vlan .
train
false
46,808
def set_timeout(name, value, power='ac', scheme=None): ret = {'name': name, 'result': True, 'comment': '', 'changes': {}} comment = [] if (name not in ['monitor', 'disk', 'standby', 'hibernate']): ret['result'] = False comment.append('{0} is not a valid setting'.format(name)) elif (power not in ['ac', 'dc']): ret['result'] = False comment.append('{0} is not a power type'.format(power)) else: check_func = __salt__['powercfg.get_{0}_timeout'.format(name)] set_func = __salt__['powercfg.set_{0}_timeout'.format(name)] values = check_func(scheme=scheme) if (values[power] == value): comment.append('{0} {1} is already set with the value {2}.'.format(name, power, value)) else: ret['changes'] = {name: {power: value}} set_func(value, power, scheme=scheme) ret['comment'] = ' '.join(comment) return ret
[ "def", "set_timeout", "(", "name", ",", "value", ",", "power", "=", "'ac'", ",", "scheme", "=", "None", ")", ":", "ret", "=", "{", "'name'", ":", "name", ",", "'result'", ":", "True", ",", "'comment'", ":", "''", ",", "'changes'", ":", "{", "}", "}", "comment", "=", "[", "]", "if", "(", "name", "not", "in", "[", "'monitor'", ",", "'disk'", ",", "'standby'", ",", "'hibernate'", "]", ")", ":", "ret", "[", "'result'", "]", "=", "False", "comment", ".", "append", "(", "'{0} is not a valid setting'", ".", "format", "(", "name", ")", ")", "elif", "(", "power", "not", "in", "[", "'ac'", ",", "'dc'", "]", ")", ":", "ret", "[", "'result'", "]", "=", "False", "comment", ".", "append", "(", "'{0} is not a power type'", ".", "format", "(", "power", ")", ")", "else", ":", "check_func", "=", "__salt__", "[", "'powercfg.get_{0}_timeout'", ".", "format", "(", "name", ")", "]", "set_func", "=", "__salt__", "[", "'powercfg.set_{0}_timeout'", ".", "format", "(", "name", ")", "]", "values", "=", "check_func", "(", "scheme", "=", "scheme", ")", "if", "(", "values", "[", "power", "]", "==", "value", ")", ":", "comment", ".", "append", "(", "'{0} {1} is already set with the value {2}.'", ".", "format", "(", "name", ",", "power", ",", "value", ")", ")", "else", ":", "ret", "[", "'changes'", "]", "=", "{", "name", ":", "{", "power", ":", "value", "}", "}", "set_func", "(", "value", ",", "power", ",", "scheme", "=", "scheme", ")", "ret", "[", "'comment'", "]", "=", "' '", ".", "join", "(", "comment", ")", "return", "ret" ]
set the sleep timeouts of specific items such as disk .
train
false
46,809
def iterhosts(resources): for (module_name, key, resource) in resources: (resource_type, name) = key.split(u'.', 1) try: parser = PARSERS[resource_type] except KeyError: continue (yield parser(resource, module_name))
[ "def", "iterhosts", "(", "resources", ")", ":", "for", "(", "module_name", ",", "key", ",", "resource", ")", "in", "resources", ":", "(", "resource_type", ",", "name", ")", "=", "key", ".", "split", "(", "u'.'", ",", "1", ")", "try", ":", "parser", "=", "PARSERS", "[", "resource_type", "]", "except", "KeyError", ":", "continue", "(", "yield", "parser", "(", "resource", ",", "module_name", ")", ")" ]
yield host tuples of .
train
false
46,810
def directed_hausdorff(u, v, seed=0): u = np.asarray(u, dtype=np.float64, order='c') v = np.asarray(v, dtype=np.float64, order='c') result = _hausdorff.directed_hausdorff(u, v, seed) return result
[ "def", "directed_hausdorff", "(", "u", ",", "v", ",", "seed", "=", "0", ")", ":", "u", "=", "np", ".", "asarray", "(", "u", ",", "dtype", "=", "np", ".", "float64", ",", "order", "=", "'c'", ")", "v", "=", "np", ".", "asarray", "(", "v", ",", "dtype", "=", "np", ".", "float64", ",", "order", "=", "'c'", ")", "result", "=", "_hausdorff", ".", "directed_hausdorff", "(", "u", ",", "v", ",", "seed", ")", "return", "result" ]
computes the directed hausdorff distance between two n-d arrays .
train
false
46,811
def mount_volume(volume, mnt_base, configfile=None): fileutils.ensure_tree(mnt_base) command = ['mount.quobyte', volume, mnt_base] if configfile: command.extend(['-c', configfile]) LOG.debug('Mounting volume %s at mount point %s ...', volume, mnt_base) utils.execute(check_exit_code=[0, 4], *command) LOG.info(_LI('Mounted volume: %s'), volume)
[ "def", "mount_volume", "(", "volume", ",", "mnt_base", ",", "configfile", "=", "None", ")", ":", "fileutils", ".", "ensure_tree", "(", "mnt_base", ")", "command", "=", "[", "'mount.quobyte'", ",", "volume", ",", "mnt_base", "]", "if", "configfile", ":", "command", ".", "extend", "(", "[", "'-c'", ",", "configfile", "]", ")", "LOG", ".", "debug", "(", "'Mounting volume %s at mount point %s ...'", ",", "volume", ",", "mnt_base", ")", "utils", ".", "execute", "(", "check_exit_code", "=", "[", "0", ",", "4", "]", ",", "*", "command", ")", "LOG", ".", "info", "(", "_LI", "(", "'Mounted volume: %s'", ")", ",", "volume", ")" ]
wraps execute calls for mounting a quobyte volume .
train
false
46,812
def get_python_version(): return sys.version[:3]
[ "def", "get_python_version", "(", ")", ":", "return", "sys", ".", "version", "[", ":", "3", "]" ]
get a 3 element tuple with the python version .
train
false
46,813
def program_version_greater(ver1, ver2): v1f = version_to_float(ver1) v2f = version_to_float(ver2) return (v1f > v2f)
[ "def", "program_version_greater", "(", "ver1", ",", "ver2", ")", ":", "v1f", "=", "version_to_float", "(", "ver1", ")", "v2f", "=", "version_to_float", "(", "ver2", ")", "return", "(", "v1f", ">", "v2f", ")" ]
return true if ver1 > ver2 using semantics of comparing version numbers .
train
false
46,814
def _setHashDB(): if (not conf.hashDBFile): conf.hashDBFile = (conf.sessionFile or os.path.join(conf.outputPath, 'session.sqlite')) if os.path.exists(conf.hashDBFile): if conf.flushSession: try: os.remove(conf.hashDBFile) logger.info('flushing session file') except OSError as msg: errMsg = ('unable to flush the session file (%s)' % msg) raise SqlmapFilePathException(errMsg) conf.hashDB = HashDB(conf.hashDBFile)
[ "def", "_setHashDB", "(", ")", ":", "if", "(", "not", "conf", ".", "hashDBFile", ")", ":", "conf", ".", "hashDBFile", "=", "(", "conf", ".", "sessionFile", "or", "os", ".", "path", ".", "join", "(", "conf", ".", "outputPath", ",", "'session.sqlite'", ")", ")", "if", "os", ".", "path", ".", "exists", "(", "conf", ".", "hashDBFile", ")", ":", "if", "conf", ".", "flushSession", ":", "try", ":", "os", ".", "remove", "(", "conf", ".", "hashDBFile", ")", "logger", ".", "info", "(", "'flushing session file'", ")", "except", "OSError", "as", "msg", ":", "errMsg", "=", "(", "'unable to flush the session file (%s)'", "%", "msg", ")", "raise", "SqlmapFilePathException", "(", "errMsg", ")", "conf", ".", "hashDB", "=", "HashDB", "(", "conf", ".", "hashDBFile", ")" ]
check and set the hashdb sqlite file for query resume functionality .
train
false
46,815
@pytest.mark.cmd @pytest.mark.django_db def test_dump_stop_level(capfd): call_command('dump', '--stats', '--stop-level=1') (out, err) = capfd.readouterr() assert ('None,None' in out) assert out.startswith('/') assert ('/language0/project0/' in out) assert ('/language1/project0/' in out) assert ('/projects/project0/' in out) assert ('/language0/project0/store0.po' not in out) assert ('/language0/project0/subdir0' not in out)
[ "@", "pytest", ".", "mark", ".", "cmd", "@", "pytest", ".", "mark", ".", "django_db", "def", "test_dump_stop_level", "(", "capfd", ")", ":", "call_command", "(", "'dump'", ",", "'--stats'", ",", "'--stop-level=1'", ")", "(", "out", ",", "err", ")", "=", "capfd", ".", "readouterr", "(", ")", "assert", "(", "'None,None'", "in", "out", ")", "assert", "out", ".", "startswith", "(", "'/'", ")", "assert", "(", "'/language0/project0/'", "in", "out", ")", "assert", "(", "'/language1/project0/'", "in", "out", ")", "assert", "(", "'/projects/project0/'", "in", "out", ")", "assert", "(", "'/language0/project0/store0.po'", "not", "in", "out", ")", "assert", "(", "'/language0/project0/subdir0'", "not", "in", "out", ")" ]
set the depth for data .
train
false
46,816
@scopes.add_arg_scope def avg_pool(inputs, kernel_size, stride=2, padding='VALID', scope=None): with tf.name_scope(scope, 'AvgPool', [inputs]): (kernel_h, kernel_w) = _two_element_tuple(kernel_size) (stride_h, stride_w) = _two_element_tuple(stride) return tf.nn.avg_pool(inputs, ksize=[1, kernel_h, kernel_w, 1], strides=[1, stride_h, stride_w, 1], padding=padding)
[ "@", "scopes", ".", "add_arg_scope", "def", "avg_pool", "(", "inputs", ",", "kernel_size", ",", "stride", "=", "2", ",", "padding", "=", "'VALID'", ",", "scope", "=", "None", ")", ":", "with", "tf", ".", "name_scope", "(", "scope", ",", "'AvgPool'", ",", "[", "inputs", "]", ")", ":", "(", "kernel_h", ",", "kernel_w", ")", "=", "_two_element_tuple", "(", "kernel_size", ")", "(", "stride_h", ",", "stride_w", ")", "=", "_two_element_tuple", "(", "stride", ")", "return", "tf", ".", "nn", ".", "avg_pool", "(", "inputs", ",", "ksize", "=", "[", "1", ",", "kernel_h", ",", "kernel_w", ",", "1", "]", ",", "strides", "=", "[", "1", ",", "stride_h", ",", "stride_w", ",", "1", "]", ",", "padding", "=", "padding", ")" ]
adds a avg pooling layer .
train
false
46,817
def fake_disk_info_json(instance, type='qcow2'): disk_info = fake_disk_info_byname(instance, type) return jsonutils.dumps(disk_info.values())
[ "def", "fake_disk_info_json", "(", "instance", ",", "type", "=", "'qcow2'", ")", ":", "disk_info", "=", "fake_disk_info_byname", "(", "instance", ",", "type", ")", "return", "jsonutils", ".", "dumps", "(", "disk_info", ".", "values", "(", ")", ")" ]
return fake instance_disk_info corresponding accurately to the properties of the given instance object .
train
false
46,818
def ParseExpiration(expiration): delta = 0 for match in re.finditer(_DELTA_REGEX, expiration): amount = int(match.group(1)) units = _EXPIRATION_CONVERSIONS.get(match.group(2).lower(), 1) delta += (amount * units) return delta
[ "def", "ParseExpiration", "(", "expiration", ")", ":", "delta", "=", "0", "for", "match", "in", "re", ".", "finditer", "(", "_DELTA_REGEX", ",", "expiration", ")", ":", "amount", "=", "int", "(", "match", ".", "group", "(", "1", ")", ")", "units", "=", "_EXPIRATION_CONVERSIONS", ".", "get", "(", "match", ".", "group", "(", "2", ")", ".", "lower", "(", ")", ",", "1", ")", "delta", "+=", "(", "amount", "*", "units", ")", "return", "delta" ]
parses an expiration delta string .
train
false
46,819
def inplace_tanh_derivative(Z, delta): delta *= (1 - (Z ** 2))
[ "def", "inplace_tanh_derivative", "(", "Z", ",", "delta", ")", ":", "delta", "*=", "(", "1", "-", "(", "Z", "**", "2", ")", ")" ]
apply the derivative of the hyperbolic tanh function .
train
false
46,820
def SocketClient(address): family = address_type(address) with socket.socket(getattr(socket, family)) as s: s.setblocking(True) s.connect(address) return Connection(s.detach())
[ "def", "SocketClient", "(", "address", ")", ":", "family", "=", "address_type", "(", "address", ")", "with", "socket", ".", "socket", "(", "getattr", "(", "socket", ",", "family", ")", ")", "as", "s", ":", "s", ".", "setblocking", "(", "True", ")", "s", ".", "connect", "(", "address", ")", "return", "Connection", "(", "s", ".", "detach", "(", ")", ")" ]
return a connection object connected to the socket given by address .
train
false
46,821
def get_topic_similarities_as_csv(): output = StringIO.StringIO() writer = csv.writer(output) writer.writerow(RECOMMENDATION_CATEGORIES) topic_similarities = get_topic_similarities_dict() for topic in RECOMMENDATION_CATEGORIES: topic_similarities_row = [value for (_, value) in sorted(topic_similarities[topic].iteritems())] writer.writerow(topic_similarities_row) return output.getvalue()
[ "def", "get_topic_similarities_as_csv", "(", ")", ":", "output", "=", "StringIO", ".", "StringIO", "(", ")", "writer", "=", "csv", ".", "writer", "(", "output", ")", "writer", ".", "writerow", "(", "RECOMMENDATION_CATEGORIES", ")", "topic_similarities", "=", "get_topic_similarities_dict", "(", ")", "for", "topic", "in", "RECOMMENDATION_CATEGORIES", ":", "topic_similarities_row", "=", "[", "value", "for", "(", "_", ",", "value", ")", "in", "sorted", "(", "topic_similarities", "[", "topic", "]", ".", "iteritems", "(", ")", ")", "]", "writer", ".", "writerow", "(", "topic_similarities_row", ")", "return", "output", ".", "getvalue", "(", ")" ]
downloads all similarities corresponding to the current topics as a string which contains the contents of a csv file .
train
false
46,822
def cross_entropy_seq_with_mask(logits, target_seqs, input_mask, return_details=False): targets = tf.reshape(target_seqs, [(-1)]) weights = tf.to_float(tf.reshape(input_mask, [(-1)])) losses = tf.nn.sparse_softmax_cross_entropy_with_logits(logits, targets) loss = tf.div(tf.reduce_sum(tf.mul(losses, weights)), tf.reduce_sum(weights), name='seq_loss_with_mask') if return_details: return (loss, losses, weights, targets) else: return loss
[ "def", "cross_entropy_seq_with_mask", "(", "logits", ",", "target_seqs", ",", "input_mask", ",", "return_details", "=", "False", ")", ":", "targets", "=", "tf", ".", "reshape", "(", "target_seqs", ",", "[", "(", "-", "1", ")", "]", ")", "weights", "=", "tf", ".", "to_float", "(", "tf", ".", "reshape", "(", "input_mask", ",", "[", "(", "-", "1", ")", "]", ")", ")", "losses", "=", "tf", ".", "nn", ".", "sparse_softmax_cross_entropy_with_logits", "(", "logits", ",", "targets", ")", "loss", "=", "tf", ".", "div", "(", "tf", ".", "reduce_sum", "(", "tf", ".", "mul", "(", "losses", ",", "weights", ")", ")", ",", "tf", ".", "reduce_sum", "(", "weights", ")", ",", "name", "=", "'seq_loss_with_mask'", ")", "if", "return_details", ":", "return", "(", "loss", ",", "losses", ",", "weights", ",", "targets", ")", "else", ":", "return", "loss" ]
returns the expression of cross-entropy of two sequences .
train
true
46,823
def _get_destination(script_parts): for part in script_parts: if ((part not in {'ln', '-s', '--symbolic'}) and os.path.exists(part)): return part
[ "def", "_get_destination", "(", "script_parts", ")", ":", "for", "part", "in", "script_parts", ":", "if", "(", "(", "part", "not", "in", "{", "'ln'", ",", "'-s'", ",", "'--symbolic'", "}", ")", "and", "os", ".", "path", ".", "exists", "(", "part", ")", ")", ":", "return", "part" ]
when arguments order is wrong first argument will be destination .
train
true
46,824
def gaussian_random_matrix(n_components, n_features, random_state=None): _check_input_size(n_components, n_features) rng = check_random_state(random_state) components = rng.normal(loc=0.0, scale=(1.0 / np.sqrt(n_components)), size=(n_components, n_features)) return components
[ "def", "gaussian_random_matrix", "(", "n_components", ",", "n_features", ",", "random_state", "=", "None", ")", ":", "_check_input_size", "(", "n_components", ",", "n_features", ")", "rng", "=", "check_random_state", "(", "random_state", ")", "components", "=", "rng", ".", "normal", "(", "loc", "=", "0.0", ",", "scale", "=", "(", "1.0", "/", "np", ".", "sqrt", "(", "n_components", ")", ")", ",", "size", "=", "(", "n_components", ",", "n_features", ")", ")", "return", "components" ]
generate a dense gaussian random matrix .
train
false
46,826
def get_install_lng(): lng = 0 try: hive = _winreg.ConnectRegistry(None, _winreg.HKEY_LOCAL_MACHINE) key = _winreg.OpenKey(hive, 'Software\\SABnzbd') for i in range(0, _winreg.QueryInfoKey(key)[1]): (name, value, val_type) = _winreg.EnumValue(key, i) if (name == 'Installer Language'): lng = value _winreg.CloseKey(key) except WindowsError: pass finally: _winreg.CloseKey(hive) return lng
[ "def", "get_install_lng", "(", ")", ":", "lng", "=", "0", "try", ":", "hive", "=", "_winreg", ".", "ConnectRegistry", "(", "None", ",", "_winreg", ".", "HKEY_LOCAL_MACHINE", ")", "key", "=", "_winreg", ".", "OpenKey", "(", "hive", ",", "'Software\\\\SABnzbd'", ")", "for", "i", "in", "range", "(", "0", ",", "_winreg", ".", "QueryInfoKey", "(", "key", ")", "[", "1", "]", ")", ":", "(", "name", ",", "value", ",", "val_type", ")", "=", "_winreg", ".", "EnumValue", "(", "key", ",", "i", ")", "if", "(", "name", "==", "'Installer Language'", ")", ":", "lng", "=", "value", "_winreg", ".", "CloseKey", "(", "key", ")", "except", "WindowsError", ":", "pass", "finally", ":", "_winreg", ".", "CloseKey", "(", "hive", ")", "return", "lng" ]
return language-code used by the installer .
train
false
46,827
def GASpecificConfig(r, info): assert isinstance(info, DecoderSpecificInfo) r.skip(1) dependsOnCoreCoder = r.bits(1) if dependsOnCoreCoder: r.skip(14) extensionFlag = r.bits(1) if (not info.channelConfiguration): pce = ProgramConfigElement(r) info.pce_channels = pce.channels if ((info.audioObjectType == 6) or (info.audioObjectType == 20)): r.skip(3) if extensionFlag: if (info.audioObjectType == 22): r.skip((5 + 11)) if (info.audioObjectType in (17, 19, 20, 23)): r.skip(((1 + 1) + 1)) extensionFlag3 = r.bits(1) if (extensionFlag3 != 0): raise NotImplementedError('extensionFlag3 set')
[ "def", "GASpecificConfig", "(", "r", ",", "info", ")", ":", "assert", "isinstance", "(", "info", ",", "DecoderSpecificInfo", ")", "r", ".", "skip", "(", "1", ")", "dependsOnCoreCoder", "=", "r", ".", "bits", "(", "1", ")", "if", "dependsOnCoreCoder", ":", "r", ".", "skip", "(", "14", ")", "extensionFlag", "=", "r", ".", "bits", "(", "1", ")", "if", "(", "not", "info", ".", "channelConfiguration", ")", ":", "pce", "=", "ProgramConfigElement", "(", "r", ")", "info", ".", "pce_channels", "=", "pce", ".", "channels", "if", "(", "(", "info", ".", "audioObjectType", "==", "6", ")", "or", "(", "info", ".", "audioObjectType", "==", "20", ")", ")", ":", "r", ".", "skip", "(", "3", ")", "if", "extensionFlag", ":", "if", "(", "info", ".", "audioObjectType", "==", "22", ")", ":", "r", ".", "skip", "(", "(", "5", "+", "11", ")", ")", "if", "(", "info", ".", "audioObjectType", "in", "(", "17", ",", "19", ",", "20", ",", "23", ")", ")", ":", "r", ".", "skip", "(", "(", "(", "1", "+", "1", ")", "+", "1", ")", ")", "extensionFlag3", "=", "r", ".", "bits", "(", "1", ")", "if", "(", "extensionFlag3", "!=", "0", ")", ":", "raise", "NotImplementedError", "(", "'extensionFlag3 set'", ")" ]
reads gaspecificconfig which is needed to get the data after that and to read program_config_element which can contain channel counts .
train
true
46,828
def flipPoints(points, prefix, xmlElement): origin = evaluate.getVector3ByPrefix(Vector3(), (prefix + 'origin'), xmlElement) axis = evaluate.getVector3ByPrefix(Vector3(1.0, 0.0, 0.0), (prefix + 'axis'), xmlElement).getNormalized() for point in points: point.setToVector3((point - ((2.0 * axis.dot((point - origin))) * axis)))
[ "def", "flipPoints", "(", "points", ",", "prefix", ",", "xmlElement", ")", ":", "origin", "=", "evaluate", ".", "getVector3ByPrefix", "(", "Vector3", "(", ")", ",", "(", "prefix", "+", "'origin'", ")", ",", "xmlElement", ")", "axis", "=", "evaluate", ".", "getVector3ByPrefix", "(", "Vector3", "(", "1.0", ",", "0.0", ",", "0.0", ")", ",", "(", "prefix", "+", "'axis'", ")", ",", "xmlElement", ")", ".", "getNormalized", "(", ")", "for", "point", "in", "points", ":", "point", ".", "setToVector3", "(", "(", "point", "-", "(", "(", "2.0", "*", "axis", ".", "dot", "(", "(", "point", "-", "origin", ")", ")", ")", "*", "axis", ")", ")", ")" ]
flip the points .
train
false
46,829
def check_db(cursor): _check_balances(cursor) _check_no_team_balances(cursor) _check_tips(cursor) _check_orphans(cursor) _check_orphans_no_tips(cursor)
[ "def", "check_db", "(", "cursor", ")", ":", "_check_balances", "(", "cursor", ")", "_check_no_team_balances", "(", "cursor", ")", "_check_tips", "(", "cursor", ")", "_check_orphans", "(", "cursor", ")", "_check_orphans_no_tips", "(", "cursor", ")" ]
runs all available self checks on the given cursor .
train
false
46,830
def json_mapping_parser(path, template): mapping_blocks = {} with open(path, 'r') as handle: for (tid, rows) in json.load(handle).iteritems(): mappings = {} for (key, values) in rows.iteritems(): mapping = {template.get(k, k): v for (k, v) in values.iteritems()} mappings[int(key)] = mapping mapping_blocks[tid] = mappings return mapping_blocks
[ "def", "json_mapping_parser", "(", "path", ",", "template", ")", ":", "mapping_blocks", "=", "{", "}", "with", "open", "(", "path", ",", "'r'", ")", "as", "handle", ":", "for", "(", "tid", ",", "rows", ")", "in", "json", ".", "load", "(", "handle", ")", ".", "iteritems", "(", ")", ":", "mappings", "=", "{", "}", "for", "(", "key", ",", "values", ")", "in", "rows", ".", "iteritems", "(", ")", ":", "mapping", "=", "{", "template", ".", "get", "(", "k", ",", "k", ")", ":", "v", "for", "(", "k", ",", "v", ")", "in", "values", ".", "iteritems", "(", ")", "}", "mappings", "[", "int", "(", "key", ")", "]", "=", "mapping", "mapping_blocks", "[", "tid", "]", "=", "mappings", "return", "mapping_blocks" ]
given a json file of the the mapping data for a modbus device .
train
false
46,832
def _get_ansi_code(color=None, style=None): ansi_code = [] if style: style_attrs = splitstrip(style) for effect in style_attrs: ansi_code.append(ANSI_STYLES[effect]) if color: if color.isdigit(): ansi_code.extend(['38', '5']) ansi_code.append(color) else: ansi_code.append(ANSI_COLORS[color]) if ansi_code: return ((ANSI_PREFIX + ';'.join(ansi_code)) + ANSI_END) return ''
[ "def", "_get_ansi_code", "(", "color", "=", "None", ",", "style", "=", "None", ")", ":", "ansi_code", "=", "[", "]", "if", "style", ":", "style_attrs", "=", "splitstrip", "(", "style", ")", "for", "effect", "in", "style_attrs", ":", "ansi_code", ".", "append", "(", "ANSI_STYLES", "[", "effect", "]", ")", "if", "color", ":", "if", "color", ".", "isdigit", "(", ")", ":", "ansi_code", ".", "extend", "(", "[", "'38'", ",", "'5'", "]", ")", "ansi_code", ".", "append", "(", "color", ")", "else", ":", "ansi_code", ".", "append", "(", "ANSI_COLORS", "[", "color", "]", ")", "if", "ansi_code", ":", "return", "(", "(", "ANSI_PREFIX", "+", "';'", ".", "join", "(", "ansi_code", ")", ")", "+", "ANSI_END", ")", "return", "''" ]
return ansi escape code corresponding to color and style :type color: str or none .
train
true
46,833
def exactly_one(l): return exactly_n(l)
[ "def", "exactly_one", "(", "l", ")", ":", "return", "exactly_n", "(", "l", ")" ]
check if only one item is not none .
train
false
46,835
def ascii_lower(string): return string.encode(u'utf8').lower().decode(u'utf8')
[ "def", "ascii_lower", "(", "string", ")", ":", "return", "string", ".", "encode", "(", "u'utf8'", ")", ".", "lower", "(", ")", ".", "decode", "(", "u'utf8'", ")" ]
transform ascii letters to lower case: a-z is mapped to a-z .
train
false
46,836
def identity(x): return x
[ "def", "identity", "(", "x", ")", ":", "return", "x" ]
simply return the input array .
train
false
46,837
def _fastq_sanger_convert_fastq_sanger(in_handle, out_handle, alphabet=None): mapping = ''.join((([chr(0) for ascii in range(0, 33)] + [chr(ascii) for ascii in range(33, 127)]) + [chr(0) for ascii in range(127, 256)])) assert (len(mapping) == 256) return _fastq_generic(in_handle, out_handle, mapping)
[ "def", "_fastq_sanger_convert_fastq_sanger", "(", "in_handle", ",", "out_handle", ",", "alphabet", "=", "None", ")", ":", "mapping", "=", "''", ".", "join", "(", "(", "(", "[", "chr", "(", "0", ")", "for", "ascii", "in", "range", "(", "0", ",", "33", ")", "]", "+", "[", "chr", "(", "ascii", ")", "for", "ascii", "in", "range", "(", "33", ",", "127", ")", "]", ")", "+", "[", "chr", "(", "0", ")", "for", "ascii", "in", "range", "(", "127", ",", "256", ")", "]", ")", ")", "assert", "(", "len", "(", "mapping", ")", "==", "256", ")", "return", "_fastq_generic", "(", "in_handle", ",", "out_handle", ",", "mapping", ")" ]
fast sanger fastq to sanger fastq conversion .
train
false
46,838
def all_queries(fn, obj, *param_lists): results = [] params = [[obj]] for pl in param_lists: new_params = [] for p in pl: for c in params: new_param = list(c) new_param.append(p) new_params.append(new_param) params = new_params results = [fn(*p) for p in params] return results
[ "def", "all_queries", "(", "fn", ",", "obj", ",", "*", "param_lists", ")", ":", "results", "=", "[", "]", "params", "=", "[", "[", "obj", "]", "]", "for", "pl", "in", "param_lists", ":", "new_params", "=", "[", "]", "for", "p", "in", "pl", ":", "for", "c", "in", "params", ":", "new_param", "=", "list", "(", "c", ")", "new_param", ".", "append", "(", "p", ")", "new_params", ".", "append", "(", "new_param", ")", "params", "=", "new_params", "results", "=", "[", "fn", "(", "*", "p", ")", "for", "p", "in", "params", "]", "return", "results" ]
given a fn and a first argument obj .
train
false
46,839
@core_helper @maintain.deprecated("h.subnav_link is deprecated please use h.nav_link\nNOTE: if action is passed as the second parameter make sure it is passed as a named parameter eg. `action='my_action'") def subnav_link(text, action, **kwargs): kwargs['action'] = action return nav_link(text, **kwargs)
[ "@", "core_helper", "@", "maintain", ".", "deprecated", "(", "\"h.subnav_link is deprecated please use h.nav_link\\nNOTE: if action is passed as the second parameter make sure it is passed as a named parameter eg. `action='my_action'\"", ")", "def", "subnav_link", "(", "text", ",", "action", ",", "**", "kwargs", ")", ":", "kwargs", "[", "'action'", "]", "=", "action", "return", "nav_link", "(", "text", ",", "**", "kwargs", ")" ]
create a link for a named route .
train
false
46,840
def translate_masks(modifiers): masks = [] for i in modifiers: try: masks.append(xcbq.ModMasks[i]) except KeyError: raise KeyError(('Unknown modifier: %s' % i)) if masks: return reduce(operator.or_, masks) else: return 0
[ "def", "translate_masks", "(", "modifiers", ")", ":", "masks", "=", "[", "]", "for", "i", "in", "modifiers", ":", "try", ":", "masks", ".", "append", "(", "xcbq", ".", "ModMasks", "[", "i", "]", ")", "except", "KeyError", ":", "raise", "KeyError", "(", "(", "'Unknown modifier: %s'", "%", "i", ")", ")", "if", "masks", ":", "return", "reduce", "(", "operator", ".", "or_", ",", "masks", ")", "else", ":", "return", "0" ]
translate a modifier mask specified as a list of strings into an or-ed bit representation .
train
false
46,841
def optimizeAngle(angle): if (angle < 0): angle %= (-360) else: angle %= 360 if (angle >= 270): angle -= 360 elif (angle < (-90)): angle += 360 return angle
[ "def", "optimizeAngle", "(", "angle", ")", ":", "if", "(", "angle", "<", "0", ")", ":", "angle", "%=", "(", "-", "360", ")", "else", ":", "angle", "%=", "360", "if", "(", "angle", ">=", "270", ")", ":", "angle", "-=", "360", "elif", "(", "angle", "<", "(", "-", "90", ")", ")", ":", "angle", "+=", "360", "return", "angle" ]
because any rotation can be expressed within 360 degrees of any given number .
train
true
46,842
def _create_dicts(): short2rgb_dict = dict(CLUT) rgb2short_dict = {} for (k, v) in short2rgb_dict.items(): rgb2short_dict[v] = k return (rgb2short_dict, short2rgb_dict)
[ "def", "_create_dicts", "(", ")", ":", "short2rgb_dict", "=", "dict", "(", "CLUT", ")", "rgb2short_dict", "=", "{", "}", "for", "(", "k", ",", "v", ")", "in", "short2rgb_dict", ".", "items", "(", ")", ":", "rgb2short_dict", "[", "v", "]", "=", "k", "return", "(", "rgb2short_dict", ",", "short2rgb_dict", ")" ]
create dictionary .
train
true
46,845
def _responseFromMessage(responseConstructor, message, **kwargs): response = responseConstructor(id=message.id, answer=True, **kwargs) response.queries = message.queries[:] return response
[ "def", "_responseFromMessage", "(", "responseConstructor", ",", "message", ",", "**", "kwargs", ")", ":", "response", "=", "responseConstructor", "(", "id", "=", "message", ".", "id", ",", "answer", "=", "True", ",", "**", "kwargs", ")", "response", ".", "queries", "=", "message", ".", "queries", "[", ":", "]", "return", "response" ]
generate a l{message} like instance suitable for use as the response to c{message} .
train
false
46,847
def only_if_coordinator(func): def wrapper(*args, **kwargs): 'Decorator wrapper.' if args[0].is_coordinator: from soco.exceptions import SoCoUPnPException try: func(*args, **kwargs) except SoCoUPnPException: _LOGGER.error('command "%s" for Sonos device "%s" not available in this mode', func.__name__, args[0].name) else: _LOGGER.debug('Ignore command "%s" for Sonos device "%s" (%s)', func.__name__, args[0].name, 'not coordinator') return wrapper
[ "def", "only_if_coordinator", "(", "func", ")", ":", "def", "wrapper", "(", "*", "args", ",", "**", "kwargs", ")", ":", "if", "args", "[", "0", "]", ".", "is_coordinator", ":", "from", "soco", ".", "exceptions", "import", "SoCoUPnPException", "try", ":", "func", "(", "*", "args", ",", "**", "kwargs", ")", "except", "SoCoUPnPException", ":", "_LOGGER", ".", "error", "(", "'command \"%s\" for Sonos device \"%s\" not available in this mode'", ",", "func", ".", "__name__", ",", "args", "[", "0", "]", ".", "name", ")", "else", ":", "_LOGGER", ".", "debug", "(", "'Ignore command \"%s\" for Sonos device \"%s\" (%s)'", ",", "func", ".", "__name__", ",", "args", "[", "0", "]", ".", "name", ",", "'not coordinator'", ")", "return", "wrapper" ]
decorator for coordinator .
train
false
46,848
@register.filter def microsite_template_path(template_name): template_name = theming_helpers.get_template_path(template_name) return (template_name[1:] if (template_name[0] == '/') else template_name)
[ "@", "register", ".", "filter", "def", "microsite_template_path", "(", "template_name", ")", ":", "template_name", "=", "theming_helpers", ".", "get_template_path", "(", "template_name", ")", "return", "(", "template_name", "[", "1", ":", "]", "if", "(", "template_name", "[", "0", "]", "==", "'/'", ")", "else", "template_name", ")" ]
django template filter to apply template overriding to microsites .
train
false
46,852
def test_login_logout(client): rv = register_and_login(client, 'user1', 'default') assert ('You were logged in' in rv.data) rv = logout(client) assert ('You were logged out' in rv.data) rv = login(client, 'user1', 'wrongpassword') assert ('Invalid password' in rv.data) rv = login(client, 'user2', 'wrongpassword') assert ('Invalid username' in rv.data)
[ "def", "test_login_logout", "(", "client", ")", ":", "rv", "=", "register_and_login", "(", "client", ",", "'user1'", ",", "'default'", ")", "assert", "(", "'You were logged in'", "in", "rv", ".", "data", ")", "rv", "=", "logout", "(", "client", ")", "assert", "(", "'You were logged out'", "in", "rv", ".", "data", ")", "rv", "=", "login", "(", "client", ",", "'user1'", ",", "'wrongpassword'", ")", "assert", "(", "'Invalid password'", "in", "rv", ".", "data", ")", "rv", "=", "login", "(", "client", ",", "'user2'", ",", "'wrongpassword'", ")", "assert", "(", "'Invalid username'", "in", "rv", ".", "data", ")" ]
make sure logging in and logging out works .
train
false
46,853
def WriteAutoRegenerationRule(params, root_makefile, makefile_name, build_files): options = params['options'] build_files_args = [gyp.common.RelativePath(filename, options.toplevel_dir) for filename in params['build_files_arg']] gyp_binary = gyp.common.FixIfRelativePath(params['gyp_binary'], options.toplevel_dir) if (not gyp_binary.startswith(os.sep)): gyp_binary = os.path.join('.', gyp_binary) root_makefile.write(('quiet_cmd_regen_makefile = ACTION Regenerating $@\ncmd_regen_makefile = cd $(srcdir); %(cmd)s\n%(makefile_name)s: %(deps)s\n DCTB $(call do_cmd,regen_makefile)\n\n' % {'makefile_name': makefile_name, 'deps': ' '.join(map(Sourceify, build_files)), 'cmd': gyp.common.EncodePOSIXShellList((([gyp_binary, '-fmake'] + gyp.RegenerateFlags(options)) + build_files_args))}))
[ "def", "WriteAutoRegenerationRule", "(", "params", ",", "root_makefile", ",", "makefile_name", ",", "build_files", ")", ":", "options", "=", "params", "[", "'options'", "]", "build_files_args", "=", "[", "gyp", ".", "common", ".", "RelativePath", "(", "filename", ",", "options", ".", "toplevel_dir", ")", "for", "filename", "in", "params", "[", "'build_files_arg'", "]", "]", "gyp_binary", "=", "gyp", ".", "common", ".", "FixIfRelativePath", "(", "params", "[", "'gyp_binary'", "]", ",", "options", ".", "toplevel_dir", ")", "if", "(", "not", "gyp_binary", ".", "startswith", "(", "os", ".", "sep", ")", ")", ":", "gyp_binary", "=", "os", ".", "path", ".", "join", "(", "'.'", ",", "gyp_binary", ")", "root_makefile", ".", "write", "(", "(", "'quiet_cmd_regen_makefile = ACTION Regenerating $@\\ncmd_regen_makefile = cd $(srcdir); %(cmd)s\\n%(makefile_name)s: %(deps)s\\n DCTB $(call do_cmd,regen_makefile)\\n\\n'", "%", "{", "'makefile_name'", ":", "makefile_name", ",", "'deps'", ":", "' '", ".", "join", "(", "map", "(", "Sourceify", ",", "build_files", ")", ")", ",", "'cmd'", ":", "gyp", ".", "common", ".", "EncodePOSIXShellList", "(", "(", "(", "[", "gyp_binary", ",", "'-fmake'", "]", "+", "gyp", ".", "RegenerateFlags", "(", "options", ")", ")", "+", "build_files_args", ")", ")", "}", ")", ")" ]
write the target to regenerate the makefile .
train
false
46,854
@cleanup def test_log_scatter(): (fig, ax) = plt.subplots(1) x = np.arange(10) y = (np.arange(10) - 1) ax.scatter(x, y) buf = io.BytesIO() fig.savefig(buf, format='pdf') buf = io.BytesIO() fig.savefig(buf, format='eps') buf = io.BytesIO() fig.savefig(buf, format='svg')
[ "@", "cleanup", "def", "test_log_scatter", "(", ")", ":", "(", "fig", ",", "ax", ")", "=", "plt", ".", "subplots", "(", "1", ")", "x", "=", "np", ".", "arange", "(", "10", ")", "y", "=", "(", "np", ".", "arange", "(", "10", ")", "-", "1", ")", "ax", ".", "scatter", "(", "x", ",", "y", ")", "buf", "=", "io", ".", "BytesIO", "(", ")", "fig", ".", "savefig", "(", "buf", ",", "format", "=", "'pdf'", ")", "buf", "=", "io", ".", "BytesIO", "(", ")", "fig", ".", "savefig", "(", "buf", ",", "format", "=", "'eps'", ")", "buf", "=", "io", ".", "BytesIO", "(", ")", "fig", ".", "savefig", "(", "buf", ",", "format", "=", "'svg'", ")" ]
issue #1799 .
train
false
46,855
def get_unsupported_upper_protocol(): if (Version(CASSANDRA_VERSION) >= Version('2.2')): return None if (Version(CASSANDRA_VERSION) >= Version('2.1')): return 4 elif (Version(CASSANDRA_VERSION) >= Version('2.0')): return 3 else: return None
[ "def", "get_unsupported_upper_protocol", "(", ")", ":", "if", "(", "Version", "(", "CASSANDRA_VERSION", ")", ">=", "Version", "(", "'2.2'", ")", ")", ":", "return", "None", "if", "(", "Version", "(", "CASSANDRA_VERSION", ")", ">=", "Version", "(", "'2.1'", ")", ")", ":", "return", "4", "elif", "(", "Version", "(", "CASSANDRA_VERSION", ")", ">=", "Version", "(", "'2.0'", ")", ")", ":", "return", "3", "else", ":", "return", "None" ]
this is used to determine the highest protocol version that is not supported by the version of c* running .
train
false
46,856
def get_include_dirs(): include_dirs = [os.path.join(d, 'include') for d in get_base_dirs()] if (sys.platform != 'win32'): include_dirs.extend(os.environ.get('CPLUS_INCLUDE_PATH', '').split(os.pathsep)) return include_dirs
[ "def", "get_include_dirs", "(", ")", ":", "include_dirs", "=", "[", "os", ".", "path", ".", "join", "(", "d", ",", "'include'", ")", "for", "d", "in", "get_base_dirs", "(", ")", "]", "if", "(", "sys", ".", "platform", "!=", "'win32'", ")", ":", "include_dirs", ".", "extend", "(", "os", ".", "environ", ".", "get", "(", "'CPLUS_INCLUDE_PATH'", ",", "''", ")", ".", "split", "(", "os", ".", "pathsep", ")", ")", "return", "include_dirs" ]
returns a list of standard include directories on this platform .
train
false
46,857
def remove_extension(template): return template[:(- len('.template'))]
[ "def", "remove_extension", "(", "template", ")", ":", "return", "template", "[", ":", "(", "-", "len", "(", "'.template'", ")", ")", "]" ]
given a filename or path of a template file .
train
false
46,858
def Frechet(name, a, s=1, m=0): return rv(name, FrechetDistribution, (a, s, m))
[ "def", "Frechet", "(", "name", ",", "a", ",", "s", "=", "1", ",", "m", "=", "0", ")", ":", "return", "rv", "(", "name", ",", "FrechetDistribution", ",", "(", "a", ",", "s", ",", "m", ")", ")" ]
create a continuous random variable with a frechet distribution .
train
false
46,860
@register.filter(is_safe=True) @stringfilter def iriencode(value): return force_text(iri_to_uri(value))
[ "@", "register", ".", "filter", "(", "is_safe", "=", "True", ")", "@", "stringfilter", "def", "iriencode", "(", "value", ")", ":", "return", "force_text", "(", "iri_to_uri", "(", "value", ")", ")" ]
escapes an iri value for use in a url .
train
false
46,861
def BetaPrime(name, alpha, beta): return rv(name, BetaPrimeDistribution, (alpha, beta))
[ "def", "BetaPrime", "(", "name", ",", "alpha", ",", "beta", ")", ":", "return", "rv", "(", "name", ",", "BetaPrimeDistribution", ",", "(", "alpha", ",", "beta", ")", ")" ]
create a continuous random variable with a beta prime distribution .
train
false
46,862
def get_field_type(model, fieldname): field = getattr(model, fieldname) if isinstance(field, ColumnElement): return field.type if isinstance(field, AssociationProxy): field = field.remote_attr if hasattr(field, 'property'): prop = field.property if isinstance(prop, RelProperty): return None return prop.columns[0].type return None
[ "def", "get_field_type", "(", "model", ",", "fieldname", ")", ":", "field", "=", "getattr", "(", "model", ",", "fieldname", ")", "if", "isinstance", "(", "field", ",", "ColumnElement", ")", ":", "return", "field", ".", "type", "if", "isinstance", "(", "field", ",", "AssociationProxy", ")", ":", "field", "=", "field", ".", "remote_attr", "if", "hasattr", "(", "field", ",", "'property'", ")", ":", "prop", "=", "field", ".", "property", "if", "isinstance", "(", "prop", ",", "RelProperty", ")", ":", "return", "None", "return", "prop", ".", "columns", "[", "0", "]", ".", "type", "return", "None" ]
given the database connection .
train
false
46,863
def get_user_model_path(): return getattr(settings, u'AUTH_USER_MODEL', u'auth.User')
[ "def", "get_user_model_path", "(", ")", ":", "return", "getattr", "(", "settings", ",", "u'AUTH_USER_MODEL'", ",", "u'auth.User'", ")" ]
returns app_label .
train
false
46,864
def test_qapp_name(qapp): assert (qapp.applicationName() == 'qute_test')
[ "def", "test_qapp_name", "(", "qapp", ")", ":", "assert", "(", "qapp", ".", "applicationName", "(", ")", "==", "'qute_test'", ")" ]
make sure the qapplication name is changed when we use qapp .
train
false
46,865
@register(u'unix-word-rubout') def unix_word_rubout(event, WORD=True): buff = event.current_buffer pos = buff.document.find_start_of_previous_word(count=event.arg, WORD=WORD) if (pos is None): pos = (- buff.cursor_position) if pos: deleted = buff.delete_before_cursor(count=(- pos)) if event.is_repeat: deleted += event.cli.clipboard.get_data().text event.cli.clipboard.set_text(deleted) else: event.cli.output.bell()
[ "@", "register", "(", "u'unix-word-rubout'", ")", "def", "unix_word_rubout", "(", "event", ",", "WORD", "=", "True", ")", ":", "buff", "=", "event", ".", "current_buffer", "pos", "=", "buff", ".", "document", ".", "find_start_of_previous_word", "(", "count", "=", "event", ".", "arg", ",", "WORD", "=", "WORD", ")", "if", "(", "pos", "is", "None", ")", ":", "pos", "=", "(", "-", "buff", ".", "cursor_position", ")", "if", "pos", ":", "deleted", "=", "buff", ".", "delete_before_cursor", "(", "count", "=", "(", "-", "pos", ")", ")", "if", "event", ".", "is_repeat", ":", "deleted", "+=", "event", ".", "cli", ".", "clipboard", ".", "get_data", "(", ")", ".", "text", "event", ".", "cli", ".", "clipboard", ".", "set_text", "(", "deleted", ")", "else", ":", "event", ".", "cli", ".", "output", ".", "bell", "(", ")" ]
kill the word behind point .
train
true
46,867
def visit_snippet_literal(self, node): self.visit_literal_block(node)
[ "def", "visit_snippet_literal", "(", "self", ",", "node", ")", ":", "self", ".", "visit_literal_block", "(", "node", ")" ]
default literal block handler .
train
false
46,868
def list_entries(logger_name): logging_client = logging.Client() logger = logging_client.logger(logger_name) print 'Listing entries for logger {}:'.format(logger.name) for entry in logger.list_entries(): timestamp = entry.timestamp.isoformat() print '* {}: {}'.format(timestamp, entry.payload)
[ "def", "list_entries", "(", "logger_name", ")", ":", "logging_client", "=", "logging", ".", "Client", "(", ")", "logger", "=", "logging_client", ".", "logger", "(", "logger_name", ")", "print", "'Listing entries for logger {}:'", ".", "format", "(", "logger", ".", "name", ")", "for", "entry", "in", "logger", ".", "list_entries", "(", ")", ":", "timestamp", "=", "entry", ".", "timestamp", ".", "isoformat", "(", ")", "print", "'* {}: {}'", ".", "format", "(", "timestamp", ",", "entry", ".", "payload", ")" ]
lists the most recent entries for a given logger .
train
false
46,869
@utils.arg('server', metavar='<server>', help=_('Name or ID of server.')) @utils.arg('action', metavar='<action>', choices=['set', 'delete'], help=_("Actions: 'set' or 'delete'.")) @utils.arg('metadata', metavar='<key=value>', nargs='+', action='append', default=[], help=_('Metadata to set or delete (only key is necessary on delete).')) def do_meta(cs, args): server = _find_server(cs, args.server) metadata = _extract_metadata(args) if (args.action == 'set'): cs.servers.set_meta(server, metadata) elif (args.action == 'delete'): cs.servers.delete_meta(server, sorted(metadata.keys(), reverse=True))
[ "@", "utils", ".", "arg", "(", "'server'", ",", "metavar", "=", "'<server>'", ",", "help", "=", "_", "(", "'Name or ID of server.'", ")", ")", "@", "utils", ".", "arg", "(", "'action'", ",", "metavar", "=", "'<action>'", ",", "choices", "=", "[", "'set'", ",", "'delete'", "]", ",", "help", "=", "_", "(", "\"Actions: 'set' or 'delete'.\"", ")", ")", "@", "utils", ".", "arg", "(", "'metadata'", ",", "metavar", "=", "'<key=value>'", ",", "nargs", "=", "'+'", ",", "action", "=", "'append'", ",", "default", "=", "[", "]", ",", "help", "=", "_", "(", "'Metadata to set or delete (only key is necessary on delete).'", ")", ")", "def", "do_meta", "(", "cs", ",", "args", ")", ":", "server", "=", "_find_server", "(", "cs", ",", "args", ".", "server", ")", "metadata", "=", "_extract_metadata", "(", "args", ")", "if", "(", "args", ".", "action", "==", "'set'", ")", ":", "cs", ".", "servers", ".", "set_meta", "(", "server", ",", "metadata", ")", "elif", "(", "args", ".", "action", "==", "'delete'", ")", ":", "cs", ".", "servers", ".", "delete_meta", "(", "server", ",", "sorted", "(", "metadata", ".", "keys", "(", ")", ",", "reverse", "=", "True", ")", ")" ]
set or delete metadata on a server .
train
false
46,870
def Clf(): global LOC LOC = None _Brewer.ClearIter() pyplot.clf() fig = pyplot.gcf() fig.set_size_inches(8, 6)
[ "def", "Clf", "(", ")", ":", "global", "LOC", "LOC", "=", "None", "_Brewer", ".", "ClearIter", "(", ")", "pyplot", ".", "clf", "(", ")", "fig", "=", "pyplot", ".", "gcf", "(", ")", "fig", ".", "set_size_inches", "(", "8", ",", "6", ")" ]
clears the figure and any hints that have been set .
train
false
46,871
def check_url(url): good_codes = [httplib.OK, httplib.FOUND, httplib.MOVED_PERMANENTLY] return (_get_server_status_code(url) in good_codes)
[ "def", "check_url", "(", "url", ")", ":", "good_codes", "=", "[", "httplib", ".", "OK", ",", "httplib", ".", "FOUND", ",", "httplib", ".", "MOVED_PERMANENTLY", "]", "return", "(", "_get_server_status_code", "(", "url", ")", "in", "good_codes", ")" ]
check if a url exists without downloading the whole file .
train
false
46,873
def salt_master(): import salt.cli.daemons master = salt.cli.daemons.Master() master.start()
[ "def", "salt_master", "(", ")", ":", "import", "salt", ".", "cli", ".", "daemons", "master", "=", "salt", ".", "cli", ".", "daemons", ".", "Master", "(", ")", "master", ".", "start", "(", ")" ]
start the salt master .
train
false
46,874
def get_alias(ip): hosts = _list_hosts() if (ip in hosts): return hosts[ip] return []
[ "def", "get_alias", "(", "ip", ")", ":", "hosts", "=", "_list_hosts", "(", ")", "if", "(", "ip", "in", "hosts", ")", ":", "return", "hosts", "[", "ip", "]", "return", "[", "]" ]
return the list of aliases associated with an ip aliases are returned in the order in which they appear in the hosts file .
train
false
46,876
def signal_handler(signal, frame): logging.warning('Caught signal: {0}'.format(signal)) IOLoop.instance().add_callback(shutdown)
[ "def", "signal_handler", "(", "signal", ",", "frame", ")", ":", "logging", ".", "warning", "(", "'Caught signal: {0}'", ".", "format", "(", "signal", ")", ")", "IOLoop", ".", "instance", "(", ")", ".", "add_callback", "(", "shutdown", ")" ]
signal handler for graceful shutdown .
train
false
46,877
def send_saml_audit_notification(action, request, user_id, group_ids, identity_provider, protocol, token_id, outcome): initiator = request.audit_initiator target = resource.Resource(typeURI=taxonomy.ACCOUNT_USER) audit_type = SAML_AUDIT_TYPE user_id = (user_id or taxonomy.UNKNOWN) token_id = (token_id or taxonomy.UNKNOWN) group_ids = (group_ids or []) cred = credential.FederatedCredential(token=token_id, type=audit_type, identity_provider=identity_provider, user=user_id, groups=group_ids) initiator.credential = cred event_type = ('%s.%s' % (SERVICE, action)) _send_audit_notification(action, initiator, outcome, target, event_type)
[ "def", "send_saml_audit_notification", "(", "action", ",", "request", ",", "user_id", ",", "group_ids", ",", "identity_provider", ",", "protocol", ",", "token_id", ",", "outcome", ")", ":", "initiator", "=", "request", ".", "audit_initiator", "target", "=", "resource", ".", "Resource", "(", "typeURI", "=", "taxonomy", ".", "ACCOUNT_USER", ")", "audit_type", "=", "SAML_AUDIT_TYPE", "user_id", "=", "(", "user_id", "or", "taxonomy", ".", "UNKNOWN", ")", "token_id", "=", "(", "token_id", "or", "taxonomy", ".", "UNKNOWN", ")", "group_ids", "=", "(", "group_ids", "or", "[", "]", ")", "cred", "=", "credential", ".", "FederatedCredential", "(", "token", "=", "token_id", ",", "type", "=", "audit_type", ",", "identity_provider", "=", "identity_provider", ",", "user", "=", "user_id", ",", "groups", "=", "group_ids", ")", "initiator", ".", "credential", "=", "cred", "event_type", "=", "(", "'%s.%s'", "%", "(", "SERVICE", ",", "action", ")", ")", "_send_audit_notification", "(", "action", ",", "initiator", ",", "outcome", ",", "target", ",", "event_type", ")" ]
send notification to inform observers about saml events .
train
false
46,878
def import_submodules(module): submodules = {} for (loader, name, ispkg) in pkgutil.iter_modules(module.__path__, (module.__name__ + '.')): try: submodule = import_module(name) except ImportError as e: logging.warning(('Error importing %s' % name)) logging.exception(e) else: (parent, child) = name.rsplit('.', 1) submodules[child] = submodule return submodules
[ "def", "import_submodules", "(", "module", ")", ":", "submodules", "=", "{", "}", "for", "(", "loader", ",", "name", ",", "ispkg", ")", "in", "pkgutil", ".", "iter_modules", "(", "module", ".", "__path__", ",", "(", "module", ".", "__name__", "+", "'.'", ")", ")", ":", "try", ":", "submodule", "=", "import_module", "(", "name", ")", "except", "ImportError", "as", "e", ":", "logging", ".", "warning", "(", "(", "'Error importing %s'", "%", "name", ")", ")", "logging", ".", "exception", "(", "e", ")", "else", ":", "(", "parent", ",", "child", ")", "=", "name", ".", "rsplit", "(", "'.'", ",", "1", ")", "submodules", "[", "child", "]", "=", "submodule", "return", "submodules" ]
import all submodules and make them available in a dict .
train
true
46,880
def test_step_clipping_no_threshold_regression(): rule1 = StepClipping() assert (rule1.threshold is None) gradients = {0: shared_floatx(3.0), 1: shared_floatx(4.0)} (clipped1, updates) = rule1.compute_steps(gradients) assert (len(updates) == 0) assert (clipped1 == gradients)
[ "def", "test_step_clipping_no_threshold_regression", "(", ")", ":", "rule1", "=", "StepClipping", "(", ")", "assert", "(", "rule1", ".", "threshold", "is", "None", ")", "gradients", "=", "{", "0", ":", "shared_floatx", "(", "3.0", ")", ",", "1", ":", "shared_floatx", "(", "4.0", ")", "}", "(", "clipped1", ",", "updates", ")", "=", "rule1", ".", "compute_steps", "(", "gradients", ")", "assert", "(", "len", "(", "updates", ")", "==", "0", ")", "assert", "(", "clipped1", "==", "gradients", ")" ]
test regression for #1145 .
train
false
46,881
def distribute(A, B): def distribute_rl(expr): for (i, arg) in enumerate(expr.args): if isinstance(arg, B): (first, b, tail) = (expr.args[:i], expr.args[i], expr.args[(i + 1):]) return B(*[A(*((first + (arg,)) + tail)) for arg in b.args]) return expr return distribute_rl
[ "def", "distribute", "(", "A", ",", "B", ")", ":", "def", "distribute_rl", "(", "expr", ")", ":", "for", "(", "i", ",", "arg", ")", "in", "enumerate", "(", "expr", ".", "args", ")", ":", "if", "isinstance", "(", "arg", ",", "B", ")", ":", "(", "first", ",", "b", ",", "tail", ")", "=", "(", "expr", ".", "args", "[", ":", "i", "]", ",", "expr", ".", "args", "[", "i", "]", ",", "expr", ".", "args", "[", "(", "i", "+", "1", ")", ":", "]", ")", "return", "B", "(", "*", "[", "A", "(", "*", "(", "(", "first", "+", "(", "arg", ",", ")", ")", "+", "tail", ")", ")", "for", "arg", "in", "b", ".", "args", "]", ")", "return", "expr", "return", "distribute_rl" ]
turns an a containing bs into a b of as where a .
train
false
46,884
def _normalize_image_location_for_db(image_data): if (('locations' not in image_data) and ('location_data' not in image_data)): image_data['locations'] = None return image_data locations = image_data.pop('locations', []) location_data = image_data.pop('location_data', []) location_data_dict = {} for l in locations: location_data_dict[l] = {} for l in location_data: location_data_dict[l['url']] = {'metadata': l['metadata'], 'status': l['status'], 'id': (l['id'] if ('id' in l) else None)} ordered_keys = locations[:] for ld in location_data: if (ld['url'] not in ordered_keys): ordered_keys.append(ld['url']) location_data = [] for loc in ordered_keys: data = location_data_dict[loc] if data: location_data.append({'url': loc, 'metadata': data['metadata'], 'status': data['status'], 'id': data['id']}) else: location_data.append({'url': loc, 'metadata': {}, 'status': 'active', 'id': None}) image_data['locations'] = location_data return image_data
[ "def", "_normalize_image_location_for_db", "(", "image_data", ")", ":", "if", "(", "(", "'locations'", "not", "in", "image_data", ")", "and", "(", "'location_data'", "not", "in", "image_data", ")", ")", ":", "image_data", "[", "'locations'", "]", "=", "None", "return", "image_data", "locations", "=", "image_data", ".", "pop", "(", "'locations'", ",", "[", "]", ")", "location_data", "=", "image_data", ".", "pop", "(", "'location_data'", ",", "[", "]", ")", "location_data_dict", "=", "{", "}", "for", "l", "in", "locations", ":", "location_data_dict", "[", "l", "]", "=", "{", "}", "for", "l", "in", "location_data", ":", "location_data_dict", "[", "l", "[", "'url'", "]", "]", "=", "{", "'metadata'", ":", "l", "[", "'metadata'", "]", ",", "'status'", ":", "l", "[", "'status'", "]", ",", "'id'", ":", "(", "l", "[", "'id'", "]", "if", "(", "'id'", "in", "l", ")", "else", "None", ")", "}", "ordered_keys", "=", "locations", "[", ":", "]", "for", "ld", "in", "location_data", ":", "if", "(", "ld", "[", "'url'", "]", "not", "in", "ordered_keys", ")", ":", "ordered_keys", ".", "append", "(", "ld", "[", "'url'", "]", ")", "location_data", "=", "[", "]", "for", "loc", "in", "ordered_keys", ":", "data", "=", "location_data_dict", "[", "loc", "]", "if", "data", ":", "location_data", ".", "append", "(", "{", "'url'", ":", "loc", ",", "'metadata'", ":", "data", "[", "'metadata'", "]", ",", "'status'", ":", "data", "[", "'status'", "]", ",", "'id'", ":", "data", "[", "'id'", "]", "}", ")", "else", ":", "location_data", ".", "append", "(", "{", "'url'", ":", "loc", ",", "'metadata'", ":", "{", "}", ",", "'status'", ":", "'active'", ",", "'id'", ":", "None", "}", ")", "image_data", "[", "'locations'", "]", "=", "location_data", "return", "image_data" ]
this function takes the legacy locations field and the newly added location_data field from the image_data values dictionary which flows over the wire between the registry and api servers and converts it into the location_data format only which is then consumable by the image object .
train
false
46,885
def getEvaluatedString(key, xmlElement=None): if (xmlElement == None): return None if (key in xmlElement.attributeDictionary): return str(getEvaluatedValueObliviously(key, xmlElement)) return None
[ "def", "getEvaluatedString", "(", "key", ",", "xmlElement", "=", "None", ")", ":", "if", "(", "xmlElement", "==", "None", ")", ":", "return", "None", "if", "(", "key", "in", "xmlElement", ".", "attributeDictionary", ")", ":", "return", "str", "(", "getEvaluatedValueObliviously", "(", "key", ",", "xmlElement", ")", ")", "return", "None" ]
get the evaluated value as a string .
train
false
46,886
def stubout_session(stubs, cls, product_version=(5, 6, 2), product_brand='XenServer', platform_version=(1, 9, 0), **opt_args): stubs.Set(session.XenAPISession, '_create_session', (lambda s, url: cls(url, **opt_args))) stubs.Set(session.XenAPISession, '_get_product_version_and_brand', (lambda s: (product_version, product_brand))) stubs.Set(session.XenAPISession, '_get_platform_version', (lambda s: platform_version))
[ "def", "stubout_session", "(", "stubs", ",", "cls", ",", "product_version", "=", "(", "5", ",", "6", ",", "2", ")", ",", "product_brand", "=", "'XenServer'", ",", "platform_version", "=", "(", "1", ",", "9", ",", "0", ")", ",", "**", "opt_args", ")", ":", "stubs", ".", "Set", "(", "session", ".", "XenAPISession", ",", "'_create_session'", ",", "(", "lambda", "s", ",", "url", ":", "cls", "(", "url", ",", "**", "opt_args", ")", ")", ")", "stubs", ".", "Set", "(", "session", ".", "XenAPISession", ",", "'_get_product_version_and_brand'", ",", "(", "lambda", "s", ":", "(", "product_version", ",", "product_brand", ")", ")", ")", "stubs", ".", "Set", "(", "session", ".", "XenAPISession", ",", "'_get_platform_version'", ",", "(", "lambda", "s", ":", "platform_version", ")", ")" ]
stubs out methods from xenapisession .
train
false
46,887
def json_error_handler(request): errors = request.errors sorted_errors = sorted(errors, key=(lambda x: six.text_type(x['name']))) error = sorted_errors[0] name = error['name'] description = error['description'] if isinstance(description, six.binary_type): description = error['description'].decode('utf-8') if (name is not None): if (name in description): message = description else: message = ('%(name)s in %(location)s: %(description)s' % error) else: message = ('%(location)s: %(description)s' % error) response = http_error(httpexceptions.HTTPBadRequest(), code=errors.status, errno=ERRORS.INVALID_PARAMETERS.value, error='Invalid parameters', message=message, details=errors) response.status = errors.status response = reapply_cors(request, response) return response
[ "def", "json_error_handler", "(", "request", ")", ":", "errors", "=", "request", ".", "errors", "sorted_errors", "=", "sorted", "(", "errors", ",", "key", "=", "(", "lambda", "x", ":", "six", ".", "text_type", "(", "x", "[", "'name'", "]", ")", ")", ")", "error", "=", "sorted_errors", "[", "0", "]", "name", "=", "error", "[", "'name'", "]", "description", "=", "error", "[", "'description'", "]", "if", "isinstance", "(", "description", ",", "six", ".", "binary_type", ")", ":", "description", "=", "error", "[", "'description'", "]", ".", "decode", "(", "'utf-8'", ")", "if", "(", "name", "is", "not", "None", ")", ":", "if", "(", "name", "in", "description", ")", ":", "message", "=", "description", "else", ":", "message", "=", "(", "'%(name)s in %(location)s: %(description)s'", "%", "error", ")", "else", ":", "message", "=", "(", "'%(location)s: %(description)s'", "%", "error", ")", "response", "=", "http_error", "(", "httpexceptions", ".", "HTTPBadRequest", "(", ")", ",", "code", "=", "errors", ".", "status", ",", "errno", "=", "ERRORS", ".", "INVALID_PARAMETERS", ".", "value", ",", "error", "=", "'Invalid parameters'", ",", "message", "=", "message", ",", "details", "=", "errors", ")", "response", ".", "status", "=", "errors", ".", "status", "response", "=", "reapply_cors", "(", "request", ",", "response", ")", "return", "response" ]
cornice json error handler .
train
false
46,888
def remove_dark_lang_config(apps, schema_editor): raise RuntimeError(u'Cannot reverse this migration.')
[ "def", "remove_dark_lang_config", "(", "apps", ",", "schema_editor", ")", ":", "raise", "RuntimeError", "(", "u'Cannot reverse this migration.'", ")" ]
write your backwards methods here .
train
false
46,889
def clone_bs4_elem(el): if isinstance(el, NavigableString): return type(el)(el) copy = Tag(None, el.builder, el.name, el.namespace, el.nsprefix) copy.attrs = dict(el.attrs) for attr in ('can_be_empty_element', 'hidden'): setattr(copy, attr, getattr(el, attr)) for child in el.contents: copy.append(clone_bs4_elem(child)) return copy
[ "def", "clone_bs4_elem", "(", "el", ")", ":", "if", "isinstance", "(", "el", ",", "NavigableString", ")", ":", "return", "type", "(", "el", ")", "(", "el", ")", "copy", "=", "Tag", "(", "None", ",", "el", ".", "builder", ",", "el", ".", "name", ",", "el", ".", "namespace", ",", "el", ".", "nsprefix", ")", "copy", ".", "attrs", "=", "dict", "(", "el", ".", "attrs", ")", "for", "attr", "in", "(", "'can_be_empty_element'", ",", "'hidden'", ")", ":", "setattr", "(", "copy", ",", "attr", ",", "getattr", "(", "el", ",", "attr", ")", ")", "for", "child", "in", "el", ".", "contents", ":", "copy", ".", "append", "(", "clone_bs4_elem", "(", "child", ")", ")", "return", "copy" ]
clone a bs4 tag before modifying it .
train
true
46,892
def _optimize_clone(optimizer, clone, num_clones, regularization_losses, **kwargs): sum_loss = _gather_clone_loss(clone, num_clones, regularization_losses) clone_grad = None if (sum_loss is not None): with tf.device(clone.device): clone_grad = optimizer.compute_gradients(sum_loss, **kwargs) return (sum_loss, clone_grad)
[ "def", "_optimize_clone", "(", "optimizer", ",", "clone", ",", "num_clones", ",", "regularization_losses", ",", "**", "kwargs", ")", ":", "sum_loss", "=", "_gather_clone_loss", "(", "clone", ",", "num_clones", ",", "regularization_losses", ")", "clone_grad", "=", "None", "if", "(", "sum_loss", "is", "not", "None", ")", ":", "with", "tf", ".", "device", "(", "clone", ".", "device", ")", ":", "clone_grad", "=", "optimizer", ".", "compute_gradients", "(", "sum_loss", ",", "**", "kwargs", ")", "return", "(", "sum_loss", ",", "clone_grad", ")" ]
compute losses and gradients for a single clone .
train
false
46,893
def format_str_for_message(msg_str): lines = msg_str.splitlines() num_lines = len(lines) sr = '\n'.join(lines) if (num_lines == 0): return '' elif (num_lines == 1): return (' ' + sr) else: return ('\n' + sr)
[ "def", "format_str_for_message", "(", "msg_str", ")", ":", "lines", "=", "msg_str", ".", "splitlines", "(", ")", "num_lines", "=", "len", "(", "lines", ")", "sr", "=", "'\\n'", ".", "join", "(", "lines", ")", "if", "(", "num_lines", "==", "0", ")", ":", "return", "''", "elif", "(", "num_lines", "==", "1", ")", ":", "return", "(", "' '", "+", "sr", ")", "else", ":", "return", "(", "'\\n'", "+", "sr", ")" ]
format msg_str so that it can be appended to a message .
train
false
46,894
def ensure_cached(task_cls, expected_num_artifacts=None): def decorator(test_fn): def wrapper(self, *args, **kwargs): with temporary_dir() as artifact_cache: self.set_options_for_scope(u'cache.{}'.format(self.options_scope), write_to=[artifact_cache]) task_cache = os.path.join(artifact_cache, task_cls.stable_name()) os.mkdir(task_cache) test_fn(self, *args, **kwargs) num_artifacts = 0 for (_, _, files) in os.walk(task_cache): num_artifacts += len(files) if (expected_num_artifacts is None): self.assertNotEqual(num_artifacts, 0) else: self.assertEqual(num_artifacts, expected_num_artifacts) return wrapper return decorator
[ "def", "ensure_cached", "(", "task_cls", ",", "expected_num_artifacts", "=", "None", ")", ":", "def", "decorator", "(", "test_fn", ")", ":", "def", "wrapper", "(", "self", ",", "*", "args", ",", "**", "kwargs", ")", ":", "with", "temporary_dir", "(", ")", "as", "artifact_cache", ":", "self", ".", "set_options_for_scope", "(", "u'cache.{}'", ".", "format", "(", "self", ".", "options_scope", ")", ",", "write_to", "=", "[", "artifact_cache", "]", ")", "task_cache", "=", "os", ".", "path", ".", "join", "(", "artifact_cache", ",", "task_cls", ".", "stable_name", "(", ")", ")", "os", ".", "mkdir", "(", "task_cache", ")", "test_fn", "(", "self", ",", "*", "args", ",", "**", "kwargs", ")", "num_artifacts", "=", "0", "for", "(", "_", ",", "_", ",", "files", ")", "in", "os", ".", "walk", "(", "task_cache", ")", ":", "num_artifacts", "+=", "len", "(", "files", ")", "if", "(", "expected_num_artifacts", "is", "None", ")", ":", "self", ".", "assertNotEqual", "(", "num_artifacts", ",", "0", ")", "else", ":", "self", ".", "assertEqual", "(", "num_artifacts", ",", "expected_num_artifacts", ")", "return", "wrapper", "return", "decorator" ]
decorator for a task-executing unit test .
train
false
46,895
def npy_cdouble_from_double_complex(var): res = '_complexstuff.npy_cdouble_from_double_complex({})'.format(var) return res
[ "def", "npy_cdouble_from_double_complex", "(", "var", ")", ":", "res", "=", "'_complexstuff.npy_cdouble_from_double_complex({})'", ".", "format", "(", "var", ")", "return", "res" ]
cast a cython double complex to a numpy cdouble .
train
false
46,896
def deactivateAaPdpContextRequest(): a = TpPd(pd=8) b = MessageType(mesType=83) c = AaDeactivationCauseAndSpareHalfOctets() packet = ((a / b) / c) return packet
[ "def", "deactivateAaPdpContextRequest", "(", ")", ":", "a", "=", "TpPd", "(", "pd", "=", "8", ")", "b", "=", "MessageType", "(", "mesType", "=", "83", ")", "c", "=", "AaDeactivationCauseAndSpareHalfOctets", "(", ")", "packet", "=", "(", "(", "a", "/", "b", ")", "/", "c", ")", "return", "packet" ]
deactivate aa pdp context request section 9 .
train
true
46,897
def get_view_plugin(view_type): for plugin in p.PluginImplementations(p.IResourceView): info = plugin.info() name = info.get('name') if (name == view_type): return plugin
[ "def", "get_view_plugin", "(", "view_type", ")", ":", "for", "plugin", "in", "p", ".", "PluginImplementations", "(", "p", ".", "IResourceView", ")", ":", "info", "=", "plugin", ".", "info", "(", ")", "name", "=", "info", ".", "get", "(", "'name'", ")", "if", "(", "name", "==", "view_type", ")", ":", "return", "plugin" ]
returns the iresourceview plugin associated with the given view_type .
train
false
46,898
def convolution_2d(x, W, b=None, stride=1, pad=0, use_cudnn=True, cover_all=False, deterministic=False): func = Convolution2DFunction(stride, pad, use_cudnn, cover_all, deterministic) if (b is None): return func(x, W) else: return func(x, W, b)
[ "def", "convolution_2d", "(", "x", ",", "W", ",", "b", "=", "None", ",", "stride", "=", "1", ",", "pad", "=", "0", ",", "use_cudnn", "=", "True", ",", "cover_all", "=", "False", ",", "deterministic", "=", "False", ")", ":", "func", "=", "Convolution2DFunction", "(", "stride", ",", "pad", ",", "use_cudnn", ",", "cover_all", ",", "deterministic", ")", "if", "(", "b", "is", "None", ")", ":", "return", "func", "(", "x", ",", "W", ")", "else", ":", "return", "func", "(", "x", ",", "W", ",", "b", ")" ]
two-dimensional convolution function .
train
false
46,900
def proportion_effectsize(prop1, prop2, method='normal'): if (method != 'normal'): raise ValueError('only "normal" is implemented') es = (2 * (np.arcsin(np.sqrt(prop1)) - np.arcsin(np.sqrt(prop2)))) return es
[ "def", "proportion_effectsize", "(", "prop1", ",", "prop2", ",", "method", "=", "'normal'", ")", ":", "if", "(", "method", "!=", "'normal'", ")", ":", "raise", "ValueError", "(", "'only \"normal\" is implemented'", ")", "es", "=", "(", "2", "*", "(", "np", ".", "arcsin", "(", "np", ".", "sqrt", "(", "prop1", ")", ")", "-", "np", ".", "arcsin", "(", "np", ".", "sqrt", "(", "prop2", ")", ")", ")", ")", "return", "es" ]
effect size for a test comparing two proportions for use in power function parameters prop1 .
train
false
46,901
@pytest.fixture() def p(tmpdir, *args): return partial(os.path.join, tmpdir)
[ "@", "pytest", ".", "fixture", "(", ")", "def", "p", "(", "tmpdir", ",", "*", "args", ")", ":", "return", "partial", "(", "os", ".", "path", ".", "join", ",", "tmpdir", ")" ]
creates an ui property inside an :class:uielement:: @p @p @p class sectionplugin : typeid = main:section .
train
false
46,903
def last_bytes(file_like_object, num): try: file_like_object.seek((- num), os.SEEK_END) except IOError as e: if (e.errno == 22): file_like_object.seek(0, os.SEEK_SET) else: raise remaining = file_like_object.tell() return (file_like_object.read(), remaining)
[ "def", "last_bytes", "(", "file_like_object", ",", "num", ")", ":", "try", ":", "file_like_object", ".", "seek", "(", "(", "-", "num", ")", ",", "os", ".", "SEEK_END", ")", "except", "IOError", "as", "e", ":", "if", "(", "e", ".", "errno", "==", "22", ")", ":", "file_like_object", ".", "seek", "(", "0", ",", "os", ".", "SEEK_SET", ")", "else", ":", "raise", "remaining", "=", "file_like_object", ".", "tell", "(", ")", "return", "(", "file_like_object", ".", "read", "(", ")", ",", "remaining", ")" ]
return num bytes from the end of the file .
train
true
46,904
def increment_ip_cidr(ip_cidr, offset=1): net0 = netaddr.IPNetwork(ip_cidr) net = netaddr.IPNetwork(ip_cidr) net.value += offset if (not (net0.network < net.ip < net0[(-1)])): tools.fail(('Incorrect ip_cidr,offset tuple (%s,%s): "incremented" ip_cidr is outside ip_cidr' % (ip_cidr, offset))) return str(net)
[ "def", "increment_ip_cidr", "(", "ip_cidr", ",", "offset", "=", "1", ")", ":", "net0", "=", "netaddr", ".", "IPNetwork", "(", "ip_cidr", ")", "net", "=", "netaddr", ".", "IPNetwork", "(", "ip_cidr", ")", "net", ".", "value", "+=", "offset", "if", "(", "not", "(", "net0", ".", "network", "<", "net", ".", "ip", "<", "net0", "[", "(", "-", "1", ")", "]", ")", ")", ":", "tools", ".", "fail", "(", "(", "'Incorrect ip_cidr,offset tuple (%s,%s): \"incremented\" ip_cidr is outside ip_cidr'", "%", "(", "ip_cidr", ",", "offset", ")", ")", ")", "return", "str", "(", "net", ")" ]
increment ip_cidr offset times .
train
false
46,905
def get_valid_breeds(): if ('breeds' in SIGNATURE_CACHE): return SIGNATURE_CACHE['breeds'].keys() else: return []
[ "def", "get_valid_breeds", "(", ")", ":", "if", "(", "'breeds'", "in", "SIGNATURE_CACHE", ")", ":", "return", "SIGNATURE_CACHE", "[", "'breeds'", "]", ".", "keys", "(", ")", "else", ":", "return", "[", "]" ]
return a list of valid breeds found in the import signatures .
train
false
46,907
def _read_temp_ifaces(iface, data): try: template = JINJA.get_template('debian_eth.jinja') except jinja2.exceptions.TemplateNotFound: log.error('Could not load template debian_eth.jinja') return '' ifcfg = template.render({'name': iface, 'data': data}) return [(item + '\n') for item in ifcfg.split('\n')]
[ "def", "_read_temp_ifaces", "(", "iface", ",", "data", ")", ":", "try", ":", "template", "=", "JINJA", ".", "get_template", "(", "'debian_eth.jinja'", ")", "except", "jinja2", ".", "exceptions", ".", "TemplateNotFound", ":", "log", ".", "error", "(", "'Could not load template debian_eth.jinja'", ")", "return", "''", "ifcfg", "=", "template", ".", "render", "(", "{", "'name'", ":", "iface", ",", "'data'", ":", "data", "}", ")", "return", "[", "(", "item", "+", "'\\n'", ")", "for", "item", "in", "ifcfg", ".", "split", "(", "'\\n'", ")", "]" ]
return what would be written to disk for interfaces .
train
true
46,908
def _upgrade_other(t_images, t_image_members, t_image_properties, dialect): foreign_keys = _get_foreign_keys(t_images, t_image_members, t_image_properties, dialect) for fk in foreign_keys: fk.drop() t_images.c.id.alter(sqlalchemy.String(36), primary_key=True) t_image_members.c.image_id.alter(sqlalchemy.String(36)) t_image_properties.c.image_id.alter(sqlalchemy.String(36)) _update_all_ids_to_uuids(t_images, t_image_members, t_image_properties) for fk in foreign_keys: fk.create()
[ "def", "_upgrade_other", "(", "t_images", ",", "t_image_members", ",", "t_image_properties", ",", "dialect", ")", ":", "foreign_keys", "=", "_get_foreign_keys", "(", "t_images", ",", "t_image_members", ",", "t_image_properties", ",", "dialect", ")", "for", "fk", "in", "foreign_keys", ":", "fk", ".", "drop", "(", ")", "t_images", ".", "c", ".", "id", ".", "alter", "(", "sqlalchemy", ".", "String", "(", "36", ")", ",", "primary_key", "=", "True", ")", "t_image_members", ".", "c", ".", "image_id", ".", "alter", "(", "sqlalchemy", ".", "String", "(", "36", ")", ")", "t_image_properties", ".", "c", ".", "image_id", ".", "alter", "(", "sqlalchemy", ".", "String", "(", "36", ")", ")", "_update_all_ids_to_uuids", "(", "t_images", ",", "t_image_members", ",", "t_image_properties", ")", "for", "fk", "in", "foreign_keys", ":", "fk", ".", "create", "(", ")" ]
upgrade 011 -> 012 with logic for non-sqlite databases .
train
false
46,910
def get_bootstrap_modules(): mod_struct = __import__('struct') loader_mods = [] loaderpath = os.path.join(HOMEPATH, 'PyInstaller', 'loader') for mod_name in ['_struct', 'zlib']: mod = __import__(mod_name) if hasattr(mod, '__file__'): loader_mods.append((mod_name, os.path.abspath(mod.__file__), 'EXTENSION')) loader_mods += [('struct', os.path.abspath(mod_struct.__file__), 'PYMODULE'), ('pyimod01_os_path', os.path.join(loaderpath, 'pyimod01_os_path.pyc'), 'PYMODULE'), ('pyimod02_archive', os.path.join(loaderpath, 'pyimod02_archive.pyc'), 'PYMODULE'), ('pyimod03_importers', os.path.join(loaderpath, 'pyimod03_importers.pyc'), 'PYMODULE'), ('pyiboot01_bootstrap', os.path.join(loaderpath, 'pyiboot01_bootstrap.py'), 'PYSOURCE')] toc = TOC(loader_mods) return toc
[ "def", "get_bootstrap_modules", "(", ")", ":", "mod_struct", "=", "__import__", "(", "'struct'", ")", "loader_mods", "=", "[", "]", "loaderpath", "=", "os", ".", "path", ".", "join", "(", "HOMEPATH", ",", "'PyInstaller'", ",", "'loader'", ")", "for", "mod_name", "in", "[", "'_struct'", ",", "'zlib'", "]", ":", "mod", "=", "__import__", "(", "mod_name", ")", "if", "hasattr", "(", "mod", ",", "'__file__'", ")", ":", "loader_mods", ".", "append", "(", "(", "mod_name", ",", "os", ".", "path", ".", "abspath", "(", "mod", ".", "__file__", ")", ",", "'EXTENSION'", ")", ")", "loader_mods", "+=", "[", "(", "'struct'", ",", "os", ".", "path", ".", "abspath", "(", "mod_struct", ".", "__file__", ")", ",", "'PYMODULE'", ")", ",", "(", "'pyimod01_os_path'", ",", "os", ".", "path", ".", "join", "(", "loaderpath", ",", "'pyimod01_os_path.pyc'", ")", ",", "'PYMODULE'", ")", ",", "(", "'pyimod02_archive'", ",", "os", ".", "path", ".", "join", "(", "loaderpath", ",", "'pyimod02_archive.pyc'", ")", ",", "'PYMODULE'", ")", ",", "(", "'pyimod03_importers'", ",", "os", ".", "path", ".", "join", "(", "loaderpath", ",", "'pyimod03_importers.pyc'", ")", ",", "'PYMODULE'", ")", ",", "(", "'pyiboot01_bootstrap'", ",", "os", ".", "path", ".", "join", "(", "loaderpath", ",", "'pyiboot01_bootstrap.py'", ")", ",", "'PYSOURCE'", ")", "]", "toc", "=", "TOC", "(", "loader_mods", ")", "return", "toc" ]
get toc with the bootstrapping modules and their dependencies .
train
false
46,911
def push_connection(redis): _connection_stack.push(patch_connection(redis))
[ "def", "push_connection", "(", "redis", ")", ":", "_connection_stack", ".", "push", "(", "patch_connection", "(", "redis", ")", ")" ]
pushes the given connection on the stack .
train
false
46,912
def org_site_top_req_priority(row, tablename='org_facility'): if (not current.deployment_settings.has_module('req')): return None try: from req import REQ_STATUS_COMPLETE except ImportError: return None if hasattr(row, tablename): row = row[tablename] try: id = row.id except AttributeError: return None s3db = current.s3db rtable = s3db.req_req stable = s3db[tablename] query = (((((rtable.deleted != True) & (stable.id == id)) & (rtable.site_id == stable.site_id)) & (rtable.fulfil_status != REQ_STATUS_COMPLETE)) & (rtable.is_template == False)) req = current.db(query).select(rtable.id, rtable.priority, orderby=(~ rtable.priority), limitby=(0, 1)).first() if req: return req.priority else: return None
[ "def", "org_site_top_req_priority", "(", "row", ",", "tablename", "=", "'org_facility'", ")", ":", "if", "(", "not", "current", ".", "deployment_settings", ".", "has_module", "(", "'req'", ")", ")", ":", "return", "None", "try", ":", "from", "req", "import", "REQ_STATUS_COMPLETE", "except", "ImportError", ":", "return", "None", "if", "hasattr", "(", "row", ",", "tablename", ")", ":", "row", "=", "row", "[", "tablename", "]", "try", ":", "id", "=", "row", ".", "id", "except", "AttributeError", ":", "return", "None", "s3db", "=", "current", ".", "s3db", "rtable", "=", "s3db", ".", "req_req", "stable", "=", "s3db", "[", "tablename", "]", "query", "=", "(", "(", "(", "(", "(", "rtable", ".", "deleted", "!=", "True", ")", "&", "(", "stable", ".", "id", "==", "id", ")", ")", "&", "(", "rtable", ".", "site_id", "==", "stable", ".", "site_id", ")", ")", "&", "(", "rtable", ".", "fulfil_status", "!=", "REQ_STATUS_COMPLETE", ")", ")", "&", "(", "rtable", ".", "is_template", "==", "False", ")", ")", "req", "=", "current", ".", "db", "(", "query", ")", ".", "select", "(", "rtable", ".", "id", ",", "rtable", ".", "priority", ",", "orderby", "=", "(", "~", "rtable", ".", "priority", ")", ",", "limitby", "=", "(", "0", ",", "1", ")", ")", ".", "first", "(", ")", "if", "req", ":", "return", "req", ".", "priority", "else", ":", "return", "None" ]
highest priority of open requests for a site .
train
false
46,915
def usageExit(msg=None): if msg: print msg print USAGE sys.exit(2)
[ "def", "usageExit", "(", "msg", "=", "None", ")", ":", "if", "msg", ":", "print", "msg", "print", "USAGE", "sys", ".", "exit", "(", "2", ")" ]
print usage and exit .
train
false
46,916
def compile_single(source, options, full_module_name=None): return run_pipeline(source, options, full_module_name)
[ "def", "compile_single", "(", "source", ",", "options", ",", "full_module_name", "=", "None", ")", ":", "return", "run_pipeline", "(", "source", ",", "options", ",", "full_module_name", ")" ]
compile_single compile the given pyrex implementation file and return a compilationresult .
train
false
46,917
def formatFileSize(size): if (size < 1024): return ('%iB' % size) elif (size < (1024 ** 2)): return ('%iK' % (size / 1024)) elif (size < (1024 ** 3)): return ('%iM' % (size / (1024 ** 2))) else: return ('%iG' % (size / (1024 ** 3)))
[ "def", "formatFileSize", "(", "size", ")", ":", "if", "(", "size", "<", "1024", ")", ":", "return", "(", "'%iB'", "%", "size", ")", "elif", "(", "size", "<", "(", "1024", "**", "2", ")", ")", ":", "return", "(", "'%iK'", "%", "(", "size", "/", "1024", ")", ")", "elif", "(", "size", "<", "(", "1024", "**", "3", ")", ")", ":", "return", "(", "'%iM'", "%", "(", "size", "/", "(", "1024", "**", "2", ")", ")", ")", "else", ":", "return", "(", "'%iG'", "%", "(", "size", "/", "(", "1024", "**", "3", ")", ")", ")" ]
format the given file size in bytes to human readable format .
train
false
46,918
def nowtime(): return now_datetime().strftime(TIME_FORMAT)
[ "def", "nowtime", "(", ")", ":", "return", "now_datetime", "(", ")", ".", "strftime", "(", "TIME_FORMAT", ")" ]
return current time in hh:mm .
train
false
46,919
def run_git(args, git_path=_DEFAULT_GIT, input=None, capture_stdout=False, **popen_kwargs): env = popen_kwargs.pop('env', {}) env['LC_ALL'] = env['LANG'] = 'C' args = ([git_path] + args) popen_kwargs['stdin'] = subprocess.PIPE if capture_stdout: popen_kwargs['stdout'] = subprocess.PIPE else: popen_kwargs.pop('stdout', None) p = subprocess.Popen(args, env=env, **popen_kwargs) (stdout, stderr) = p.communicate(input=input) return (p.returncode, stdout)
[ "def", "run_git", "(", "args", ",", "git_path", "=", "_DEFAULT_GIT", ",", "input", "=", "None", ",", "capture_stdout", "=", "False", ",", "**", "popen_kwargs", ")", ":", "env", "=", "popen_kwargs", ".", "pop", "(", "'env'", ",", "{", "}", ")", "env", "[", "'LC_ALL'", "]", "=", "env", "[", "'LANG'", "]", "=", "'C'", "args", "=", "(", "[", "git_path", "]", "+", "args", ")", "popen_kwargs", "[", "'stdin'", "]", "=", "subprocess", ".", "PIPE", "if", "capture_stdout", ":", "popen_kwargs", "[", "'stdout'", "]", "=", "subprocess", ".", "PIPE", "else", ":", "popen_kwargs", ".", "pop", "(", "'stdout'", ",", "None", ")", "p", "=", "subprocess", ".", "Popen", "(", "args", ",", "env", "=", "env", ",", "**", "popen_kwargs", ")", "(", "stdout", ",", "stderr", ")", "=", "p", ".", "communicate", "(", "input", "=", "input", ")", "return", "(", "p", ".", "returncode", ",", "stdout", ")" ]
run a git command .
train
false
46,920
def test_add_contacts_case_insensitive(contacts_provider, contact_sync, db, default_namespace): num_original_contacts = db.session.query(Contact).filter_by(namespace_id=default_namespace.id).count() contacts_provider._next_uid = 'foo' contacts_provider._get_next_uid = (lambda current: 'FOO') contacts_provider.supply_contact('Contact One', 'contact.one@email.address') contacts_provider.supply_contact('Contact Two', 'contact.two@email.address') contact_sync.provider = contacts_provider contact_sync.sync() num_current_contacts = db.session.query(Contact).filter_by(namespace_id=default_namespace.id).count() assert ((num_current_contacts - num_original_contacts) == 2)
[ "def", "test_add_contacts_case_insensitive", "(", "contacts_provider", ",", "contact_sync", ",", "db", ",", "default_namespace", ")", ":", "num_original_contacts", "=", "db", ".", "session", ".", "query", "(", "Contact", ")", ".", "filter_by", "(", "namespace_id", "=", "default_namespace", ".", "id", ")", ".", "count", "(", ")", "contacts_provider", ".", "_next_uid", "=", "'foo'", "contacts_provider", ".", "_get_next_uid", "=", "(", "lambda", "current", ":", "'FOO'", ")", "contacts_provider", ".", "supply_contact", "(", "'Contact One'", ",", "'contact.one@email.address'", ")", "contacts_provider", ".", "supply_contact", "(", "'Contact Two'", ",", "'contact.two@email.address'", ")", "contact_sync", ".", "provider", "=", "contacts_provider", "contact_sync", ".", "sync", "(", ")", "num_current_contacts", "=", "db", ".", "session", ".", "query", "(", "Contact", ")", ".", "filter_by", "(", "namespace_id", "=", "default_namespace", ".", "id", ")", ".", "count", "(", ")", "assert", "(", "(", "num_current_contacts", "-", "num_original_contacts", ")", "==", "2", ")" ]
tests that syncing two contacts with uids that differ only in case sensitivity doesnt cause an error .
train
false
46,921
def is_rich_text_default_for_user(user): if user.is_authenticated(): try: return user.get_profile().should_use_rich_text except ObjectDoesNotExist: pass siteconfig = SiteConfiguration.objects.get_current() return siteconfig.get(u'default_use_rich_text')
[ "def", "is_rich_text_default_for_user", "(", "user", ")", ":", "if", "user", ".", "is_authenticated", "(", ")", ":", "try", ":", "return", "user", ".", "get_profile", "(", ")", ".", "should_use_rich_text", "except", "ObjectDoesNotExist", ":", "pass", "siteconfig", "=", "SiteConfiguration", ".", "objects", ".", "get_current", "(", ")", "return", "siteconfig", ".", "get", "(", "u'default_use_rich_text'", ")" ]
returns whether the user edits in markdown by default .
train
false
46,922
def create_terminal_writer(config, *args, **kwargs): tw = py.io.TerminalWriter(*args, **kwargs) if (config.option.color == 'yes'): tw.hasmarkup = True if (config.option.color == 'no'): tw.hasmarkup = False return tw
[ "def", "create_terminal_writer", "(", "config", ",", "*", "args", ",", "**", "kwargs", ")", ":", "tw", "=", "py", ".", "io", ".", "TerminalWriter", "(", "*", "args", ",", "**", "kwargs", ")", "if", "(", "config", ".", "option", ".", "color", "==", "'yes'", ")", ":", "tw", ".", "hasmarkup", "=", "True", "if", "(", "config", ".", "option", ".", "color", "==", "'no'", ")", ":", "tw", ".", "hasmarkup", "=", "False", "return", "tw" ]
create a terminalwriter instance configured according to the options in the config object .
train
false
46,923
def parallel_bulk(client, actions, thread_count=4, chunk_size=500, max_chunk_bytes=((100 * 1024) * 1024), expand_action_callback=expand_action, **kwargs): from multiprocessing.dummy import Pool actions = map(expand_action_callback, actions) pool = Pool(thread_count) try: for result in pool.imap((lambda chunk: list(_process_bulk_chunk(client, chunk, **kwargs))), _chunk_actions(actions, chunk_size, max_chunk_bytes, client.transport.serializer)): for item in result: (yield item) finally: pool.close() pool.join()
[ "def", "parallel_bulk", "(", "client", ",", "actions", ",", "thread_count", "=", "4", ",", "chunk_size", "=", "500", ",", "max_chunk_bytes", "=", "(", "(", "100", "*", "1024", ")", "*", "1024", ")", ",", "expand_action_callback", "=", "expand_action", ",", "**", "kwargs", ")", ":", "from", "multiprocessing", ".", "dummy", "import", "Pool", "actions", "=", "map", "(", "expand_action_callback", ",", "actions", ")", "pool", "=", "Pool", "(", "thread_count", ")", "try", ":", "for", "result", "in", "pool", ".", "imap", "(", "(", "lambda", "chunk", ":", "list", "(", "_process_bulk_chunk", "(", "client", ",", "chunk", ",", "**", "kwargs", ")", ")", ")", ",", "_chunk_actions", "(", "actions", ",", "chunk_size", ",", "max_chunk_bytes", ",", "client", ".", "transport", ".", "serializer", ")", ")", ":", "for", "item", "in", "result", ":", "(", "yield", "item", ")", "finally", ":", "pool", ".", "close", "(", ")", "pool", ".", "join", "(", ")" ]
parallel version of the bulk helper run in multiple threads at once .
train
true
46,924
def get_domains(name): name = name.split('.') return ('.'.join(name[i:]) for i in xrange(len(name)))
[ "def", "get_domains", "(", "name", ")", ":", "name", "=", "name", ".", "split", "(", "'.'", ")", "return", "(", "'.'", ".", "join", "(", "name", "[", "i", ":", "]", ")", "for", "i", "in", "xrange", "(", "len", "(", "name", ")", ")", ")" ]
generates the upper domains from a domain name .
train
false
46,925
@register.simple_tag def project_name(prj): return escape(force_text(Project.objects.get(slug=prj)))
[ "@", "register", ".", "simple_tag", "def", "project_name", "(", "prj", ")", ":", "return", "escape", "(", "force_text", "(", "Project", ".", "objects", ".", "get", "(", "slug", "=", "prj", ")", ")", ")" ]
gets project name based on slug .
train
false
46,926
def parse_message(message): events = [] if (message['action'] in ['create', 'delete']): events.append(parse_create_or_delete(message)) elif (message['action'] == 'change'): if message['change']['diff']: for value in message['change']['diff']: parsed_event = parse_change_event(value, message) if parsed_event: events.append(parsed_event) if message['change']['comment']: events.append(parse_comment(message)) return events
[ "def", "parse_message", "(", "message", ")", ":", "events", "=", "[", "]", "if", "(", "message", "[", "'action'", "]", "in", "[", "'create'", ",", "'delete'", "]", ")", ":", "events", ".", "append", "(", "parse_create_or_delete", "(", "message", ")", ")", "elif", "(", "message", "[", "'action'", "]", "==", "'change'", ")", ":", "if", "message", "[", "'change'", "]", "[", "'diff'", "]", ":", "for", "value", "in", "message", "[", "'change'", "]", "[", "'diff'", "]", ":", "parsed_event", "=", "parse_change_event", "(", "value", ",", "message", ")", "if", "parsed_event", ":", "events", ".", "append", "(", "parsed_event", ")", "if", "message", "[", "'change'", "]", "[", "'comment'", "]", ":", "events", ".", "append", "(", "parse_comment", "(", "message", ")", ")", "return", "events" ]
parses the payload by delegating to specialized functions .
train
false
46,930
def gc_histogram(): result = {} for o in gc.get_objects(): t = type(o) count = result.get(t, 0) result[t] = (count + 1) return result
[ "def", "gc_histogram", "(", ")", ":", "result", "=", "{", "}", "for", "o", "in", "gc", ".", "get_objects", "(", ")", ":", "t", "=", "type", "(", "o", ")", "count", "=", "result", ".", "get", "(", "t", ",", "0", ")", "result", "[", "t", "]", "=", "(", "count", "+", "1", ")", "return", "result" ]
returns per-class counts of existing objects .
train
false
46,932
def security_hash(request, form, *args): import warnings warnings.warn('security_hash is deprecated; use form_hmac instead', PendingDeprecationWarning) data = [] for bf in form: if (form.empty_permitted and (not form.has_changed())): value = (bf.data or '') else: value = (bf.field.clean(bf.data) or '') if isinstance(value, basestring): value = value.strip() data.append((bf.name, value)) data.extend(args) data.append(settings.SECRET_KEY) pickled = pickle.dumps(data, pickle.HIGHEST_PROTOCOL) return md5_constructor(pickled).hexdigest()
[ "def", "security_hash", "(", "request", ",", "form", ",", "*", "args", ")", ":", "import", "warnings", "warnings", ".", "warn", "(", "'security_hash is deprecated; use form_hmac instead'", ",", "PendingDeprecationWarning", ")", "data", "=", "[", "]", "for", "bf", "in", "form", ":", "if", "(", "form", ".", "empty_permitted", "and", "(", "not", "form", ".", "has_changed", "(", ")", ")", ")", ":", "value", "=", "(", "bf", ".", "data", "or", "''", ")", "else", ":", "value", "=", "(", "bf", ".", "field", ".", "clean", "(", "bf", ".", "data", ")", "or", "''", ")", "if", "isinstance", "(", "value", ",", "basestring", ")", ":", "value", "=", "value", ".", "strip", "(", ")", "data", ".", "append", "(", "(", "bf", ".", "name", ",", "value", ")", ")", "data", ".", "extend", "(", "args", ")", "data", ".", "append", "(", "settings", ".", "SECRET_KEY", ")", "pickled", "=", "pickle", ".", "dumps", "(", "data", ",", "pickle", ".", "HIGHEST_PROTOCOL", ")", "return", "md5_constructor", "(", "pickled", ")", ".", "hexdigest", "(", ")" ]
calculates a security hash for the given form instance .
train
false
46,933
def _createTPs(numCols, cellsPerColumn=4, checkSynapseConsistency=True): minThreshold = 4 activationThreshold = 4 newSynapseCount = 5 initialPerm = 0.6 connectedPerm = 0.5 permanenceInc = 0.1 permanenceDec = 0.001 globalDecay = 0.0 if (VERBOSITY > 1): print 'Creating TP10X instance' cppTp = TP10X2(numberOfCols=numCols, cellsPerColumn=cellsPerColumn, initialPerm=initialPerm, connectedPerm=connectedPerm, minThreshold=minThreshold, newSynapseCount=newSynapseCount, permanenceInc=permanenceInc, permanenceDec=permanenceDec, activationThreshold=activationThreshold, globalDecay=globalDecay, burnIn=1, seed=SEED, verbosity=VERBOSITY, checkSynapseConsistency=checkSynapseConsistency, pamLength=1000) if (VERBOSITY > 1): print 'Creating PY TP instance' pyTp = TP(numberOfCols=numCols, cellsPerColumn=cellsPerColumn, initialPerm=initialPerm, connectedPerm=connectedPerm, minThreshold=minThreshold, newSynapseCount=newSynapseCount, permanenceInc=permanenceInc, permanenceDec=permanenceDec, activationThreshold=activationThreshold, globalDecay=globalDecay, burnIn=1, seed=SEED, verbosity=VERBOSITY, pamLength=1000) return (cppTp, pyTp)
[ "def", "_createTPs", "(", "numCols", ",", "cellsPerColumn", "=", "4", ",", "checkSynapseConsistency", "=", "True", ")", ":", "minThreshold", "=", "4", "activationThreshold", "=", "4", "newSynapseCount", "=", "5", "initialPerm", "=", "0.6", "connectedPerm", "=", "0.5", "permanenceInc", "=", "0.1", "permanenceDec", "=", "0.001", "globalDecay", "=", "0.0", "if", "(", "VERBOSITY", ">", "1", ")", ":", "print", "'Creating TP10X instance'", "cppTp", "=", "TP10X2", "(", "numberOfCols", "=", "numCols", ",", "cellsPerColumn", "=", "cellsPerColumn", ",", "initialPerm", "=", "initialPerm", ",", "connectedPerm", "=", "connectedPerm", ",", "minThreshold", "=", "minThreshold", ",", "newSynapseCount", "=", "newSynapseCount", ",", "permanenceInc", "=", "permanenceInc", ",", "permanenceDec", "=", "permanenceDec", ",", "activationThreshold", "=", "activationThreshold", ",", "globalDecay", "=", "globalDecay", ",", "burnIn", "=", "1", ",", "seed", "=", "SEED", ",", "verbosity", "=", "VERBOSITY", ",", "checkSynapseConsistency", "=", "checkSynapseConsistency", ",", "pamLength", "=", "1000", ")", "if", "(", "VERBOSITY", ">", "1", ")", ":", "print", "'Creating PY TP instance'", "pyTp", "=", "TP", "(", "numberOfCols", "=", "numCols", ",", "cellsPerColumn", "=", "cellsPerColumn", ",", "initialPerm", "=", "initialPerm", ",", "connectedPerm", "=", "connectedPerm", ",", "minThreshold", "=", "minThreshold", ",", "newSynapseCount", "=", "newSynapseCount", ",", "permanenceInc", "=", "permanenceInc", ",", "permanenceDec", "=", "permanenceDec", ",", "activationThreshold", "=", "activationThreshold", ",", "globalDecay", "=", "globalDecay", ",", "burnIn", "=", "1", ",", "seed", "=", "SEED", ",", "verbosity", "=", "VERBOSITY", ",", "pamLength", "=", "1000", ")", "return", "(", "cppTp", ",", "pyTp", ")" ]
create tp and tp10x instances with identical parameters .
train
false
46,934
def parse_date_or_fatal(str, fatal): try: date = float(str) except ValueError as e: raise fatal(('invalid date format (should be a float): %r' % e)) else: return date
[ "def", "parse_date_or_fatal", "(", "str", ",", "fatal", ")", ":", "try", ":", "date", "=", "float", "(", "str", ")", "except", "ValueError", "as", "e", ":", "raise", "fatal", "(", "(", "'invalid date format (should be a float): %r'", "%", "e", ")", ")", "else", ":", "return", "date" ]
parses the given date or calls option .
train
false