id_within_dataset int64 1 55.5k | snippet stringlengths 19 14.2k | tokens listlengths 6 1.63k | nl stringlengths 6 352 | split_within_dataset stringclasses 1 value | is_duplicated bool 2 classes |
|---|---|---|---|---|---|
36,363 | @allow_unvouched
@never_cache
def confirm_delete(request):
return render(request, 'phonebook/confirm_delete.html')
| [
"@",
"allow_unvouched",
"@",
"never_cache",
"def",
"confirm_delete",
"(",
"request",
")",
":",
"return",
"render",
"(",
"request",
",",
"'phonebook/confirm_delete.html'",
")"
] | display a confirmation page asking the user if they want to leave . | train | false |
36,364 | def test_unnamed():
t = usertypes.Timer()
assert (not t.objectName())
assert (t._name == 'unnamed')
assert (repr(t) == "<qutebrowser.utils.usertypes.Timer name='unnamed'>")
| [
"def",
"test_unnamed",
"(",
")",
":",
"t",
"=",
"usertypes",
".",
"Timer",
"(",
")",
"assert",
"(",
"not",
"t",
".",
"objectName",
"(",
")",
")",
"assert",
"(",
"t",
".",
"_name",
"==",
"'unnamed'",
")",
"assert",
"(",
"repr",
"(",
"t",
")",
"==",
"\"<qutebrowser.utils.usertypes.Timer name='unnamed'>\"",
")"
] | make sure an unnamed timer is named correctly . | train | false |
36,365 | def getSegmentFromPath(path, pathIndex):
if (len(path) < 2):
return None
begin = path[(-1)]
end = path[(-2)]
forwardEndpoint = getEndpointFromPath(path, pathIndex)
reversePath = path[:]
reversePath.reverse()
reverseEndpoint = getEndpointFromPath(reversePath, pathIndex)
return (forwardEndpoint, reverseEndpoint)
| [
"def",
"getSegmentFromPath",
"(",
"path",
",",
"pathIndex",
")",
":",
"if",
"(",
"len",
"(",
"path",
")",
"<",
"2",
")",
":",
"return",
"None",
"begin",
"=",
"path",
"[",
"(",
"-",
"1",
")",
"]",
"end",
"=",
"path",
"[",
"(",
"-",
"2",
")",
"]",
"forwardEndpoint",
"=",
"getEndpointFromPath",
"(",
"path",
",",
"pathIndex",
")",
"reversePath",
"=",
"path",
"[",
":",
"]",
"reversePath",
".",
"reverse",
"(",
")",
"reverseEndpoint",
"=",
"getEndpointFromPath",
"(",
"reversePath",
",",
"pathIndex",
")",
"return",
"(",
"forwardEndpoint",
",",
"reverseEndpoint",
")"
] | get endpoint segment from a path . | train | false |
36,369 | def headers_to_object_info(headers, status_int=HTTP_OK):
(headers, meta, sysmeta) = _prep_headers_to_info(headers, 'object')
transient_sysmeta = {}
for (key, val) in headers.iteritems():
if is_object_transient_sysmeta(key):
key = strip_object_transient_sysmeta_prefix(key.lower())
transient_sysmeta[key] = val
info = {'status': status_int, 'length': headers.get('content-length'), 'type': headers.get('content-type'), 'etag': headers.get('etag'), 'meta': meta, 'sysmeta': sysmeta, 'transient_sysmeta': transient_sysmeta}
return info
| [
"def",
"headers_to_object_info",
"(",
"headers",
",",
"status_int",
"=",
"HTTP_OK",
")",
":",
"(",
"headers",
",",
"meta",
",",
"sysmeta",
")",
"=",
"_prep_headers_to_info",
"(",
"headers",
",",
"'object'",
")",
"transient_sysmeta",
"=",
"{",
"}",
"for",
"(",
"key",
",",
"val",
")",
"in",
"headers",
".",
"iteritems",
"(",
")",
":",
"if",
"is_object_transient_sysmeta",
"(",
"key",
")",
":",
"key",
"=",
"strip_object_transient_sysmeta_prefix",
"(",
"key",
".",
"lower",
"(",
")",
")",
"transient_sysmeta",
"[",
"key",
"]",
"=",
"val",
"info",
"=",
"{",
"'status'",
":",
"status_int",
",",
"'length'",
":",
"headers",
".",
"get",
"(",
"'content-length'",
")",
",",
"'type'",
":",
"headers",
".",
"get",
"(",
"'content-type'",
")",
",",
"'etag'",
":",
"headers",
".",
"get",
"(",
"'etag'",
")",
",",
"'meta'",
":",
"meta",
",",
"'sysmeta'",
":",
"sysmeta",
",",
"'transient_sysmeta'",
":",
"transient_sysmeta",
"}",
"return",
"info"
] | construct a cacheable dict of object info based on response headers . | train | false |
36,370 | def getAlongWayHexadecimalColor(beginBrightness, colorWidth, difference, endColorTuple, wayLength):
alongWay = 1.0
if (wayLength != 0.0):
alongWay = (0.4 + (0.6 * min(1.0, abs((float(difference) / float(wayLength))))))
hexadecimalColor = '#'
oneMinusAlongWay = (1.0 - alongWay)
for primaryIndex in xrange(3):
hexadecimalColor += getAlongWayHexadecimalPrimary(beginBrightness, oneMinusAlongWay, colorWidth, endColorTuple[primaryIndex], alongWay)
return hexadecimalColor
| [
"def",
"getAlongWayHexadecimalColor",
"(",
"beginBrightness",
",",
"colorWidth",
",",
"difference",
",",
"endColorTuple",
",",
"wayLength",
")",
":",
"alongWay",
"=",
"1.0",
"if",
"(",
"wayLength",
"!=",
"0.0",
")",
":",
"alongWay",
"=",
"(",
"0.4",
"+",
"(",
"0.6",
"*",
"min",
"(",
"1.0",
",",
"abs",
"(",
"(",
"float",
"(",
"difference",
")",
"/",
"float",
"(",
"wayLength",
")",
")",
")",
")",
")",
")",
"hexadecimalColor",
"=",
"'#'",
"oneMinusAlongWay",
"=",
"(",
"1.0",
"-",
"alongWay",
")",
"for",
"primaryIndex",
"in",
"xrange",
"(",
"3",
")",
":",
"hexadecimalColor",
"+=",
"getAlongWayHexadecimalPrimary",
"(",
"beginBrightness",
",",
"oneMinusAlongWay",
",",
"colorWidth",
",",
"endColorTuple",
"[",
"primaryIndex",
"]",
",",
"alongWay",
")",
"return",
"hexadecimalColor"
] | get a color along the way from begin brightness to the end color . | train | false |
36,371 | def func_to_time(x):
y = ((5.0 * np.asarray(x)) - 10)
time.sleep(y)
return y
| [
"def",
"func_to_time",
"(",
"x",
")",
":",
"y",
"=",
"(",
"(",
"5.0",
"*",
"np",
".",
"asarray",
"(",
"x",
")",
")",
"-",
"10",
")",
"time",
".",
"sleep",
"(",
"y",
")",
"return",
"y"
] | this sleeps for y seconds for use with timing tests . | train | false |
36,372 | def rule_create(request, **kwargs):
body = {'firewall_rule': kwargs}
rule = neutronclient(request).create_firewall_rule(body).get('firewall_rule')
return Rule(rule)
| [
"def",
"rule_create",
"(",
"request",
",",
"**",
"kwargs",
")",
":",
"body",
"=",
"{",
"'firewall_rule'",
":",
"kwargs",
"}",
"rule",
"=",
"neutronclient",
"(",
"request",
")",
".",
"create_firewall_rule",
"(",
"body",
")",
".",
"get",
"(",
"'firewall_rule'",
")",
"return",
"Rule",
"(",
"rule",
")"
] | create a firewall rule . | train | false |
36,373 | def _get_instances_with_cached_ips(orig_func, *args, **kwargs):
instances = orig_func(*args, **kwargs)
context = args[0]
fake_device = objects.PciDevice.get_by_dev_addr(context, 1, 'a')
def _info_cache_for(instance):
info_cache = dict(test_instance_info_cache.fake_info_cache, network_info=_get_fake_cache(), instance_uuid=instance['uuid'])
if isinstance(instance, obj_base.NovaObject):
_info_cache = objects.InstanceInfoCache(context)
objects.InstanceInfoCache._from_db_object(context, _info_cache, info_cache)
info_cache = _info_cache
instance['info_cache'] = info_cache
if isinstance(instances, (list, obj_base.ObjectListBase)):
for instance in instances:
_info_cache_for(instance)
fake_device.claim(instance.uuid)
fake_device.allocate(instance)
else:
_info_cache_for(instances)
fake_device.claim(instances.uuid)
fake_device.allocate(instances)
return instances
| [
"def",
"_get_instances_with_cached_ips",
"(",
"orig_func",
",",
"*",
"args",
",",
"**",
"kwargs",
")",
":",
"instances",
"=",
"orig_func",
"(",
"*",
"args",
",",
"**",
"kwargs",
")",
"context",
"=",
"args",
"[",
"0",
"]",
"fake_device",
"=",
"objects",
".",
"PciDevice",
".",
"get_by_dev_addr",
"(",
"context",
",",
"1",
",",
"'a'",
")",
"def",
"_info_cache_for",
"(",
"instance",
")",
":",
"info_cache",
"=",
"dict",
"(",
"test_instance_info_cache",
".",
"fake_info_cache",
",",
"network_info",
"=",
"_get_fake_cache",
"(",
")",
",",
"instance_uuid",
"=",
"instance",
"[",
"'uuid'",
"]",
")",
"if",
"isinstance",
"(",
"instance",
",",
"obj_base",
".",
"NovaObject",
")",
":",
"_info_cache",
"=",
"objects",
".",
"InstanceInfoCache",
"(",
"context",
")",
"objects",
".",
"InstanceInfoCache",
".",
"_from_db_object",
"(",
"context",
",",
"_info_cache",
",",
"info_cache",
")",
"info_cache",
"=",
"_info_cache",
"instance",
"[",
"'info_cache'",
"]",
"=",
"info_cache",
"if",
"isinstance",
"(",
"instances",
",",
"(",
"list",
",",
"obj_base",
".",
"ObjectListBase",
")",
")",
":",
"for",
"instance",
"in",
"instances",
":",
"_info_cache_for",
"(",
"instance",
")",
"fake_device",
".",
"claim",
"(",
"instance",
".",
"uuid",
")",
"fake_device",
".",
"allocate",
"(",
"instance",
")",
"else",
":",
"_info_cache_for",
"(",
"instances",
")",
"fake_device",
".",
"claim",
"(",
"instances",
".",
"uuid",
")",
"fake_device",
".",
"allocate",
"(",
"instances",
")",
"return",
"instances"
] | kludge the cache into instance(s) without having to create db entries . | train | false |
36,374 | def find_spec(name, package=None):
fullname = (resolve_name(name, package) if name.startswith('.') else name)
if (fullname not in sys.modules):
parent_name = fullname.rpartition('.')[0]
if parent_name:
parent = __import__(parent_name, fromlist=['__path__'])
return _find_spec(fullname, parent.__path__)
else:
return _find_spec(fullname, None)
else:
module = sys.modules[fullname]
if (module is None):
return None
try:
spec = module.__spec__
except AttributeError:
raise ValueError('{}.__spec__ is not set'.format(name))
else:
if (spec is None):
raise ValueError('{}.__spec__ is None'.format(name))
return spec
| [
"def",
"find_spec",
"(",
"name",
",",
"package",
"=",
"None",
")",
":",
"fullname",
"=",
"(",
"resolve_name",
"(",
"name",
",",
"package",
")",
"if",
"name",
".",
"startswith",
"(",
"'.'",
")",
"else",
"name",
")",
"if",
"(",
"fullname",
"not",
"in",
"sys",
".",
"modules",
")",
":",
"parent_name",
"=",
"fullname",
".",
"rpartition",
"(",
"'.'",
")",
"[",
"0",
"]",
"if",
"parent_name",
":",
"parent",
"=",
"__import__",
"(",
"parent_name",
",",
"fromlist",
"=",
"[",
"'__path__'",
"]",
")",
"return",
"_find_spec",
"(",
"fullname",
",",
"parent",
".",
"__path__",
")",
"else",
":",
"return",
"_find_spec",
"(",
"fullname",
",",
"None",
")",
"else",
":",
"module",
"=",
"sys",
".",
"modules",
"[",
"fullname",
"]",
"if",
"(",
"module",
"is",
"None",
")",
":",
"return",
"None",
"try",
":",
"spec",
"=",
"module",
".",
"__spec__",
"except",
"AttributeError",
":",
"raise",
"ValueError",
"(",
"'{}.__spec__ is not set'",
".",
"format",
"(",
"name",
")",
")",
"else",
":",
"if",
"(",
"spec",
"is",
"None",
")",
":",
"raise",
"ValueError",
"(",
"'{}.__spec__ is None'",
".",
"format",
"(",
"name",
")",
")",
"return",
"spec"
] | return the spec for the specified module . | train | true |
36,375 | def libvlc_audio_equalizer_new_from_preset(u_index):
f = (_Cfunctions.get('libvlc_audio_equalizer_new_from_preset', None) or _Cfunction('libvlc_audio_equalizer_new_from_preset', ((1,),), None, ctypes.c_void_p, ctypes.c_uint))
return f(u_index)
| [
"def",
"libvlc_audio_equalizer_new_from_preset",
"(",
"u_index",
")",
":",
"f",
"=",
"(",
"_Cfunctions",
".",
"get",
"(",
"'libvlc_audio_equalizer_new_from_preset'",
",",
"None",
")",
"or",
"_Cfunction",
"(",
"'libvlc_audio_equalizer_new_from_preset'",
",",
"(",
"(",
"1",
",",
")",
",",
")",
",",
"None",
",",
"ctypes",
".",
"c_void_p",
",",
"ctypes",
".",
"c_uint",
")",
")",
"return",
"f",
"(",
"u_index",
")"
] | create a new equalizer . | train | true |
36,376 | @register.assignment_tag
def shipping_charge_discount(method, basket):
return method.discount(basket)
| [
"@",
"register",
".",
"assignment_tag",
"def",
"shipping_charge_discount",
"(",
"method",
",",
"basket",
")",
":",
"return",
"method",
".",
"discount",
"(",
"basket",
")"
] | template tag for calculating the shipping discount for a given shipping method and basket . | train | false |
36,378 | def new_endpoint_ref_with_region(service_id, region, interface='public', **kwargs):
ref = new_endpoint_ref(service_id, interface, region=region, region_id='invalid', **kwargs)
del ref['region_id']
return ref
| [
"def",
"new_endpoint_ref_with_region",
"(",
"service_id",
",",
"region",
",",
"interface",
"=",
"'public'",
",",
"**",
"kwargs",
")",
":",
"ref",
"=",
"new_endpoint_ref",
"(",
"service_id",
",",
"interface",
",",
"region",
"=",
"region",
",",
"region_id",
"=",
"'invalid'",
",",
"**",
"kwargs",
")",
"del",
"ref",
"[",
"'region_id'",
"]",
"return",
"ref"
] | define an endpoint_ref having a pre-3 . | train | false |
36,379 | def EntityGroupName(entity):
element = entity.path().element(0)
if element.has_id():
return str(element.id())
elif element.has_name():
return element.name()
else:
return 'None'
| [
"def",
"EntityGroupName",
"(",
"entity",
")",
":",
"element",
"=",
"entity",
".",
"path",
"(",
")",
".",
"element",
"(",
"0",
")",
"if",
"element",
".",
"has_id",
"(",
")",
":",
"return",
"str",
"(",
"element",
".",
"id",
"(",
")",
")",
"elif",
"element",
".",
"has_name",
"(",
")",
":",
"return",
"element",
".",
"name",
"(",
")",
"else",
":",
"return",
"'None'"
] | given entity primary key as reference proto . | train | false |
36,380 | def _fix_dot_imports(not_consumed):
names = {}
for (name, stmts) in six.iteritems(not_consumed):
if any(((isinstance(stmt, astroid.AssName) and isinstance(stmt.ass_type(), astroid.AugAssign)) for stmt in stmts)):
continue
for stmt in stmts:
if (not isinstance(stmt, (astroid.From, astroid.Import))):
continue
for imports in stmt.names:
second_name = None
if (imports[0] == '*'):
second_name = name
elif ((imports[0].find('.') > (-1)) or (name in imports)):
second_name = imports[0]
if (second_name and (second_name not in names)):
names[second_name] = stmt
return sorted(names.items(), key=(lambda a: a[1].fromlineno))
| [
"def",
"_fix_dot_imports",
"(",
"not_consumed",
")",
":",
"names",
"=",
"{",
"}",
"for",
"(",
"name",
",",
"stmts",
")",
"in",
"six",
".",
"iteritems",
"(",
"not_consumed",
")",
":",
"if",
"any",
"(",
"(",
"(",
"isinstance",
"(",
"stmt",
",",
"astroid",
".",
"AssName",
")",
"and",
"isinstance",
"(",
"stmt",
".",
"ass_type",
"(",
")",
",",
"astroid",
".",
"AugAssign",
")",
")",
"for",
"stmt",
"in",
"stmts",
")",
")",
":",
"continue",
"for",
"stmt",
"in",
"stmts",
":",
"if",
"(",
"not",
"isinstance",
"(",
"stmt",
",",
"(",
"astroid",
".",
"From",
",",
"astroid",
".",
"Import",
")",
")",
")",
":",
"continue",
"for",
"imports",
"in",
"stmt",
".",
"names",
":",
"second_name",
"=",
"None",
"if",
"(",
"imports",
"[",
"0",
"]",
"==",
"'*'",
")",
":",
"second_name",
"=",
"name",
"elif",
"(",
"(",
"imports",
"[",
"0",
"]",
".",
"find",
"(",
"'.'",
")",
">",
"(",
"-",
"1",
")",
")",
"or",
"(",
"name",
"in",
"imports",
")",
")",
":",
"second_name",
"=",
"imports",
"[",
"0",
"]",
"if",
"(",
"second_name",
"and",
"(",
"second_name",
"not",
"in",
"names",
")",
")",
":",
"names",
"[",
"second_name",
"]",
"=",
"stmt",
"return",
"sorted",
"(",
"names",
".",
"items",
"(",
")",
",",
"key",
"=",
"(",
"lambda",
"a",
":",
"a",
"[",
"1",
"]",
".",
"fromlineno",
")",
")"
] | try to fix imports with multiple dots . | train | false |
36,381 | def inet_aton(text):
parts = text.split('.')
if (len(parts) != 4):
raise dns.exception.SyntaxError
for part in parts:
if (not part.isdigit()):
raise dns.exception.SyntaxError
if ((len(part) > 1) and (part[0] == '0')):
raise dns.exception.SyntaxError
try:
bytes = [int(part) for part in parts]
return struct.pack('BBBB', *bytes)
except:
raise dns.exception.SyntaxError
| [
"def",
"inet_aton",
"(",
"text",
")",
":",
"parts",
"=",
"text",
".",
"split",
"(",
"'.'",
")",
"if",
"(",
"len",
"(",
"parts",
")",
"!=",
"4",
")",
":",
"raise",
"dns",
".",
"exception",
".",
"SyntaxError",
"for",
"part",
"in",
"parts",
":",
"if",
"(",
"not",
"part",
".",
"isdigit",
"(",
")",
")",
":",
"raise",
"dns",
".",
"exception",
".",
"SyntaxError",
"if",
"(",
"(",
"len",
"(",
"part",
")",
">",
"1",
")",
"and",
"(",
"part",
"[",
"0",
"]",
"==",
"'0'",
")",
")",
":",
"raise",
"dns",
".",
"exception",
".",
"SyntaxError",
"try",
":",
"bytes",
"=",
"[",
"int",
"(",
"part",
")",
"for",
"part",
"in",
"parts",
"]",
"return",
"struct",
".",
"pack",
"(",
"'BBBB'",
",",
"*",
"bytes",
")",
"except",
":",
"raise",
"dns",
".",
"exception",
".",
"SyntaxError"
] | inet_aton -> packed 32-bit ip representation convert an ip address in string format to the 32-bit packed binary format used in low-level network functions . | train | false |
36,382 | def make_bpd_error(s_code, s_message):
class NewBPDError(BPDError, ):
code = s_code
message = s_message
cmd_name = ''
index = 0
def __init__(self):
pass
return NewBPDError
| [
"def",
"make_bpd_error",
"(",
"s_code",
",",
"s_message",
")",
":",
"class",
"NewBPDError",
"(",
"BPDError",
",",
")",
":",
"code",
"=",
"s_code",
"message",
"=",
"s_message",
"cmd_name",
"=",
"''",
"index",
"=",
"0",
"def",
"__init__",
"(",
"self",
")",
":",
"pass",
"return",
"NewBPDError"
] | create a bpderror subclass for a static code and message . | train | false |
36,383 | def addPegOutput(bevel, endZ, outputs, radiusArealized, sides, start, topOverBottom):
height = abs((start.z - endZ))
bevelStartRatio = max((1.0 - (bevel / height)), 0.5)
oneMinusBevelStartRatio = (1.0 - bevelStartRatio)
trunkEndZ = ((bevelStartRatio * endZ) + (oneMinusBevelStartRatio * start.z))
trunkTopOverBottom = ((bevelStartRatio * topOverBottom) + oneMinusBevelStartRatio)
cylinder.addCylinderOutputByEndStart(trunkEndZ, radiusArealized, outputs, sides, start, trunkTopOverBottom)
capRadius = (radiusArealized * trunkTopOverBottom)
capStart = ((bevelStartRatio * Vector3(start.x, start.y, endZ)) + (oneMinusBevelStartRatio * start))
radiusMaximum = max(radiusArealized.real, radiusArealized.imag)
endRadiusMaximum = ((radiusMaximum * topOverBottom) - bevel)
trunkRadiusMaximum = (radiusMaximum * trunkTopOverBottom)
capTopOverBottom = (endRadiusMaximum / trunkRadiusMaximum)
cylinder.addCylinderOutputByEndStart(endZ, capRadius, outputs, sides, capStart, capTopOverBottom)
| [
"def",
"addPegOutput",
"(",
"bevel",
",",
"endZ",
",",
"outputs",
",",
"radiusArealized",
",",
"sides",
",",
"start",
",",
"topOverBottom",
")",
":",
"height",
"=",
"abs",
"(",
"(",
"start",
".",
"z",
"-",
"endZ",
")",
")",
"bevelStartRatio",
"=",
"max",
"(",
"(",
"1.0",
"-",
"(",
"bevel",
"/",
"height",
")",
")",
",",
"0.5",
")",
"oneMinusBevelStartRatio",
"=",
"(",
"1.0",
"-",
"bevelStartRatio",
")",
"trunkEndZ",
"=",
"(",
"(",
"bevelStartRatio",
"*",
"endZ",
")",
"+",
"(",
"oneMinusBevelStartRatio",
"*",
"start",
".",
"z",
")",
")",
"trunkTopOverBottom",
"=",
"(",
"(",
"bevelStartRatio",
"*",
"topOverBottom",
")",
"+",
"oneMinusBevelStartRatio",
")",
"cylinder",
".",
"addCylinderOutputByEndStart",
"(",
"trunkEndZ",
",",
"radiusArealized",
",",
"outputs",
",",
"sides",
",",
"start",
",",
"trunkTopOverBottom",
")",
"capRadius",
"=",
"(",
"radiusArealized",
"*",
"trunkTopOverBottom",
")",
"capStart",
"=",
"(",
"(",
"bevelStartRatio",
"*",
"Vector3",
"(",
"start",
".",
"x",
",",
"start",
".",
"y",
",",
"endZ",
")",
")",
"+",
"(",
"oneMinusBevelStartRatio",
"*",
"start",
")",
")",
"radiusMaximum",
"=",
"max",
"(",
"radiusArealized",
".",
"real",
",",
"radiusArealized",
".",
"imag",
")",
"endRadiusMaximum",
"=",
"(",
"(",
"radiusMaximum",
"*",
"topOverBottom",
")",
"-",
"bevel",
")",
"trunkRadiusMaximum",
"=",
"(",
"radiusMaximum",
"*",
"trunkTopOverBottom",
")",
"capTopOverBottom",
"=",
"(",
"endRadiusMaximum",
"/",
"trunkRadiusMaximum",
")",
"cylinder",
".",
"addCylinderOutputByEndStart",
"(",
"endZ",
",",
"capRadius",
",",
"outputs",
",",
"sides",
",",
"capStart",
",",
"capTopOverBottom",
")"
] | add beveled cylinder to outputs given bevel . | train | false |
36,384 | def get_display_time_zone(time_zone_name):
time_zone = timezone(time_zone_name)
tz_abbr = get_time_zone_abbr(time_zone)
tz_offset = get_time_zone_offset(time_zone)
return '{name} ({abbr}, UTC{offset})'.format(name=time_zone, abbr=tz_abbr, offset=tz_offset).replace('_', ' ')
| [
"def",
"get_display_time_zone",
"(",
"time_zone_name",
")",
":",
"time_zone",
"=",
"timezone",
"(",
"time_zone_name",
")",
"tz_abbr",
"=",
"get_time_zone_abbr",
"(",
"time_zone",
")",
"tz_offset",
"=",
"get_time_zone_offset",
"(",
"time_zone",
")",
"return",
"'{name} ({abbr}, UTC{offset})'",
".",
"format",
"(",
"name",
"=",
"time_zone",
",",
"abbr",
"=",
"tz_abbr",
",",
"offset",
"=",
"tz_offset",
")",
".",
"replace",
"(",
"'_'",
",",
"' '",
")"
] | returns a formatted display time zone (e . | train | false |
36,385 | def project_to_sphere(points, center, radius):
lengths = scipy.spatial.distance.cdist(points, np.array([center]))
return ((((points - center) / lengths) * radius) + center)
| [
"def",
"project_to_sphere",
"(",
"points",
",",
"center",
",",
"radius",
")",
":",
"lengths",
"=",
"scipy",
".",
"spatial",
".",
"distance",
".",
"cdist",
"(",
"points",
",",
"np",
".",
"array",
"(",
"[",
"center",
"]",
")",
")",
"return",
"(",
"(",
"(",
"(",
"points",
"-",
"center",
")",
"/",
"lengths",
")",
"*",
"radius",
")",
"+",
"center",
")"
] | projects the elements of points onto the sphere defined by center and radius . | train | false |
36,386 | def bpm(max_strokes):
t0 = None
dt = []
for i in range(max_strokes):
s = input()
if (s == ''):
t1 = time.time()
if t0:
dt.append((t1 - t0))
t0 = t1
else:
break
ave = (sum([((1.0 / dti) * 60) for dti in dt]) / len(dt))
return ave
| [
"def",
"bpm",
"(",
"max_strokes",
")",
":",
"t0",
"=",
"None",
"dt",
"=",
"[",
"]",
"for",
"i",
"in",
"range",
"(",
"max_strokes",
")",
":",
"s",
"=",
"input",
"(",
")",
"if",
"(",
"s",
"==",
"''",
")",
":",
"t1",
"=",
"time",
".",
"time",
"(",
")",
"if",
"t0",
":",
"dt",
".",
"append",
"(",
"(",
"t1",
"-",
"t0",
")",
")",
"t0",
"=",
"t1",
"else",
":",
"break",
"ave",
"=",
"(",
"sum",
"(",
"[",
"(",
"(",
"1.0",
"/",
"dti",
")",
"*",
"60",
")",
"for",
"dti",
"in",
"dt",
"]",
")",
"/",
"len",
"(",
"dt",
")",
")",
"return",
"ave"
] | returns average bpm listening to enter keystrokes . | train | false |
36,388 | def to_cudandarray(x):
if (not isinstance(x, pycuda.gpuarray.GPUArray)):
raise ValueError('We can transfer only pycuda.gpuarray.GPUArray to CudaNdarray')
elif (x.dtype != 'float32'):
raise ValueError('CudaNdarray support only float32')
else:
strides = [1]
for i in x.shape[::(-1)][:(-1)]:
strides.append((strides[(-1)] * i))
strides = tuple(strides[::(-1)])
ptr = int(x.gpudata)
z = cuda.from_gpu_pointer(ptr, x.shape, strides, x)
return z
| [
"def",
"to_cudandarray",
"(",
"x",
")",
":",
"if",
"(",
"not",
"isinstance",
"(",
"x",
",",
"pycuda",
".",
"gpuarray",
".",
"GPUArray",
")",
")",
":",
"raise",
"ValueError",
"(",
"'We can transfer only pycuda.gpuarray.GPUArray to CudaNdarray'",
")",
"elif",
"(",
"x",
".",
"dtype",
"!=",
"'float32'",
")",
":",
"raise",
"ValueError",
"(",
"'CudaNdarray support only float32'",
")",
"else",
":",
"strides",
"=",
"[",
"1",
"]",
"for",
"i",
"in",
"x",
".",
"shape",
"[",
":",
":",
"(",
"-",
"1",
")",
"]",
"[",
":",
"(",
"-",
"1",
")",
"]",
":",
"strides",
".",
"append",
"(",
"(",
"strides",
"[",
"(",
"-",
"1",
")",
"]",
"*",
"i",
")",
")",
"strides",
"=",
"tuple",
"(",
"strides",
"[",
":",
":",
"(",
"-",
"1",
")",
"]",
")",
"ptr",
"=",
"int",
"(",
"x",
".",
"gpudata",
")",
"z",
"=",
"cuda",
".",
"from_gpu_pointer",
"(",
"ptr",
",",
"x",
".",
"shape",
",",
"strides",
",",
"x",
")",
"return",
"z"
] | take a pycuda . | train | false |
36,389 | def tst_render_template(template, context):
return '<div>{0}</div>'.format(saxutils.escape(repr(context)))
| [
"def",
"tst_render_template",
"(",
"template",
",",
"context",
")",
":",
"return",
"'<div>{0}</div>'",
".",
"format",
"(",
"saxutils",
".",
"escape",
"(",
"repr",
"(",
"context",
")",
")",
")"
] | a test version of render to template . | train | false |
36,390 | @login_required
def invitation_create(request, template_name='invitations/invitation_form.html', success_template_name='invitations/invitation_success.html'):
try:
allotment = request.user.invitation_allotment
if (allotment.amount == 0):
return invitation_error(request)
except InvitationAllotment.DoesNotExist:
return invitation_error(request)
if (request.method == 'POST'):
form = InvitationForm(request.POST)
if form.is_valid():
invitation = form.save(commit=False)
invitation.token = Invitation.objects.create_token(invitation.email)
invitation.from_user = request.user
invitation.save()
allotment.decrement(1)
send_invitation_email(invitation)
return render(request, success_template_name)
else:
form = InvitationForm()
return render(request, template_name, {'form': form})
| [
"@",
"login_required",
"def",
"invitation_create",
"(",
"request",
",",
"template_name",
"=",
"'invitations/invitation_form.html'",
",",
"success_template_name",
"=",
"'invitations/invitation_success.html'",
")",
":",
"try",
":",
"allotment",
"=",
"request",
".",
"user",
".",
"invitation_allotment",
"if",
"(",
"allotment",
".",
"amount",
"==",
"0",
")",
":",
"return",
"invitation_error",
"(",
"request",
")",
"except",
"InvitationAllotment",
".",
"DoesNotExist",
":",
"return",
"invitation_error",
"(",
"request",
")",
"if",
"(",
"request",
".",
"method",
"==",
"'POST'",
")",
":",
"form",
"=",
"InvitationForm",
"(",
"request",
".",
"POST",
")",
"if",
"form",
".",
"is_valid",
"(",
")",
":",
"invitation",
"=",
"form",
".",
"save",
"(",
"commit",
"=",
"False",
")",
"invitation",
".",
"token",
"=",
"Invitation",
".",
"objects",
".",
"create_token",
"(",
"invitation",
".",
"email",
")",
"invitation",
".",
"from_user",
"=",
"request",
".",
"user",
"invitation",
".",
"save",
"(",
")",
"allotment",
".",
"decrement",
"(",
"1",
")",
"send_invitation_email",
"(",
"invitation",
")",
"return",
"render",
"(",
"request",
",",
"success_template_name",
")",
"else",
":",
"form",
"=",
"InvitationForm",
"(",
")",
"return",
"render",
"(",
"request",
",",
"template_name",
",",
"{",
"'form'",
":",
"form",
"}",
")"
] | returns a form for a user to send an invitation . | train | false |
36,392 | def make_net_dict(net_id, net_name, ports):
res = {const.NET_ID: net_id, const.NET_NAME: net_name}
res[const.NET_PORTS] = ports
return res
| [
"def",
"make_net_dict",
"(",
"net_id",
",",
"net_name",
",",
"ports",
")",
":",
"res",
"=",
"{",
"const",
".",
"NET_ID",
":",
"net_id",
",",
"const",
".",
"NET_NAME",
":",
"net_name",
"}",
"res",
"[",
"const",
".",
"NET_PORTS",
"]",
"=",
"ports",
"return",
"res"
] | helper funciton . | train | false |
36,395 | @curry
def delayed(obj, name=None, pure=False, nout=None, traverse=True):
if isinstance(obj, Delayed):
return obj
if (isinstance(obj, base.Base) or traverse):
(task, dasks) = to_task_dasks(obj)
else:
task = quote(obj)
dasks = []
if (task is obj):
if (not ((nout is None) or ((type(nout) is int) and (nout >= 0)))):
raise ValueError(('nout must be None or a positive integer, got %s' % nout))
if (not name):
try:
prefix = obj.__name__
except AttributeError:
prefix = type(obj).__name__
token = tokenize(obj, nout, pure=pure)
name = ('%s-%s' % (prefix, token))
return DelayedLeaf(obj, name, pure=pure, nout=nout)
else:
if (not name):
name = ('%s-%s' % (type(obj).__name__, tokenize(task, pure=pure)))
dasks.append({name: task})
return Delayed(name, dasks)
| [
"@",
"curry",
"def",
"delayed",
"(",
"obj",
",",
"name",
"=",
"None",
",",
"pure",
"=",
"False",
",",
"nout",
"=",
"None",
",",
"traverse",
"=",
"True",
")",
":",
"if",
"isinstance",
"(",
"obj",
",",
"Delayed",
")",
":",
"return",
"obj",
"if",
"(",
"isinstance",
"(",
"obj",
",",
"base",
".",
"Base",
")",
"or",
"traverse",
")",
":",
"(",
"task",
",",
"dasks",
")",
"=",
"to_task_dasks",
"(",
"obj",
")",
"else",
":",
"task",
"=",
"quote",
"(",
"obj",
")",
"dasks",
"=",
"[",
"]",
"if",
"(",
"task",
"is",
"obj",
")",
":",
"if",
"(",
"not",
"(",
"(",
"nout",
"is",
"None",
")",
"or",
"(",
"(",
"type",
"(",
"nout",
")",
"is",
"int",
")",
"and",
"(",
"nout",
">=",
"0",
")",
")",
")",
")",
":",
"raise",
"ValueError",
"(",
"(",
"'nout must be None or a positive integer, got %s'",
"%",
"nout",
")",
")",
"if",
"(",
"not",
"name",
")",
":",
"try",
":",
"prefix",
"=",
"obj",
".",
"__name__",
"except",
"AttributeError",
":",
"prefix",
"=",
"type",
"(",
"obj",
")",
".",
"__name__",
"token",
"=",
"tokenize",
"(",
"obj",
",",
"nout",
",",
"pure",
"=",
"pure",
")",
"name",
"=",
"(",
"'%s-%s'",
"%",
"(",
"prefix",
",",
"token",
")",
")",
"return",
"DelayedLeaf",
"(",
"obj",
",",
"name",
",",
"pure",
"=",
"pure",
",",
"nout",
"=",
"nout",
")",
"else",
":",
"if",
"(",
"not",
"name",
")",
":",
"name",
"=",
"(",
"'%s-%s'",
"%",
"(",
"type",
"(",
"obj",
")",
".",
"__name__",
",",
"tokenize",
"(",
"task",
",",
"pure",
"=",
"pure",
")",
")",
")",
"dasks",
".",
"append",
"(",
"{",
"name",
":",
"task",
"}",
")",
"return",
"Delayed",
"(",
"name",
",",
"dasks",
")"
] | decorator used to capture the arguments of a function . | train | false |
36,396 | def getPathsByIntersectedLoop(begin, end, loop):
closestBeginDistanceIndex = euclidean.getClosestDistanceIndexToLine(begin, loop)
closestEndDistanceIndex = euclidean.getClosestDistanceIndexToLine(end, loop)
beginIndex = ((closestBeginDistanceIndex.index + 1) % len(loop))
endIndex = ((closestEndDistanceIndex.index + 1) % len(loop))
closestBegin = euclidean.getClosestPointOnSegment(loop[closestBeginDistanceIndex.index], loop[beginIndex], begin)
closestEnd = euclidean.getClosestPointOnSegment(loop[closestEndDistanceIndex.index], loop[endIndex], end)
clockwisePath = [closestBegin]
widdershinsPath = [closestBegin]
if (closestBeginDistanceIndex.index != closestEndDistanceIndex.index):
widdershinsPath += euclidean.getAroundLoop(beginIndex, endIndex, loop)
clockwisePath += euclidean.getAroundLoop(endIndex, beginIndex, loop)[::(-1)]
clockwisePath.append(closestEnd)
widdershinsPath.append(closestEnd)
return [clockwisePath, widdershinsPath]
| [
"def",
"getPathsByIntersectedLoop",
"(",
"begin",
",",
"end",
",",
"loop",
")",
":",
"closestBeginDistanceIndex",
"=",
"euclidean",
".",
"getClosestDistanceIndexToLine",
"(",
"begin",
",",
"loop",
")",
"closestEndDistanceIndex",
"=",
"euclidean",
".",
"getClosestDistanceIndexToLine",
"(",
"end",
",",
"loop",
")",
"beginIndex",
"=",
"(",
"(",
"closestBeginDistanceIndex",
".",
"index",
"+",
"1",
")",
"%",
"len",
"(",
"loop",
")",
")",
"endIndex",
"=",
"(",
"(",
"closestEndDistanceIndex",
".",
"index",
"+",
"1",
")",
"%",
"len",
"(",
"loop",
")",
")",
"closestBegin",
"=",
"euclidean",
".",
"getClosestPointOnSegment",
"(",
"loop",
"[",
"closestBeginDistanceIndex",
".",
"index",
"]",
",",
"loop",
"[",
"beginIndex",
"]",
",",
"begin",
")",
"closestEnd",
"=",
"euclidean",
".",
"getClosestPointOnSegment",
"(",
"loop",
"[",
"closestEndDistanceIndex",
".",
"index",
"]",
",",
"loop",
"[",
"endIndex",
"]",
",",
"end",
")",
"clockwisePath",
"=",
"[",
"closestBegin",
"]",
"widdershinsPath",
"=",
"[",
"closestBegin",
"]",
"if",
"(",
"closestBeginDistanceIndex",
".",
"index",
"!=",
"closestEndDistanceIndex",
".",
"index",
")",
":",
"widdershinsPath",
"+=",
"euclidean",
".",
"getAroundLoop",
"(",
"beginIndex",
",",
"endIndex",
",",
"loop",
")",
"clockwisePath",
"+=",
"euclidean",
".",
"getAroundLoop",
"(",
"endIndex",
",",
"beginIndex",
",",
"loop",
")",
"[",
":",
":",
"(",
"-",
"1",
")",
"]",
"clockwisePath",
".",
"append",
"(",
"closestEnd",
")",
"widdershinsPath",
".",
"append",
"(",
"closestEnd",
")",
"return",
"[",
"clockwisePath",
",",
"widdershinsPath",
"]"
] | get both paths along the loop from the point nearest to the begin to the point nearest to the end . | train | false |
36,397 | def prepare_site_db_and_overrides():
_SITE_DB.clear()
_SITE_DB[_MAIN_LANG] = _MAIN_SITEURL
main_siteurl = ('/' if (_MAIN_SITEURL == '') else _MAIN_SITEURL)
for (lang, overrides) in _SUBSITE_QUEUE.items():
if ('SITEURL' not in overrides):
overrides['SITEURL'] = posixpath.join(main_siteurl, lang)
_SITE_DB[lang] = overrides['SITEURL']
if ('OUTPUT_PATH' not in overrides):
overrides['OUTPUT_PATH'] = os.path.join(_MAIN_SETTINGS['OUTPUT_PATH'], lang)
if ('CACHE_PATH' not in overrides):
overrides['CACHE_PATH'] = os.path.join(_MAIN_SETTINGS['CACHE_PATH'], lang)
if ('STATIC_PATHS' not in overrides):
overrides['STATIC_PATHS'] = []
if (('THEME' not in overrides) and ('THEME_STATIC_DIR' not in overrides) and ('THEME_STATIC_PATHS' not in overrides)):
relpath = relpath_to_site(lang, _MAIN_LANG)
overrides['THEME_STATIC_DIR'] = posixpath.join(relpath, _MAIN_SETTINGS['THEME_STATIC_DIR'])
overrides['THEME_STATIC_PATHS'] = []
overrides['DEFAULT_LANG'] = lang
| [
"def",
"prepare_site_db_and_overrides",
"(",
")",
":",
"_SITE_DB",
".",
"clear",
"(",
")",
"_SITE_DB",
"[",
"_MAIN_LANG",
"]",
"=",
"_MAIN_SITEURL",
"main_siteurl",
"=",
"(",
"'/'",
"if",
"(",
"_MAIN_SITEURL",
"==",
"''",
")",
"else",
"_MAIN_SITEURL",
")",
"for",
"(",
"lang",
",",
"overrides",
")",
"in",
"_SUBSITE_QUEUE",
".",
"items",
"(",
")",
":",
"if",
"(",
"'SITEURL'",
"not",
"in",
"overrides",
")",
":",
"overrides",
"[",
"'SITEURL'",
"]",
"=",
"posixpath",
".",
"join",
"(",
"main_siteurl",
",",
"lang",
")",
"_SITE_DB",
"[",
"lang",
"]",
"=",
"overrides",
"[",
"'SITEURL'",
"]",
"if",
"(",
"'OUTPUT_PATH'",
"not",
"in",
"overrides",
")",
":",
"overrides",
"[",
"'OUTPUT_PATH'",
"]",
"=",
"os",
".",
"path",
".",
"join",
"(",
"_MAIN_SETTINGS",
"[",
"'OUTPUT_PATH'",
"]",
",",
"lang",
")",
"if",
"(",
"'CACHE_PATH'",
"not",
"in",
"overrides",
")",
":",
"overrides",
"[",
"'CACHE_PATH'",
"]",
"=",
"os",
".",
"path",
".",
"join",
"(",
"_MAIN_SETTINGS",
"[",
"'CACHE_PATH'",
"]",
",",
"lang",
")",
"if",
"(",
"'STATIC_PATHS'",
"not",
"in",
"overrides",
")",
":",
"overrides",
"[",
"'STATIC_PATHS'",
"]",
"=",
"[",
"]",
"if",
"(",
"(",
"'THEME'",
"not",
"in",
"overrides",
")",
"and",
"(",
"'THEME_STATIC_DIR'",
"not",
"in",
"overrides",
")",
"and",
"(",
"'THEME_STATIC_PATHS'",
"not",
"in",
"overrides",
")",
")",
":",
"relpath",
"=",
"relpath_to_site",
"(",
"lang",
",",
"_MAIN_LANG",
")",
"overrides",
"[",
"'THEME_STATIC_DIR'",
"]",
"=",
"posixpath",
".",
"join",
"(",
"relpath",
",",
"_MAIN_SETTINGS",
"[",
"'THEME_STATIC_DIR'",
"]",
")",
"overrides",
"[",
"'THEME_STATIC_PATHS'",
"]",
"=",
"[",
"]",
"overrides",
"[",
"'DEFAULT_LANG'",
"]",
"=",
"lang"
] | prepare overrides and create _site_db _site_db . | train | true |
36,400 | @docstring.dedent_interpd
def xscale(*args, **kwargs):
gca().set_xscale(*args, **kwargs)
| [
"@",
"docstring",
".",
"dedent_interpd",
"def",
"xscale",
"(",
"*",
"args",
",",
"**",
"kwargs",
")",
":",
"gca",
"(",
")",
".",
"set_xscale",
"(",
"*",
"args",
",",
"**",
"kwargs",
")"
] | set the scaling of the *x*-axis . | train | false |
36,401 | def collect_all_bears_from_sections(sections, log_printer):
local_bears = {}
global_bears = {}
for section in sections:
bear_dirs = sections[section].bear_dirs()
(local_bears[section], global_bears[section]) = collect_bears(bear_dirs, ['**'], [BEAR_KIND.LOCAL, BEAR_KIND.GLOBAL], log_printer, warn_if_unused_glob=False)
return (local_bears, global_bears)
| [
"def",
"collect_all_bears_from_sections",
"(",
"sections",
",",
"log_printer",
")",
":",
"local_bears",
"=",
"{",
"}",
"global_bears",
"=",
"{",
"}",
"for",
"section",
"in",
"sections",
":",
"bear_dirs",
"=",
"sections",
"[",
"section",
"]",
".",
"bear_dirs",
"(",
")",
"(",
"local_bears",
"[",
"section",
"]",
",",
"global_bears",
"[",
"section",
"]",
")",
"=",
"collect_bears",
"(",
"bear_dirs",
",",
"[",
"'**'",
"]",
",",
"[",
"BEAR_KIND",
".",
"LOCAL",
",",
"BEAR_KIND",
".",
"GLOBAL",
"]",
",",
"log_printer",
",",
"warn_if_unused_glob",
"=",
"False",
")",
"return",
"(",
"local_bears",
",",
"global_bears",
")"
] | collect all kinds of bears from bear directories given in the sections . | train | false |
36,402 | def scroll(clip, h=None, w=None, x_speed=0, y_speed=0, x_start=0, y_start=0, apply_to='mask'):
if (h is None):
h = clip.h
if (w is None):
w = clip.w
xmax = ((clip.w - w) - 1)
ymax = ((clip.h - h) - 1)
def f(gf, t):
x = max(0, min(xmax, (x_start + np.round((x_speed * t)))))
y = max(0, min(ymax, (y_start + np.round((y_speed * t)))))
return gf(t)[y:(y + h), x:(x + w)]
return clip.fl(f, apply_to=apply_to)
| [
"def",
"scroll",
"(",
"clip",
",",
"h",
"=",
"None",
",",
"w",
"=",
"None",
",",
"x_speed",
"=",
"0",
",",
"y_speed",
"=",
"0",
",",
"x_start",
"=",
"0",
",",
"y_start",
"=",
"0",
",",
"apply_to",
"=",
"'mask'",
")",
":",
"if",
"(",
"h",
"is",
"None",
")",
":",
"h",
"=",
"clip",
".",
"h",
"if",
"(",
"w",
"is",
"None",
")",
":",
"w",
"=",
"clip",
".",
"w",
"xmax",
"=",
"(",
"(",
"clip",
".",
"w",
"-",
"w",
")",
"-",
"1",
")",
"ymax",
"=",
"(",
"(",
"clip",
".",
"h",
"-",
"h",
")",
"-",
"1",
")",
"def",
"f",
"(",
"gf",
",",
"t",
")",
":",
"x",
"=",
"max",
"(",
"0",
",",
"min",
"(",
"xmax",
",",
"(",
"x_start",
"+",
"np",
".",
"round",
"(",
"(",
"x_speed",
"*",
"t",
")",
")",
")",
")",
")",
"y",
"=",
"max",
"(",
"0",
",",
"min",
"(",
"ymax",
",",
"(",
"y_start",
"+",
"np",
".",
"round",
"(",
"(",
"y_speed",
"*",
"t",
")",
")",
")",
")",
")",
"return",
"gf",
"(",
"t",
")",
"[",
"y",
":",
"(",
"y",
"+",
"h",
")",
",",
"x",
":",
"(",
"x",
"+",
"w",
")",
"]",
"return",
"clip",
".",
"fl",
"(",
"f",
",",
"apply_to",
"=",
"apply_to",
")"
] | scrolls horizontally or vertically a clip . | train | false |
36,403 | def limited_join(sep, items, max_chars=30, overflow_marker='...'):
full_str = sep.join(items)
if (len(full_str) < max_chars):
return full_str
n_chars = 0
n_items = 0
for (j, item) in enumerate(items):
n_chars += (len(item) + len(sep))
if (n_chars < (max_chars - len(overflow_marker))):
n_items += 1
else:
break
return sep.join((list(items[:n_items]) + [overflow_marker]))
| [
"def",
"limited_join",
"(",
"sep",
",",
"items",
",",
"max_chars",
"=",
"30",
",",
"overflow_marker",
"=",
"'...'",
")",
":",
"full_str",
"=",
"sep",
".",
"join",
"(",
"items",
")",
"if",
"(",
"len",
"(",
"full_str",
")",
"<",
"max_chars",
")",
":",
"return",
"full_str",
"n_chars",
"=",
"0",
"n_items",
"=",
"0",
"for",
"(",
"j",
",",
"item",
")",
"in",
"enumerate",
"(",
"items",
")",
":",
"n_chars",
"+=",
"(",
"len",
"(",
"item",
")",
"+",
"len",
"(",
"sep",
")",
")",
"if",
"(",
"n_chars",
"<",
"(",
"max_chars",
"-",
"len",
"(",
"overflow_marker",
")",
")",
")",
":",
"n_items",
"+=",
"1",
"else",
":",
"break",
"return",
"sep",
".",
"join",
"(",
"(",
"list",
"(",
"items",
"[",
":",
"n_items",
"]",
")",
"+",
"[",
"overflow_marker",
"]",
")",
")"
] | join a number of strings to one . | train | true |
36,404 | def lnsum0(lx):
lx = numpy.asarray(lx)
base = lx.max()
x = numpy.exp((lx - base))
ssum = x.sum(0)
normalized = nsum0(lx)
result = numpy.log(normalized)
conventional = numpy.log((numpy.exp(lx).sum(0) / numpy.exp(lx).sum()))
assert similar(result, conventional)
return result
| [
"def",
"lnsum0",
"(",
"lx",
")",
":",
"lx",
"=",
"numpy",
".",
"asarray",
"(",
"lx",
")",
"base",
"=",
"lx",
".",
"max",
"(",
")",
"x",
"=",
"numpy",
".",
"exp",
"(",
"(",
"lx",
"-",
"base",
")",
")",
"ssum",
"=",
"x",
".",
"sum",
"(",
"0",
")",
"normalized",
"=",
"nsum0",
"(",
"lx",
")",
"result",
"=",
"numpy",
".",
"log",
"(",
"normalized",
")",
"conventional",
"=",
"numpy",
".",
"log",
"(",
"(",
"numpy",
".",
"exp",
"(",
"lx",
")",
".",
"sum",
"(",
"0",
")",
"/",
"numpy",
".",
"exp",
"(",
"lx",
")",
".",
"sum",
"(",
")",
")",
")",
"assert",
"similar",
"(",
"result",
",",
"conventional",
")",
"return",
"result"
] | accepts log-values as input . | train | true |
36,405 | def initialize_collector(cls, name=None, configfile=None, handlers=[]):
log = logging.getLogger('diamond')
collector = None
try:
collector = cls(name=name, configfile=configfile, handlers=handlers)
except Exception:
log.error('Failed to initialize Collector: %s. %s', cls.__name__, traceback.format_exc())
return collector
| [
"def",
"initialize_collector",
"(",
"cls",
",",
"name",
"=",
"None",
",",
"configfile",
"=",
"None",
",",
"handlers",
"=",
"[",
"]",
")",
":",
"log",
"=",
"logging",
".",
"getLogger",
"(",
"'diamond'",
")",
"collector",
"=",
"None",
"try",
":",
"collector",
"=",
"cls",
"(",
"name",
"=",
"name",
",",
"configfile",
"=",
"configfile",
",",
"handlers",
"=",
"handlers",
")",
"except",
"Exception",
":",
"log",
".",
"error",
"(",
"'Failed to initialize Collector: %s. %s'",
",",
"cls",
".",
"__name__",
",",
"traceback",
".",
"format_exc",
"(",
")",
")",
"return",
"collector"
] | initialize collector . | train | false |
36,406 | def post_save_user(instance, raw, created, **kwargs):
from cms.utils.permissions import get_current_user
creator = get_current_user()
if ((not creator) or (not created) or creator.is_anonymous()):
return
page_user = PageUser(user_ptr_id=instance.pk, created_by=creator)
page_user.__dict__.update(instance.__dict__)
page_user.save()
| [
"def",
"post_save_user",
"(",
"instance",
",",
"raw",
",",
"created",
",",
"**",
"kwargs",
")",
":",
"from",
"cms",
".",
"utils",
".",
"permissions",
"import",
"get_current_user",
"creator",
"=",
"get_current_user",
"(",
")",
"if",
"(",
"(",
"not",
"creator",
")",
"or",
"(",
"not",
"created",
")",
"or",
"creator",
".",
"is_anonymous",
"(",
")",
")",
":",
"return",
"page_user",
"=",
"PageUser",
"(",
"user_ptr_id",
"=",
"instance",
".",
"pk",
",",
"created_by",
"=",
"creator",
")",
"page_user",
".",
"__dict__",
".",
"update",
"(",
"instance",
".",
"__dict__",
")",
"page_user",
".",
"save",
"(",
")"
] | signal called when new user is created . | train | false |
36,409 | def dup_zz_factor(f, K):
(cont, g) = dup_primitive(f, K)
n = dup_degree(g)
if (dup_LC(g, K) < 0):
(cont, g) = ((- cont), dup_neg(g, K))
if (n <= 0):
return (cont, [])
elif (n == 1):
return (cont, [(g, 1)])
if query('USE_IRREDUCIBLE_IN_FACTOR'):
if dup_zz_irreducible_p(g, K):
return (cont, [(g, 1)])
g = dup_sqf_part(g, K)
H = None
if query('USE_CYCLOTOMIC_FACTOR'):
H = dup_zz_cyclotomic_factor(g, K)
if (H is None):
H = dup_zz_zassenhaus(g, K)
factors = dup_trial_division(f, H, K)
return (cont, factors)
| [
"def",
"dup_zz_factor",
"(",
"f",
",",
"K",
")",
":",
"(",
"cont",
",",
"g",
")",
"=",
"dup_primitive",
"(",
"f",
",",
"K",
")",
"n",
"=",
"dup_degree",
"(",
"g",
")",
"if",
"(",
"dup_LC",
"(",
"g",
",",
"K",
")",
"<",
"0",
")",
":",
"(",
"cont",
",",
"g",
")",
"=",
"(",
"(",
"-",
"cont",
")",
",",
"dup_neg",
"(",
"g",
",",
"K",
")",
")",
"if",
"(",
"n",
"<=",
"0",
")",
":",
"return",
"(",
"cont",
",",
"[",
"]",
")",
"elif",
"(",
"n",
"==",
"1",
")",
":",
"return",
"(",
"cont",
",",
"[",
"(",
"g",
",",
"1",
")",
"]",
")",
"if",
"query",
"(",
"'USE_IRREDUCIBLE_IN_FACTOR'",
")",
":",
"if",
"dup_zz_irreducible_p",
"(",
"g",
",",
"K",
")",
":",
"return",
"(",
"cont",
",",
"[",
"(",
"g",
",",
"1",
")",
"]",
")",
"g",
"=",
"dup_sqf_part",
"(",
"g",
",",
"K",
")",
"H",
"=",
"None",
"if",
"query",
"(",
"'USE_CYCLOTOMIC_FACTOR'",
")",
":",
"H",
"=",
"dup_zz_cyclotomic_factor",
"(",
"g",
",",
"K",
")",
"if",
"(",
"H",
"is",
"None",
")",
":",
"H",
"=",
"dup_zz_zassenhaus",
"(",
"g",
",",
"K",
")",
"factors",
"=",
"dup_trial_division",
"(",
"f",
",",
"H",
",",
"K",
")",
"return",
"(",
"cont",
",",
"factors",
")"
] | factor polynomials in z[x] . | train | false |
36,410 | def project_root(parser, projdir=os.getcwd()):
previousdir = curdir = osp.abspath(projdir)
testercls = PyTester
conf_file_path = osp.join(curdir, CONF_FILE)
if osp.isfile(conf_file_path):
testercls = load_pytest_conf(conf_file_path, parser)
while (this_is_a_testdir(curdir) or osp.isfile(osp.join(curdir, '__init__.py'))):
newdir = osp.normpath(osp.join(curdir, os.pardir))
if (newdir == curdir):
break
previousdir = curdir
curdir = newdir
conf_file_path = osp.join(curdir, CONF_FILE)
if osp.isfile(conf_file_path):
testercls = load_pytest_conf(conf_file_path, parser)
return (previousdir, testercls)
| [
"def",
"project_root",
"(",
"parser",
",",
"projdir",
"=",
"os",
".",
"getcwd",
"(",
")",
")",
":",
"previousdir",
"=",
"curdir",
"=",
"osp",
".",
"abspath",
"(",
"projdir",
")",
"testercls",
"=",
"PyTester",
"conf_file_path",
"=",
"osp",
".",
"join",
"(",
"curdir",
",",
"CONF_FILE",
")",
"if",
"osp",
".",
"isfile",
"(",
"conf_file_path",
")",
":",
"testercls",
"=",
"load_pytest_conf",
"(",
"conf_file_path",
",",
"parser",
")",
"while",
"(",
"this_is_a_testdir",
"(",
"curdir",
")",
"or",
"osp",
".",
"isfile",
"(",
"osp",
".",
"join",
"(",
"curdir",
",",
"'__init__.py'",
")",
")",
")",
":",
"newdir",
"=",
"osp",
".",
"normpath",
"(",
"osp",
".",
"join",
"(",
"curdir",
",",
"os",
".",
"pardir",
")",
")",
"if",
"(",
"newdir",
"==",
"curdir",
")",
":",
"break",
"previousdir",
"=",
"curdir",
"curdir",
"=",
"newdir",
"conf_file_path",
"=",
"osp",
".",
"join",
"(",
"curdir",
",",
"CONF_FILE",
")",
"if",
"osp",
".",
"isfile",
"(",
"conf_file_path",
")",
":",
"testercls",
"=",
"load_pytest_conf",
"(",
"conf_file_path",
",",
"parser",
")",
"return",
"(",
"previousdir",
",",
"testercls",
")"
] | try to find projects root and add it to sys . | train | false |
36,411 | def preserve_plugin_listeners():
if (not ImportAddedPlugin.listeners):
ImportAddedPlugin.listeners = _listeners
| [
"def",
"preserve_plugin_listeners",
"(",
")",
":",
"if",
"(",
"not",
"ImportAddedPlugin",
".",
"listeners",
")",
":",
"ImportAddedPlugin",
".",
"listeners",
"=",
"_listeners"
] | preserve the initial plugin listeners as they would otherwise be deleted after the first setup / tear down cycle . | train | false |
36,412 | def css(state, text, i, formats, user_data):
pat = cdata_close_pats[u'style']
m = pat.search(text, i)
if (m is None):
css_text = text[i:]
else:
css_text = text[i:m.start()]
ans = []
css_user_data = user_data.css_user_data = (user_data.css_user_data or CSSUserData())
state.sub_parser_state = css_user_data.state = (state.sub_parser_state or CSSState())
for (j, num, fmt) in run_loop(css_user_data, css_state_map, formats[u'css_sub_formats'], css_text):
ans.append((num, fmt))
if (m is not None):
state.sub_parser_state = None
state.parse = IN_CLOSING_TAG
add_tag_data(user_data, TagStart(m.start(), u'', u'style', True, True))
ans.extend([(2, formats[u'end_tag']), ((len(m.group()) - 2), formats[u'tag_name'])])
return ans
| [
"def",
"css",
"(",
"state",
",",
"text",
",",
"i",
",",
"formats",
",",
"user_data",
")",
":",
"pat",
"=",
"cdata_close_pats",
"[",
"u'style'",
"]",
"m",
"=",
"pat",
".",
"search",
"(",
"text",
",",
"i",
")",
"if",
"(",
"m",
"is",
"None",
")",
":",
"css_text",
"=",
"text",
"[",
"i",
":",
"]",
"else",
":",
"css_text",
"=",
"text",
"[",
"i",
":",
"m",
".",
"start",
"(",
")",
"]",
"ans",
"=",
"[",
"]",
"css_user_data",
"=",
"user_data",
".",
"css_user_data",
"=",
"(",
"user_data",
".",
"css_user_data",
"or",
"CSSUserData",
"(",
")",
")",
"state",
".",
"sub_parser_state",
"=",
"css_user_data",
".",
"state",
"=",
"(",
"state",
".",
"sub_parser_state",
"or",
"CSSState",
"(",
")",
")",
"for",
"(",
"j",
",",
"num",
",",
"fmt",
")",
"in",
"run_loop",
"(",
"css_user_data",
",",
"css_state_map",
",",
"formats",
"[",
"u'css_sub_formats'",
"]",
",",
"css_text",
")",
":",
"ans",
".",
"append",
"(",
"(",
"num",
",",
"fmt",
")",
")",
"if",
"(",
"m",
"is",
"not",
"None",
")",
":",
"state",
".",
"sub_parser_state",
"=",
"None",
"state",
".",
"parse",
"=",
"IN_CLOSING_TAG",
"add_tag_data",
"(",
"user_data",
",",
"TagStart",
"(",
"m",
".",
"start",
"(",
")",
",",
"u''",
",",
"u'style'",
",",
"True",
",",
"True",
")",
")",
"ans",
".",
"extend",
"(",
"[",
"(",
"2",
",",
"formats",
"[",
"u'end_tag'",
"]",
")",
",",
"(",
"(",
"len",
"(",
"m",
".",
"group",
"(",
")",
")",
"-",
"2",
")",
",",
"formats",
"[",
"u'tag_name'",
"]",
")",
"]",
")",
"return",
"ans"
] | inside a <style> tag . | train | false |
36,416 | def _get_rar_version(xfile):
with XFile(xfile) as fd:
buf = fd.read(len(RAR5_ID))
if buf.startswith(RAR_ID):
return 3
elif buf.startswith(RAR5_ID):
return 5
return 0
| [
"def",
"_get_rar_version",
"(",
"xfile",
")",
":",
"with",
"XFile",
"(",
"xfile",
")",
"as",
"fd",
":",
"buf",
"=",
"fd",
".",
"read",
"(",
"len",
"(",
"RAR5_ID",
")",
")",
"if",
"buf",
".",
"startswith",
"(",
"RAR_ID",
")",
":",
"return",
"3",
"elif",
"buf",
".",
"startswith",
"(",
"RAR5_ID",
")",
":",
"return",
"5",
"return",
"0"
] | check quickly whether file is rar archive . | train | true |
36,417 | def nzf_cmp_date(nzf1, nzf2):
return nzf_cmp_name(nzf1, nzf2, name=False)
| [
"def",
"nzf_cmp_date",
"(",
"nzf1",
",",
"nzf2",
")",
":",
"return",
"nzf_cmp_name",
"(",
"nzf1",
",",
"nzf2",
",",
"name",
"=",
"False",
")"
] | compare files based on date . | train | false |
36,419 | def virtual_interface_update(context, address, values):
return IMPL.virtual_interface_update(context, address, values)
| [
"def",
"virtual_interface_update",
"(",
"context",
",",
"address",
",",
"values",
")",
":",
"return",
"IMPL",
".",
"virtual_interface_update",
"(",
"context",
",",
"address",
",",
"values",
")"
] | create a virtual interface record in the database . | train | false |
36,420 | def prompt_for_mc_auth():
print(u'These tests will never delete or modify your music.\n\nIf the tests fail, you *might* end up with a test song/playlist in your library, though.\n')
mclient = Mobileclient()
valid_mc_auth = False
while (not valid_mc_auth):
print()
email = input(u'Email: ')
passwd = getpass()
try:
android_id = os.environ[u'GM_AA_D_ID']
except KeyError:
android_id = input(u"Device ID ('mac' for FROM_MAC_ADDRESS): ")
if (android_id == u'mac'):
android_id = Mobileclient.FROM_MAC_ADDRESS
if (not android_id):
print(u'a device id must be provided')
sys.exit(1)
valid_mc_auth = mclient.login(email, passwd, android_id)
return (email, passwd, android_id)
| [
"def",
"prompt_for_mc_auth",
"(",
")",
":",
"print",
"(",
"u'These tests will never delete or modify your music.\\n\\nIf the tests fail, you *might* end up with a test song/playlist in your library, though.\\n'",
")",
"mclient",
"=",
"Mobileclient",
"(",
")",
"valid_mc_auth",
"=",
"False",
"while",
"(",
"not",
"valid_mc_auth",
")",
":",
"print",
"(",
")",
"email",
"=",
"input",
"(",
"u'Email: '",
")",
"passwd",
"=",
"getpass",
"(",
")",
"try",
":",
"android_id",
"=",
"os",
".",
"environ",
"[",
"u'GM_AA_D_ID'",
"]",
"except",
"KeyError",
":",
"android_id",
"=",
"input",
"(",
"u\"Device ID ('mac' for FROM_MAC_ADDRESS): \"",
")",
"if",
"(",
"android_id",
"==",
"u'mac'",
")",
":",
"android_id",
"=",
"Mobileclient",
".",
"FROM_MAC_ADDRESS",
"if",
"(",
"not",
"android_id",
")",
":",
"print",
"(",
"u'a device id must be provided'",
")",
"sys",
".",
"exit",
"(",
"1",
")",
"valid_mc_auth",
"=",
"mclient",
".",
"login",
"(",
"email",
",",
"passwd",
",",
"android_id",
")",
"return",
"(",
"email",
",",
"passwd",
",",
"android_id",
")"
] | return a valid tuple by continually prompting the user . | train | false |
36,421 | @permission_required([('Websites', 'ReadAbuse')])
def queue_abuse_websites(request):
queues_helper = ReviewersQueuesHelper(request)
sites = queues_helper.get_abuse_queue_websites()
page = paginate(request, sites, per_page=20)
abuse_formset = WebsiteAbuseViewFormSet((request.POST or None), queryset=page.object_list, request=request)
if abuse_formset.is_valid():
abuse_formset.save()
return redirect(reverse('reviewers.websites.queue_abuse'))
return render(request, 'reviewers/queue.html', context(request, abuse_formset=abuse_formset, tab='abusewebsites', page=page))
| [
"@",
"permission_required",
"(",
"[",
"(",
"'Websites'",
",",
"'ReadAbuse'",
")",
"]",
")",
"def",
"queue_abuse_websites",
"(",
"request",
")",
":",
"queues_helper",
"=",
"ReviewersQueuesHelper",
"(",
"request",
")",
"sites",
"=",
"queues_helper",
".",
"get_abuse_queue_websites",
"(",
")",
"page",
"=",
"paginate",
"(",
"request",
",",
"sites",
",",
"per_page",
"=",
"20",
")",
"abuse_formset",
"=",
"WebsiteAbuseViewFormSet",
"(",
"(",
"request",
".",
"POST",
"or",
"None",
")",
",",
"queryset",
"=",
"page",
".",
"object_list",
",",
"request",
"=",
"request",
")",
"if",
"abuse_formset",
".",
"is_valid",
"(",
")",
":",
"abuse_formset",
".",
"save",
"(",
")",
"return",
"redirect",
"(",
"reverse",
"(",
"'reviewers.websites.queue_abuse'",
")",
")",
"return",
"render",
"(",
"request",
",",
"'reviewers/queue.html'",
",",
"context",
"(",
"request",
",",
"abuse_formset",
"=",
"abuse_formset",
",",
"tab",
"=",
"'abusewebsites'",
",",
"page",
"=",
"page",
")",
")"
] | queue for reviewing abuse reports for websites . | train | false |
36,422 | def get_total_citations(soup):
results = soup.find('div', attrs={'id': 'gs_ab_md'}).contents[0]
matches = re.search('About\\s(\\d+)\\s', results)
if matches:
hits = matches.groups()[0]
return hits
| [
"def",
"get_total_citations",
"(",
"soup",
")",
":",
"results",
"=",
"soup",
".",
"find",
"(",
"'div'",
",",
"attrs",
"=",
"{",
"'id'",
":",
"'gs_ab_md'",
"}",
")",
".",
"contents",
"[",
"0",
"]",
"matches",
"=",
"re",
".",
"search",
"(",
"'About\\\\s(\\\\d+)\\\\s'",
",",
"results",
")",
"if",
"matches",
":",
"hits",
"=",
"matches",
".",
"groups",
"(",
")",
"[",
"0",
"]",
"return",
"hits"
] | get total citations . | train | false |
36,425 | def make_column_label(command_name, description, default):
for (pattern, label) in description_map:
if fnmatch.fnmatch(description, pattern):
return label
return default
| [
"def",
"make_column_label",
"(",
"command_name",
",",
"description",
",",
"default",
")",
":",
"for",
"(",
"pattern",
",",
"label",
")",
"in",
"description_map",
":",
"if",
"fnmatch",
".",
"fnmatch",
"(",
"description",
",",
"pattern",
")",
":",
"return",
"label",
"return",
"default"
] | try to work out a reasonable column name from parameter description . | train | true |
36,426 | def list_storage_accounts(call=None, kwargs=None):
global storconn
if (not storconn):
storconn = get_conn(StorageManagementClient)
if (kwargs is None):
kwargs = {}
ret = {}
for acct in pages_to_list(storconn.storage_accounts.list()):
ret[acct.name] = object_to_dict(acct)
return ret
| [
"def",
"list_storage_accounts",
"(",
"call",
"=",
"None",
",",
"kwargs",
"=",
"None",
")",
":",
"global",
"storconn",
"if",
"(",
"not",
"storconn",
")",
":",
"storconn",
"=",
"get_conn",
"(",
"StorageManagementClient",
")",
"if",
"(",
"kwargs",
"is",
"None",
")",
":",
"kwargs",
"=",
"{",
"}",
"ret",
"=",
"{",
"}",
"for",
"acct",
"in",
"pages_to_list",
"(",
"storconn",
".",
"storage_accounts",
".",
"list",
"(",
")",
")",
":",
"ret",
"[",
"acct",
".",
"name",
"]",
"=",
"object_to_dict",
"(",
"acct",
")",
"return",
"ret"
] | list storage accounts . | train | false |
36,427 | @real_memoize
def is_openbsd():
return sys.platform.startswith('openbsd')
| [
"@",
"real_memoize",
"def",
"is_openbsd",
"(",
")",
":",
"return",
"sys",
".",
"platform",
".",
"startswith",
"(",
"'openbsd'",
")"
] | simple function to return if host is openbsd or not . | train | false |
36,428 | def SelectFromList(prompt, items):
vim_items = [prompt]
vim_items.extend([u'{0}: {1}'.format((i + 1), item) for (i, item) in enumerate(items)])
vim.eval(u'inputsave()')
try:
selected = (GetIntValue(((u'inputlist( ' + json.dumps(vim_items)) + u' )')) - 1)
except KeyboardInterrupt:
selected = (-1)
finally:
vim.eval(u'inputrestore()')
if ((selected < 0) or (selected >= len(items))):
raise RuntimeError(NO_SELECTION_MADE_MSG)
return selected
| [
"def",
"SelectFromList",
"(",
"prompt",
",",
"items",
")",
":",
"vim_items",
"=",
"[",
"prompt",
"]",
"vim_items",
".",
"extend",
"(",
"[",
"u'{0}: {1}'",
".",
"format",
"(",
"(",
"i",
"+",
"1",
")",
",",
"item",
")",
"for",
"(",
"i",
",",
"item",
")",
"in",
"enumerate",
"(",
"items",
")",
"]",
")",
"vim",
".",
"eval",
"(",
"u'inputsave()'",
")",
"try",
":",
"selected",
"=",
"(",
"GetIntValue",
"(",
"(",
"(",
"u'inputlist( '",
"+",
"json",
".",
"dumps",
"(",
"vim_items",
")",
")",
"+",
"u' )'",
")",
")",
"-",
"1",
")",
"except",
"KeyboardInterrupt",
":",
"selected",
"=",
"(",
"-",
"1",
")",
"finally",
":",
"vim",
".",
"eval",
"(",
"u'inputrestore()'",
")",
"if",
"(",
"(",
"selected",
"<",
"0",
")",
"or",
"(",
"selected",
">=",
"len",
"(",
"items",
")",
")",
")",
":",
"raise",
"RuntimeError",
"(",
"NO_SELECTION_MADE_MSG",
")",
"return",
"selected"
] | ask the user to select an item from the list |items| . | train | false |
36,429 | def SetComputerInfo():
info = win32net.NetWkstaGetInfo(None, 502)
| [
"def",
"SetComputerInfo",
"(",
")",
":",
"info",
"=",
"win32net",
".",
"NetWkstaGetInfo",
"(",
"None",
",",
"502",
")"
] | doesnt actually change anything . | train | false |
36,430 | def get_child_environment(parent_env, child_params, item_to_remove=None, child_resource_name=None):
def is_flat_params(env_or_param):
if (env_or_param is None):
return False
for sect in env_fmt.SECTIONS:
if (sect in env_or_param):
return False
return True
child_env = parent_env.user_env_as_dict()
child_env[env_fmt.PARAMETERS] = {}
flat_params = is_flat_params(child_params)
new_env = Environment()
if (flat_params and (child_params is not None)):
child_env[env_fmt.PARAMETERS] = child_params
new_env.load(child_env)
if ((not flat_params) and (child_params is not None)):
new_env.load(child_params)
if (item_to_remove is not None):
new_env.registry.remove_item(item_to_remove)
if child_resource_name:
new_env.registry.remove_resources_except(child_resource_name)
return new_env
| [
"def",
"get_child_environment",
"(",
"parent_env",
",",
"child_params",
",",
"item_to_remove",
"=",
"None",
",",
"child_resource_name",
"=",
"None",
")",
":",
"def",
"is_flat_params",
"(",
"env_or_param",
")",
":",
"if",
"(",
"env_or_param",
"is",
"None",
")",
":",
"return",
"False",
"for",
"sect",
"in",
"env_fmt",
".",
"SECTIONS",
":",
"if",
"(",
"sect",
"in",
"env_or_param",
")",
":",
"return",
"False",
"return",
"True",
"child_env",
"=",
"parent_env",
".",
"user_env_as_dict",
"(",
")",
"child_env",
"[",
"env_fmt",
".",
"PARAMETERS",
"]",
"=",
"{",
"}",
"flat_params",
"=",
"is_flat_params",
"(",
"child_params",
")",
"new_env",
"=",
"Environment",
"(",
")",
"if",
"(",
"flat_params",
"and",
"(",
"child_params",
"is",
"not",
"None",
")",
")",
":",
"child_env",
"[",
"env_fmt",
".",
"PARAMETERS",
"]",
"=",
"child_params",
"new_env",
".",
"load",
"(",
"child_env",
")",
"if",
"(",
"(",
"not",
"flat_params",
")",
"and",
"(",
"child_params",
"is",
"not",
"None",
")",
")",
":",
"new_env",
".",
"load",
"(",
"child_params",
")",
"if",
"(",
"item_to_remove",
"is",
"not",
"None",
")",
":",
"new_env",
".",
"registry",
".",
"remove_item",
"(",
"item_to_remove",
")",
"if",
"child_resource_name",
":",
"new_env",
".",
"registry",
".",
"remove_resources_except",
"(",
"child_resource_name",
")",
"return",
"new_env"
] | build a child environment using the parent environment and params . | train | false |
36,431 | def unpack_tarfile(filename, extract_dir, progress_filter=default_filter):
try:
tarobj = tarfile.open(filename)
except tarfile.TarError:
raise UnrecognizedFormat(('%s is not a compressed or uncompressed tar file' % (filename,)))
try:
tarobj.chown = (lambda *args: None)
for member in tarobj:
name = member.name
if ((not name.startswith('/')) and ('..' not in name.split('/'))):
prelim_dst = os.path.join(extract_dir, *name.split('/'))
while ((member is not None) and (member.islnk() or member.issym())):
linkpath = member.linkname
if member.issym():
linkpath = posixpath.join(posixpath.dirname(member.name), linkpath)
linkpath = posixpath.normpath(linkpath)
member = tarobj._getmember(linkpath)
if ((member is not None) and (member.isfile() or member.isdir())):
final_dst = progress_filter(name, prelim_dst)
if final_dst:
if final_dst.endswith(os.sep):
final_dst = final_dst[:(-1)]
try:
tarobj._extract_member(member, final_dst)
except tarfile.ExtractError:
pass
return True
finally:
tarobj.close()
| [
"def",
"unpack_tarfile",
"(",
"filename",
",",
"extract_dir",
",",
"progress_filter",
"=",
"default_filter",
")",
":",
"try",
":",
"tarobj",
"=",
"tarfile",
".",
"open",
"(",
"filename",
")",
"except",
"tarfile",
".",
"TarError",
":",
"raise",
"UnrecognizedFormat",
"(",
"(",
"'%s is not a compressed or uncompressed tar file'",
"%",
"(",
"filename",
",",
")",
")",
")",
"try",
":",
"tarobj",
".",
"chown",
"=",
"(",
"lambda",
"*",
"args",
":",
"None",
")",
"for",
"member",
"in",
"tarobj",
":",
"name",
"=",
"member",
".",
"name",
"if",
"(",
"(",
"not",
"name",
".",
"startswith",
"(",
"'/'",
")",
")",
"and",
"(",
"'..'",
"not",
"in",
"name",
".",
"split",
"(",
"'/'",
")",
")",
")",
":",
"prelim_dst",
"=",
"os",
".",
"path",
".",
"join",
"(",
"extract_dir",
",",
"*",
"name",
".",
"split",
"(",
"'/'",
")",
")",
"while",
"(",
"(",
"member",
"is",
"not",
"None",
")",
"and",
"(",
"member",
".",
"islnk",
"(",
")",
"or",
"member",
".",
"issym",
"(",
")",
")",
")",
":",
"linkpath",
"=",
"member",
".",
"linkname",
"if",
"member",
".",
"issym",
"(",
")",
":",
"linkpath",
"=",
"posixpath",
".",
"join",
"(",
"posixpath",
".",
"dirname",
"(",
"member",
".",
"name",
")",
",",
"linkpath",
")",
"linkpath",
"=",
"posixpath",
".",
"normpath",
"(",
"linkpath",
")",
"member",
"=",
"tarobj",
".",
"_getmember",
"(",
"linkpath",
")",
"if",
"(",
"(",
"member",
"is",
"not",
"None",
")",
"and",
"(",
"member",
".",
"isfile",
"(",
")",
"or",
"member",
".",
"isdir",
"(",
")",
")",
")",
":",
"final_dst",
"=",
"progress_filter",
"(",
"name",
",",
"prelim_dst",
")",
"if",
"final_dst",
":",
"if",
"final_dst",
".",
"endswith",
"(",
"os",
".",
"sep",
")",
":",
"final_dst",
"=",
"final_dst",
"[",
":",
"(",
"-",
"1",
")",
"]",
"try",
":",
"tarobj",
".",
"_extract_member",
"(",
"member",
",",
"final_dst",
")",
"except",
"tarfile",
".",
"ExtractError",
":",
"pass",
"return",
"True",
"finally",
":",
"tarobj",
".",
"close",
"(",
")"
] | unpack tar/tar . | train | true |
36,432 | def get_py_filename(name, force_win32=None):
name = os.path.expanduser(name)
if (force_win32 is not None):
warn("The 'force_win32' argument to 'get_py_filename' is deprecated since IPython 5.0 and should not be used anymore", DeprecationWarning, stacklevel=2)
if ((not os.path.isfile(name)) and (not name.endswith('.py'))):
name += '.py'
if os.path.isfile(name):
return name
else:
raise IOError(('File `%r` not found.' % name))
| [
"def",
"get_py_filename",
"(",
"name",
",",
"force_win32",
"=",
"None",
")",
":",
"name",
"=",
"os",
".",
"path",
".",
"expanduser",
"(",
"name",
")",
"if",
"(",
"force_win32",
"is",
"not",
"None",
")",
":",
"warn",
"(",
"\"The 'force_win32' argument to 'get_py_filename' is deprecated since IPython 5.0 and should not be used anymore\"",
",",
"DeprecationWarning",
",",
"stacklevel",
"=",
"2",
")",
"if",
"(",
"(",
"not",
"os",
".",
"path",
".",
"isfile",
"(",
"name",
")",
")",
"and",
"(",
"not",
"name",
".",
"endswith",
"(",
"'.py'",
")",
")",
")",
":",
"name",
"+=",
"'.py'",
"if",
"os",
".",
"path",
".",
"isfile",
"(",
"name",
")",
":",
"return",
"name",
"else",
":",
"raise",
"IOError",
"(",
"(",
"'File `%r` not found.'",
"%",
"name",
")",
")"
] | return a valid python filename in the current directory . | train | false |
36,433 | def write_t4b(t4bfile, coverdata=None):
from PIL import Image
if (coverdata is not None):
coverdata = BytesIO(coverdata)
cover = Image.open(coverdata).convert(u'L')
cover.thumbnail((96, 144), Image.ANTIALIAS)
t4bcover = Image.new(u'L', (96, 144), u'white')
(x, y) = cover.size
t4bcover.paste(cover, (((96 - x) // 2), ((144 - y) // 2)))
pxs = t4bcover.getdata()
t4bfile.write('t4bp')
data = (((16 * reduce_color(pxs[i])) + reduce_color(pxs[(i + 1)])) for i in xrange(0, len(pxs), 2))
t4bfile.write(bytes(bytearray(data)))
else:
t4bfile.write(DEFAULT_T4B_DATA)
| [
"def",
"write_t4b",
"(",
"t4bfile",
",",
"coverdata",
"=",
"None",
")",
":",
"from",
"PIL",
"import",
"Image",
"if",
"(",
"coverdata",
"is",
"not",
"None",
")",
":",
"coverdata",
"=",
"BytesIO",
"(",
"coverdata",
")",
"cover",
"=",
"Image",
".",
"open",
"(",
"coverdata",
")",
".",
"convert",
"(",
"u'L'",
")",
"cover",
".",
"thumbnail",
"(",
"(",
"96",
",",
"144",
")",
",",
"Image",
".",
"ANTIALIAS",
")",
"t4bcover",
"=",
"Image",
".",
"new",
"(",
"u'L'",
",",
"(",
"96",
",",
"144",
")",
",",
"u'white'",
")",
"(",
"x",
",",
"y",
")",
"=",
"cover",
".",
"size",
"t4bcover",
".",
"paste",
"(",
"cover",
",",
"(",
"(",
"(",
"96",
"-",
"x",
")",
"//",
"2",
")",
",",
"(",
"(",
"144",
"-",
"y",
")",
"//",
"2",
")",
")",
")",
"pxs",
"=",
"t4bcover",
".",
"getdata",
"(",
")",
"t4bfile",
".",
"write",
"(",
"'t4bp'",
")",
"data",
"=",
"(",
"(",
"(",
"16",
"*",
"reduce_color",
"(",
"pxs",
"[",
"i",
"]",
")",
")",
"+",
"reduce_color",
"(",
"pxs",
"[",
"(",
"i",
"+",
"1",
")",
"]",
")",
")",
"for",
"i",
"in",
"xrange",
"(",
"0",
",",
"len",
"(",
"pxs",
")",
",",
"2",
")",
")",
"t4bfile",
".",
"write",
"(",
"bytes",
"(",
"bytearray",
"(",
"data",
")",
")",
")",
"else",
":",
"t4bfile",
".",
"write",
"(",
"DEFAULT_T4B_DATA",
")"
] | t4bfile is a file handle ready to write binary data to disk . | train | false |
36,434 | def runs_last(func):
def Wrapper():
calls = func.num_host_calls
if (calls >= (len(env.hosts) - 1)):
return func()
else:
func.num_host_calls = (calls + 1)
return None
setattr(func, 'num_host_calls', 0)
return Wrapper
| [
"def",
"runs_last",
"(",
"func",
")",
":",
"def",
"Wrapper",
"(",
")",
":",
"calls",
"=",
"func",
".",
"num_host_calls",
"if",
"(",
"calls",
">=",
"(",
"len",
"(",
"env",
".",
"hosts",
")",
"-",
"1",
")",
")",
":",
"return",
"func",
"(",
")",
"else",
":",
"func",
".",
"num_host_calls",
"=",
"(",
"calls",
"+",
"1",
")",
"return",
"None",
"setattr",
"(",
"func",
",",
"'num_host_calls'",
",",
"0",
")",
"return",
"Wrapper"
] | decorator to run a function only on the last invocation . | train | false |
36,435 | def checkForSerialEvents(read_time, rx_data, parser_state, **kwargs):
print2err('checkForSerialEvents called: ', (read_time, rx_data, parser_state, kwargs))
parser_state['last_time'] = read_time
serial_events = []
if (rx_data == 'TEST'):
serial_events.append({'data': rx_data})
return serial_events
| [
"def",
"checkForSerialEvents",
"(",
"read_time",
",",
"rx_data",
",",
"parser_state",
",",
"**",
"kwargs",
")",
":",
"print2err",
"(",
"'checkForSerialEvents called: '",
",",
"(",
"read_time",
",",
"rx_data",
",",
"parser_state",
",",
"kwargs",
")",
")",
"parser_state",
"[",
"'last_time'",
"]",
"=",
"read_time",
"serial_events",
"=",
"[",
"]",
"if",
"(",
"rx_data",
"==",
"'TEST'",
")",
":",
"serial_events",
".",
"append",
"(",
"{",
"'data'",
":",
"rx_data",
"}",
")",
"return",
"serial_events"
] | must have the following signature: evt_list = somecustomparsername where: read_time: the time when the serial device read() returned with the new rx_data . | train | false |
36,436 | @login_required
def upgrade_account(request, acc_type):
user = request.user
bill = None
try:
bill = user.bill
except:
bill = Bill(user=user)
bill.upgrade_type = acc_type
bill.save()
tn = bill.pk
if (acc_type == 'bronze'):
url = create_partner_trade_by_buyer(tn, u'ikindle\u6742\u5fd7\u8ba2\u9605(4\u4efd)', u'\u8ba2\u9605\u6742\u5fd7\u5230\u4f60\u7684Kindle\uff0c 2.99x6\u4e2a\u6708', '0.01')
return HttpResponseRedirect(url)
elif (acc_type == 'silver'):
url = create_partner_trade_by_buyer(tn, u'ikindle\u6742\u5fd7\u8ba2\u9605(6\u4efd)', u'\u8ba2\u9605\u6742\u5fd7\u5230\u4f60\u7684Kindle\uff0c3.99x6\u4e2a\u6708', '0.01')
return HttpResponseRedirect(url)
elif (acc_type == 'gold'):
url = create_partner_trade_by_buyer(tn, u'ikindle\u6742\u5fd7\u8ba2\u9605(\u65e0\u9650\u5236)', u'\u8ba2\u9605\u6742\u5fd7\u5230\u4f60\u7684Kindle\uff0c5.99x6\u4e2a\u6708', '0.01')
return HttpResponseRedirect(url)
else:
return HttpResponseRedirect(reverse('payment_error'))
| [
"@",
"login_required",
"def",
"upgrade_account",
"(",
"request",
",",
"acc_type",
")",
":",
"user",
"=",
"request",
".",
"user",
"bill",
"=",
"None",
"try",
":",
"bill",
"=",
"user",
".",
"bill",
"except",
":",
"bill",
"=",
"Bill",
"(",
"user",
"=",
"user",
")",
"bill",
".",
"upgrade_type",
"=",
"acc_type",
"bill",
".",
"save",
"(",
")",
"tn",
"=",
"bill",
".",
"pk",
"if",
"(",
"acc_type",
"==",
"'bronze'",
")",
":",
"url",
"=",
"create_partner_trade_by_buyer",
"(",
"tn",
",",
"u'ikindle\\u6742\\u5fd7\\u8ba2\\u9605(4\\u4efd)'",
",",
"u'\\u8ba2\\u9605\\u6742\\u5fd7\\u5230\\u4f60\\u7684Kindle\\uff0c 2.99x6\\u4e2a\\u6708'",
",",
"'0.01'",
")",
"return",
"HttpResponseRedirect",
"(",
"url",
")",
"elif",
"(",
"acc_type",
"==",
"'silver'",
")",
":",
"url",
"=",
"create_partner_trade_by_buyer",
"(",
"tn",
",",
"u'ikindle\\u6742\\u5fd7\\u8ba2\\u9605(6\\u4efd)'",
",",
"u'\\u8ba2\\u9605\\u6742\\u5fd7\\u5230\\u4f60\\u7684Kindle\\uff0c3.99x6\\u4e2a\\u6708'",
",",
"'0.01'",
")",
"return",
"HttpResponseRedirect",
"(",
"url",
")",
"elif",
"(",
"acc_type",
"==",
"'gold'",
")",
":",
"url",
"=",
"create_partner_trade_by_buyer",
"(",
"tn",
",",
"u'ikindle\\u6742\\u5fd7\\u8ba2\\u9605(\\u65e0\\u9650\\u5236)'",
",",
"u'\\u8ba2\\u9605\\u6742\\u5fd7\\u5230\\u4f60\\u7684Kindle\\uff0c5.99x6\\u4e2a\\u6708'",
",",
"'0.01'",
")",
"return",
"HttpResponseRedirect",
"(",
"url",
")",
"else",
":",
"return",
"HttpResponseRedirect",
"(",
"reverse",
"(",
"'payment_error'",
")",
")"
] | request for upgrade account to acc_type . | train | false |
36,437 | def create_index():
for (k, v) in INDEX.items():
for (key, kwargs) in v.items():
client[DATABASE_NAME][k].ensure_index((list(key) if (type(key) == types.TupleType) else key), **kwargs)
| [
"def",
"create_index",
"(",
")",
":",
"for",
"(",
"k",
",",
"v",
")",
"in",
"INDEX",
".",
"items",
"(",
")",
":",
"for",
"(",
"key",
",",
"kwargs",
")",
"in",
"v",
".",
"items",
"(",
")",
":",
"client",
"[",
"DATABASE_NAME",
"]",
"[",
"k",
"]",
".",
"ensure_index",
"(",
"(",
"list",
"(",
"key",
")",
"if",
"(",
"type",
"(",
"key",
")",
"==",
"types",
".",
"TupleType",
")",
"else",
"key",
")",
",",
"**",
"kwargs",
")"
] | create an index for a single column . | train | false |
36,438 | def remove_outliers(seqs, num_stds, fraction_seqs_for_stats=0.95):
aln = Alignment.from_fasta_records(parse_fasta(seqs), DNA)
consensus_seq = aln.majority_consensus()
dists_to_consensus = [s.distance(consensus_seq) for s in aln]
average_distance = mean(dists_to_consensus)
std_distance = std(dists_to_consensus)
dist_cutoff = (average_distance + (num_stds * std_distance))
seqs_to_keep = []
for (seq_id, dist_to_consensus) in izip(aln.ids(), dists_to_consensus):
if (dist_to_consensus <= dist_cutoff):
seqs_to_keep.append(seq_id)
filtered_aln = aln.subalignment(seqs_to_keep=seqs_to_keep)
return filtered_aln
| [
"def",
"remove_outliers",
"(",
"seqs",
",",
"num_stds",
",",
"fraction_seqs_for_stats",
"=",
"0.95",
")",
":",
"aln",
"=",
"Alignment",
".",
"from_fasta_records",
"(",
"parse_fasta",
"(",
"seqs",
")",
",",
"DNA",
")",
"consensus_seq",
"=",
"aln",
".",
"majority_consensus",
"(",
")",
"dists_to_consensus",
"=",
"[",
"s",
".",
"distance",
"(",
"consensus_seq",
")",
"for",
"s",
"in",
"aln",
"]",
"average_distance",
"=",
"mean",
"(",
"dists_to_consensus",
")",
"std_distance",
"=",
"std",
"(",
"dists_to_consensus",
")",
"dist_cutoff",
"=",
"(",
"average_distance",
"+",
"(",
"num_stds",
"*",
"std_distance",
")",
")",
"seqs_to_keep",
"=",
"[",
"]",
"for",
"(",
"seq_id",
",",
"dist_to_consensus",
")",
"in",
"izip",
"(",
"aln",
".",
"ids",
"(",
")",
",",
"dists_to_consensus",
")",
":",
"if",
"(",
"dist_to_consensus",
"<=",
"dist_cutoff",
")",
":",
"seqs_to_keep",
".",
"append",
"(",
"seq_id",
")",
"filtered_aln",
"=",
"aln",
".",
"subalignment",
"(",
"seqs_to_keep",
"=",
"seqs_to_keep",
")",
"return",
"filtered_aln"
] | remove sequences very different from the majority consensus given aligned sequences . | train | false |
36,439 | def get_unused_port_and_socket():
s = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
s.bind(('localhost', 0))
(addr, port) = s.getsockname()
return (port, s)
| [
"def",
"get_unused_port_and_socket",
"(",
")",
":",
"s",
"=",
"socket",
".",
"socket",
"(",
"socket",
".",
"AF_INET",
",",
"socket",
".",
"SOCK_STREAM",
")",
"s",
".",
"bind",
"(",
"(",
"'localhost'",
",",
"0",
")",
")",
"(",
"addr",
",",
"port",
")",
"=",
"s",
".",
"getsockname",
"(",
")",
"return",
"(",
"port",
",",
"s",
")"
] | returns an unused port on localhost and the open socket from which it was created . | train | false |
36,440 | def convert_bytes(bytes):
if (bytes is None):
bytes = 0
bytes = float(bytes)
if (bytes >= 1099511627776):
terabytes = (bytes / 1099511627776)
size = ('%.2fTB' % terabytes)
elif (bytes >= 1073741824):
gigabytes = (bytes / 1073741824)
size = ('%.2fGB' % gigabytes)
elif (bytes >= 1048576):
megabytes = (bytes / 1048576)
size = ('%.2fMB' % megabytes)
elif (bytes >= 1024):
kilobytes = (bytes / 1024)
size = ('%.2fKB' % kilobytes)
else:
size = ('%.2fb' % bytes)
return size
| [
"def",
"convert_bytes",
"(",
"bytes",
")",
":",
"if",
"(",
"bytes",
"is",
"None",
")",
":",
"bytes",
"=",
"0",
"bytes",
"=",
"float",
"(",
"bytes",
")",
"if",
"(",
"bytes",
">=",
"1099511627776",
")",
":",
"terabytes",
"=",
"(",
"bytes",
"/",
"1099511627776",
")",
"size",
"=",
"(",
"'%.2fTB'",
"%",
"terabytes",
")",
"elif",
"(",
"bytes",
">=",
"1073741824",
")",
":",
"gigabytes",
"=",
"(",
"bytes",
"/",
"1073741824",
")",
"size",
"=",
"(",
"'%.2fGB'",
"%",
"gigabytes",
")",
"elif",
"(",
"bytes",
">=",
"1048576",
")",
":",
"megabytes",
"=",
"(",
"bytes",
"/",
"1048576",
")",
"size",
"=",
"(",
"'%.2fMB'",
"%",
"megabytes",
")",
"elif",
"(",
"bytes",
">=",
"1024",
")",
":",
"kilobytes",
"=",
"(",
"bytes",
"/",
"1024",
")",
"size",
"=",
"(",
"'%.2fKB'",
"%",
"kilobytes",
")",
"else",
":",
"size",
"=",
"(",
"'%.2fb'",
"%",
"bytes",
")",
"return",
"size"
] | returns given bytes as prettified string . | train | true |
36,441 | def getLoopConvex(polygonComplex):
if (len(polygonComplex) < 4):
return polygonComplex
leftPointIndex = getLeftPointIndex(polygonComplex)
around = (polygonComplex[(leftPointIndex + 1):] + polygonComplex[:(leftPointIndex + 1)])
lastAddedIndex = (-1)
lastPoint = around[(-1)]
aroundLengthMinusOne = (len(around) - 1)
polygonConvex = []
segment = complex(0.0, (-1.0))
while (lastAddedIndex < aroundLengthMinusOne):
lastAddedIndex = getLoopConvexAddedIndex(around, lastAddedIndex, lastPoint, segment)
segment = getNormalized((around[lastAddedIndex] - lastPoint))
lastPoint = around[lastAddedIndex]
polygonConvex.append(lastPoint)
return polygonConvex
| [
"def",
"getLoopConvex",
"(",
"polygonComplex",
")",
":",
"if",
"(",
"len",
"(",
"polygonComplex",
")",
"<",
"4",
")",
":",
"return",
"polygonComplex",
"leftPointIndex",
"=",
"getLeftPointIndex",
"(",
"polygonComplex",
")",
"around",
"=",
"(",
"polygonComplex",
"[",
"(",
"leftPointIndex",
"+",
"1",
")",
":",
"]",
"+",
"polygonComplex",
"[",
":",
"(",
"leftPointIndex",
"+",
"1",
")",
"]",
")",
"lastAddedIndex",
"=",
"(",
"-",
"1",
")",
"lastPoint",
"=",
"around",
"[",
"(",
"-",
"1",
")",
"]",
"aroundLengthMinusOne",
"=",
"(",
"len",
"(",
"around",
")",
"-",
"1",
")",
"polygonConvex",
"=",
"[",
"]",
"segment",
"=",
"complex",
"(",
"0.0",
",",
"(",
"-",
"1.0",
")",
")",
"while",
"(",
"lastAddedIndex",
"<",
"aroundLengthMinusOne",
")",
":",
"lastAddedIndex",
"=",
"getLoopConvexAddedIndex",
"(",
"around",
",",
"lastAddedIndex",
",",
"lastPoint",
",",
"segment",
")",
"segment",
"=",
"getNormalized",
"(",
"(",
"around",
"[",
"lastAddedIndex",
"]",
"-",
"lastPoint",
")",
")",
"lastPoint",
"=",
"around",
"[",
"lastAddedIndex",
"]",
"polygonConvex",
".",
"append",
"(",
"lastPoint",
")",
"return",
"polygonConvex"
] | get convex hull of a complex polygon using gift wrap algorithm . | train | false |
36,444 | def vm_attach(name, kwargs=None, call=None):
if (call != 'action'):
raise SaltCloudSystemExit('The vm_attach action must be called with -a or --action.')
if (kwargs is None):
kwargs = {}
path = kwargs.get('path', None)
data = kwargs.get('data', None)
if data:
if path:
log.warning("Both the 'data' and 'path' arguments were provided. 'data' will take precedence.")
elif path:
data = salt.utils.fopen(path, mode='r').read()
else:
raise SaltCloudSystemExit("The vm_attach function requires either 'data' or a file 'path' to be provided.")
(server, user, password) = _get_xml_rpc()
auth = ':'.join([user, password])
vm_id = int(get_vm_id(kwargs={'name': name}))
response = server.one.vm.attach(auth, vm_id, data)
ret = {'action': 'vm.attach', 'attached': response[0], 'vm_id': response[1], 'error_code': response[2]}
return ret
| [
"def",
"vm_attach",
"(",
"name",
",",
"kwargs",
"=",
"None",
",",
"call",
"=",
"None",
")",
":",
"if",
"(",
"call",
"!=",
"'action'",
")",
":",
"raise",
"SaltCloudSystemExit",
"(",
"'The vm_attach action must be called with -a or --action.'",
")",
"if",
"(",
"kwargs",
"is",
"None",
")",
":",
"kwargs",
"=",
"{",
"}",
"path",
"=",
"kwargs",
".",
"get",
"(",
"'path'",
",",
"None",
")",
"data",
"=",
"kwargs",
".",
"get",
"(",
"'data'",
",",
"None",
")",
"if",
"data",
":",
"if",
"path",
":",
"log",
".",
"warning",
"(",
"\"Both the 'data' and 'path' arguments were provided. 'data' will take precedence.\"",
")",
"elif",
"path",
":",
"data",
"=",
"salt",
".",
"utils",
".",
"fopen",
"(",
"path",
",",
"mode",
"=",
"'r'",
")",
".",
"read",
"(",
")",
"else",
":",
"raise",
"SaltCloudSystemExit",
"(",
"\"The vm_attach function requires either 'data' or a file 'path' to be provided.\"",
")",
"(",
"server",
",",
"user",
",",
"password",
")",
"=",
"_get_xml_rpc",
"(",
")",
"auth",
"=",
"':'",
".",
"join",
"(",
"[",
"user",
",",
"password",
"]",
")",
"vm_id",
"=",
"int",
"(",
"get_vm_id",
"(",
"kwargs",
"=",
"{",
"'name'",
":",
"name",
"}",
")",
")",
"response",
"=",
"server",
".",
"one",
".",
"vm",
".",
"attach",
"(",
"auth",
",",
"vm_id",
",",
"data",
")",
"ret",
"=",
"{",
"'action'",
":",
"'vm.attach'",
",",
"'attached'",
":",
"response",
"[",
"0",
"]",
",",
"'vm_id'",
":",
"response",
"[",
"1",
"]",
",",
"'error_code'",
":",
"response",
"[",
"2",
"]",
"}",
"return",
"ret"
] | attaches a new disk to the given virtual machine . | train | true |
36,445 | def NowUTCToISO8601():
return TimestampUTCToISO8601(time.time())
| [
"def",
"NowUTCToISO8601",
"(",
")",
":",
"return",
"TimestampUTCToISO8601",
"(",
"time",
".",
"time",
"(",
")",
")"
] | return the current date in iso 8601 format: yyyy-mm-dd . | train | false |
36,446 | def _format_unicode_password(pwd):
return '"{0}"'.format(pwd).encode('utf-16-le')
| [
"def",
"_format_unicode_password",
"(",
"pwd",
")",
":",
"return",
"'\"{0}\"'",
".",
"format",
"(",
"pwd",
")",
".",
"encode",
"(",
"'utf-16-le'",
")"
] | formats a string per microsoft ad password specifications . | train | false |
36,447 | def require_certificate_permission(func):
@wraps(func)
def inner(request, *args, **kwargs):
if has_access(request.user, 'certificates', 'global'):
return func(request, *args, **kwargs)
else:
return HttpResponseForbidden()
return inner
| [
"def",
"require_certificate_permission",
"(",
"func",
")",
":",
"@",
"wraps",
"(",
"func",
")",
"def",
"inner",
"(",
"request",
",",
"*",
"args",
",",
"**",
"kwargs",
")",
":",
"if",
"has_access",
"(",
"request",
".",
"user",
",",
"'certificates'",
",",
"'global'",
")",
":",
"return",
"func",
"(",
"request",
",",
"*",
"args",
",",
"**",
"kwargs",
")",
"else",
":",
"return",
"HttpResponseForbidden",
"(",
")",
"return",
"inner"
] | view decorator that requires permission to view and regenerate certificates . | train | false |
36,450 | def prepare_request_uri_query(oauth_params, uri):
(sch, net, path, par, query, fra) = urlparse(uri)
query = urlencode(_append_params(oauth_params, (extract_params(query) or [])))
return urlunparse((sch, net, path, par, query, fra))
| [
"def",
"prepare_request_uri_query",
"(",
"oauth_params",
",",
"uri",
")",
":",
"(",
"sch",
",",
"net",
",",
"path",
",",
"par",
",",
"query",
",",
"fra",
")",
"=",
"urlparse",
"(",
"uri",
")",
"query",
"=",
"urlencode",
"(",
"_append_params",
"(",
"oauth_params",
",",
"(",
"extract_params",
"(",
"query",
")",
"or",
"[",
"]",
")",
")",
")",
"return",
"urlunparse",
"(",
"(",
"sch",
",",
"net",
",",
"path",
",",
"par",
",",
"query",
",",
"fra",
")",
")"
] | prepare the request uri query . | train | true |
36,451 | def count_changes(change):
if isinstance(change, ChangeSet):
result = 0
for child in change.changes:
result += count_changes(child)
return result
return 1
| [
"def",
"count_changes",
"(",
"change",
")",
":",
"if",
"isinstance",
"(",
"change",
",",
"ChangeSet",
")",
":",
"result",
"=",
"0",
"for",
"child",
"in",
"change",
".",
"changes",
":",
"result",
"+=",
"count_changes",
"(",
"child",
")",
"return",
"result",
"return",
"1"
] | counts the number of basic changes a change will make . | train | true |
36,452 | def model_to_dict(instance, fields=None, exclude=None):
from django.db.models.fields.related import ManyToManyField
opts = instance._meta
data = {}
for f in (opts.fields + opts.many_to_many):
if (not f.editable):
continue
if (fields and (not (f.name in fields))):
continue
if (exclude and (f.name in exclude)):
continue
if isinstance(f, ManyToManyField):
if (instance.pk is None):
data[f.name] = []
else:
data[f.name] = list(f.value_from_object(instance).values_list(u'pk', flat=True))
else:
data[f.name] = f.value_from_object(instance)
return data
| [
"def",
"model_to_dict",
"(",
"instance",
",",
"fields",
"=",
"None",
",",
"exclude",
"=",
"None",
")",
":",
"from",
"django",
".",
"db",
".",
"models",
".",
"fields",
".",
"related",
"import",
"ManyToManyField",
"opts",
"=",
"instance",
".",
"_meta",
"data",
"=",
"{",
"}",
"for",
"f",
"in",
"(",
"opts",
".",
"fields",
"+",
"opts",
".",
"many_to_many",
")",
":",
"if",
"(",
"not",
"f",
".",
"editable",
")",
":",
"continue",
"if",
"(",
"fields",
"and",
"(",
"not",
"(",
"f",
".",
"name",
"in",
"fields",
")",
")",
")",
":",
"continue",
"if",
"(",
"exclude",
"and",
"(",
"f",
".",
"name",
"in",
"exclude",
")",
")",
":",
"continue",
"if",
"isinstance",
"(",
"f",
",",
"ManyToManyField",
")",
":",
"if",
"(",
"instance",
".",
"pk",
"is",
"None",
")",
":",
"data",
"[",
"f",
".",
"name",
"]",
"=",
"[",
"]",
"else",
":",
"data",
"[",
"f",
".",
"name",
"]",
"=",
"list",
"(",
"f",
".",
"value_from_object",
"(",
"instance",
")",
".",
"values_list",
"(",
"u'pk'",
",",
"flat",
"=",
"True",
")",
")",
"else",
":",
"data",
"[",
"f",
".",
"name",
"]",
"=",
"f",
".",
"value_from_object",
"(",
"instance",
")",
"return",
"data"
] | returns a dict containing the data in instance suitable for passing as a forms initial keyword argument . | train | false |
36,453 | def render_markup(text):
return Markup(markdown.render(text))
| [
"def",
"render_markup",
"(",
"text",
")",
":",
"return",
"Markup",
"(",
"markdown",
".",
"render",
"(",
"text",
")",
")"
] | renders the given text as markdown . | train | false |
36,454 | def register_email_hook(signal, handler):
assert (signal in (review_request_published, review_request_closed, review_published, reply_published)), (u'Invalid signal %r' % signal)
_hooks[signal].add(handler)
| [
"def",
"register_email_hook",
"(",
"signal",
",",
"handler",
")",
":",
"assert",
"(",
"signal",
"in",
"(",
"review_request_published",
",",
"review_request_closed",
",",
"review_published",
",",
"reply_published",
")",
")",
",",
"(",
"u'Invalid signal %r'",
"%",
"signal",
")",
"_hooks",
"[",
"signal",
"]",
".",
"add",
"(",
"handler",
")"
] | register an e-mail hook . | train | false |
36,455 | @bp.route('/')
def topics():
page = force_int(request.args.get('page', 1), 0)
if (not page):
return abort(404)
paginator = Topic.query.order_by(Topic.updated.desc()).paginate(page)
paginator.items = fill_topics(paginator.items)
return render_template('topic/topics.html', paginator=paginator, endpoint='topic.topics')
| [
"@",
"bp",
".",
"route",
"(",
"'/'",
")",
"def",
"topics",
"(",
")",
":",
"page",
"=",
"force_int",
"(",
"request",
".",
"args",
".",
"get",
"(",
"'page'",
",",
"1",
")",
",",
"0",
")",
"if",
"(",
"not",
"page",
")",
":",
"return",
"abort",
"(",
"404",
")",
"paginator",
"=",
"Topic",
".",
"query",
".",
"order_by",
"(",
"Topic",
".",
"updated",
".",
"desc",
"(",
")",
")",
".",
"paginate",
"(",
"page",
")",
"paginator",
".",
"items",
"=",
"fill_topics",
"(",
"paginator",
".",
"items",
")",
"return",
"render_template",
"(",
"'topic/topics.html'",
",",
"paginator",
"=",
"paginator",
",",
"endpoint",
"=",
"'topic.topics'",
")"
] | the topics list page . | train | false |
36,458 | def single_pre_save_handler(sender, instance, **kwargs):
instance._ct_inventory = None
| [
"def",
"single_pre_save_handler",
"(",
"sender",
",",
"instance",
",",
"**",
"kwargs",
")",
":",
"instance",
".",
"_ct_inventory",
"=",
"None"
] | clobber the _ct_inventory attribute of this object . | train | false |
36,460 | def sanitizeSceneName(name, ezrss=False):
if (not ezrss):
bad_chars = u",:()'!?\u2019"
else:
bad_chars = u",()'?\u2019"
for x in bad_chars:
name = name.replace(x, '')
name = name.replace('- ', '.').replace(' ', '.').replace('&', 'and').replace('/', '.')
name = re.sub('\\.\\.*', '.', name)
if name.endswith('.'):
name = name[:(-1)]
return name
| [
"def",
"sanitizeSceneName",
"(",
"name",
",",
"ezrss",
"=",
"False",
")",
":",
"if",
"(",
"not",
"ezrss",
")",
":",
"bad_chars",
"=",
"u\",:()'!?\\u2019\"",
"else",
":",
"bad_chars",
"=",
"u\",()'?\\u2019\"",
"for",
"x",
"in",
"bad_chars",
":",
"name",
"=",
"name",
".",
"replace",
"(",
"x",
",",
"''",
")",
"name",
"=",
"name",
".",
"replace",
"(",
"'- '",
",",
"'.'",
")",
".",
"replace",
"(",
"' '",
",",
"'.'",
")",
".",
"replace",
"(",
"'&'",
",",
"'and'",
")",
".",
"replace",
"(",
"'/'",
",",
"'.'",
")",
"name",
"=",
"re",
".",
"sub",
"(",
"'\\\\.\\\\.*'",
",",
"'.'",
",",
"name",
")",
"if",
"name",
".",
"endswith",
"(",
"'.'",
")",
":",
"name",
"=",
"name",
"[",
":",
"(",
"-",
"1",
")",
"]",
"return",
"name"
] | takes a show name and returns the "scenified" version of it . | train | false |
36,461 | def iconvergence_agent_tests_factory(fixture):
class IConvergenceAgentTests(TestCase, ):
'\n Tests for ``IConvergenceAgent``.\n '
def test_connected(self):
'\n ``IConvergenceAgent.connected()`` takes an AMP instance.\n '
agent = fixture(self)
agent.connected(connected_amp_protocol())
def test_disconnected(self):
'\n ``IConvergenceAgent.disconnected()`` can be called after\n ``IConvergenceAgent.connected()``.\n '
agent = fixture(self)
agent.connected(connected_amp_protocol())
agent.disconnected()
def test_reconnected(self):
'\n ``IConvergenceAgent.connected()`` can be called after\n ``IConvergenceAgent.disconnected()``.\n '
agent = fixture(self)
agent.connected(connected_amp_protocol())
agent.disconnected()
agent.connected(connected_amp_protocol())
def test_cluster_updated(self):
'\n ``IConvergenceAgent.cluster_updated()`` takes two ``Deployment``\n instances.\n '
agent = fixture(self)
agent.connected(connected_amp_protocol())
agent.cluster_updated(Deployment(nodes=frozenset()), DeploymentState(nodes=[]))
def test_interface(self):
'\n The object provides ``IConvergenceAgent``.\n '
agent = fixture(self)
self.assertTrue(verifyObject(IConvergenceAgent, agent))
return IConvergenceAgentTests
| [
"def",
"iconvergence_agent_tests_factory",
"(",
"fixture",
")",
":",
"class",
"IConvergenceAgentTests",
"(",
"TestCase",
",",
")",
":",
"def",
"test_connected",
"(",
"self",
")",
":",
"agent",
"=",
"fixture",
"(",
"self",
")",
"agent",
".",
"connected",
"(",
"connected_amp_protocol",
"(",
")",
")",
"def",
"test_disconnected",
"(",
"self",
")",
":",
"agent",
"=",
"fixture",
"(",
"self",
")",
"agent",
".",
"connected",
"(",
"connected_amp_protocol",
"(",
")",
")",
"agent",
".",
"disconnected",
"(",
")",
"def",
"test_reconnected",
"(",
"self",
")",
":",
"agent",
"=",
"fixture",
"(",
"self",
")",
"agent",
".",
"connected",
"(",
"connected_amp_protocol",
"(",
")",
")",
"agent",
".",
"disconnected",
"(",
")",
"agent",
".",
"connected",
"(",
"connected_amp_protocol",
"(",
")",
")",
"def",
"test_cluster_updated",
"(",
"self",
")",
":",
"agent",
"=",
"fixture",
"(",
"self",
")",
"agent",
".",
"connected",
"(",
"connected_amp_protocol",
"(",
")",
")",
"agent",
".",
"cluster_updated",
"(",
"Deployment",
"(",
"nodes",
"=",
"frozenset",
"(",
")",
")",
",",
"DeploymentState",
"(",
"nodes",
"=",
"[",
"]",
")",
")",
"def",
"test_interface",
"(",
"self",
")",
":",
"agent",
"=",
"fixture",
"(",
"self",
")",
"self",
".",
"assertTrue",
"(",
"verifyObject",
"(",
"IConvergenceAgent",
",",
"agent",
")",
")",
"return",
"IConvergenceAgentTests"
] | create tests that verify basic iconvergenceagent compliance . | train | false |
36,462 | def cachedir_index_add(minion_id, profile, driver, provider, base=None):
base = init_cachedir(base)
index_file = os.path.join(base, 'index.p')
lock_file(index_file)
if os.path.exists(index_file):
with salt.utils.fopen(index_file, 'r') as fh_:
index = msgpack.load(fh_)
else:
index = {}
prov_comps = provider.split(':')
index.update({minion_id: {'id': minion_id, 'profile': profile, 'driver': driver, 'provider': prov_comps[0]}})
with salt.utils.fopen(index_file, 'w') as fh_:
msgpack.dump(index, fh_)
unlock_file(index_file)
| [
"def",
"cachedir_index_add",
"(",
"minion_id",
",",
"profile",
",",
"driver",
",",
"provider",
",",
"base",
"=",
"None",
")",
":",
"base",
"=",
"init_cachedir",
"(",
"base",
")",
"index_file",
"=",
"os",
".",
"path",
".",
"join",
"(",
"base",
",",
"'index.p'",
")",
"lock_file",
"(",
"index_file",
")",
"if",
"os",
".",
"path",
".",
"exists",
"(",
"index_file",
")",
":",
"with",
"salt",
".",
"utils",
".",
"fopen",
"(",
"index_file",
",",
"'r'",
")",
"as",
"fh_",
":",
"index",
"=",
"msgpack",
".",
"load",
"(",
"fh_",
")",
"else",
":",
"index",
"=",
"{",
"}",
"prov_comps",
"=",
"provider",
".",
"split",
"(",
"':'",
")",
"index",
".",
"update",
"(",
"{",
"minion_id",
":",
"{",
"'id'",
":",
"minion_id",
",",
"'profile'",
":",
"profile",
",",
"'driver'",
":",
"driver",
",",
"'provider'",
":",
"prov_comps",
"[",
"0",
"]",
"}",
"}",
")",
"with",
"salt",
".",
"utils",
".",
"fopen",
"(",
"index_file",
",",
"'w'",
")",
"as",
"fh_",
":",
"msgpack",
".",
"dump",
"(",
"index",
",",
"fh_",
")",
"unlock_file",
"(",
"index_file",
")"
] | add an entry to the cachedir index . | train | false |
36,463 | def unique_file(path, chmod=511, mode='w'):
(path, tail) = os.path.split(path)
return _unique_file(path, filename_pat=(lambda count: ('%04d_%s' % (count, tail))), count=0, chmod=chmod, mode=mode)
| [
"def",
"unique_file",
"(",
"path",
",",
"chmod",
"=",
"511",
",",
"mode",
"=",
"'w'",
")",
":",
"(",
"path",
",",
"tail",
")",
"=",
"os",
".",
"path",
".",
"split",
"(",
"path",
")",
"return",
"_unique_file",
"(",
"path",
",",
"filename_pat",
"=",
"(",
"lambda",
"count",
":",
"(",
"'%04d_%s'",
"%",
"(",
"count",
",",
"tail",
")",
")",
")",
",",
"count",
"=",
"0",
",",
"chmod",
"=",
"chmod",
",",
"mode",
"=",
"mode",
")"
] | safely finds a unique file . | train | false |
36,465 | def correlation_matrix(data):
(_, num_columns) = shape(data)
def matrix_entry(i, j):
return correlation(get_column(data, i), get_column(data, j))
return make_matrix(num_columns, num_columns, matrix_entry)
| [
"def",
"correlation_matrix",
"(",
"data",
")",
":",
"(",
"_",
",",
"num_columns",
")",
"=",
"shape",
"(",
"data",
")",
"def",
"matrix_entry",
"(",
"i",
",",
"j",
")",
":",
"return",
"correlation",
"(",
"get_column",
"(",
"data",
",",
"i",
")",
",",
"get_column",
"(",
"data",
",",
"j",
")",
")",
"return",
"make_matrix",
"(",
"num_columns",
",",
"num_columns",
",",
"matrix_entry",
")"
] | returns the num_columns x num_columns matrix whose th entry is the correlation between columns i and j of data . | train | false |
36,466 | def _SendFileToServer(pool, fullpath, subpath, basename, rebalance):
fp = FileCopyWrapper(rebalance, subpath, basename, fullpath)
try:
headers = {'Content-Length': 0}
res = pool.urlopen('POST', '/rebalance/copy-file', headers=headers, body=fp)
if (res.status != constants.RESPONSE_OK):
return False
except urllib3.exceptions.MaxRetryError:
logging.warning('Failed to send file %s', fullpath)
return False
finally:
fp.close()
return True
| [
"def",
"_SendFileToServer",
"(",
"pool",
",",
"fullpath",
",",
"subpath",
",",
"basename",
",",
"rebalance",
")",
":",
"fp",
"=",
"FileCopyWrapper",
"(",
"rebalance",
",",
"subpath",
",",
"basename",
",",
"fullpath",
")",
"try",
":",
"headers",
"=",
"{",
"'Content-Length'",
":",
"0",
"}",
"res",
"=",
"pool",
".",
"urlopen",
"(",
"'POST'",
",",
"'/rebalance/copy-file'",
",",
"headers",
"=",
"headers",
",",
"body",
"=",
"fp",
")",
"if",
"(",
"res",
".",
"status",
"!=",
"constants",
".",
"RESPONSE_OK",
")",
":",
"return",
"False",
"except",
"urllib3",
".",
"exceptions",
".",
"MaxRetryError",
":",
"logging",
".",
"warning",
"(",
"'Failed to send file %s'",
",",
"fullpath",
")",
"return",
"False",
"finally",
":",
"fp",
".",
"close",
"(",
")",
"return",
"True"
] | sends a specific data store file to the server . | train | false |
36,467 | def _root_task(worker):
return worker._add_task_history[0][0]
| [
"def",
"_root_task",
"(",
"worker",
")",
":",
"return",
"worker",
".",
"_add_task_history",
"[",
"0",
"]",
"[",
"0",
"]"
] | return the first task scheduled by the worker . | train | false |
36,468 | def getNewRepository():
return ExportRepository()
| [
"def",
"getNewRepository",
"(",
")",
":",
"return",
"ExportRepository",
"(",
")"
] | get the repository constructor . | train | false |
36,469 | def update_symlinks(config, unused_plugins):
cert_manager.update_live_symlinks(config)
| [
"def",
"update_symlinks",
"(",
"config",
",",
"unused_plugins",
")",
":",
"cert_manager",
".",
"update_live_symlinks",
"(",
"config",
")"
] | update the certificate file family symlinks use the information in the config file to make symlinks point to the correct archive directory . | train | false |
36,470 | def test_setup_headshape():
for hs in hs_fnames:
(dig, t) = _process_bti_headshape(hs)
expected = set(['kind', 'ident', 'r'])
found = set(reduce((lambda x, y: (list(x) + list(y))), [d.keys() for d in dig]))
assert_true((not (expected - found)))
| [
"def",
"test_setup_headshape",
"(",
")",
":",
"for",
"hs",
"in",
"hs_fnames",
":",
"(",
"dig",
",",
"t",
")",
"=",
"_process_bti_headshape",
"(",
"hs",
")",
"expected",
"=",
"set",
"(",
"[",
"'kind'",
",",
"'ident'",
",",
"'r'",
"]",
")",
"found",
"=",
"set",
"(",
"reduce",
"(",
"(",
"lambda",
"x",
",",
"y",
":",
"(",
"list",
"(",
"x",
")",
"+",
"list",
"(",
"y",
")",
")",
")",
",",
"[",
"d",
".",
"keys",
"(",
")",
"for",
"d",
"in",
"dig",
"]",
")",
")",
"assert_true",
"(",
"(",
"not",
"(",
"expected",
"-",
"found",
")",
")",
")"
] | test reading bti headshape . | train | false |
36,472 | def load_kvm():
kvm_arch = get_kvm_arch()
def load_module(mod='kvm'):
return utils.system(('modprobe %s' % mod))
loaded = load_module()
if (not loaded):
loaded = load_module(mod=kvm_arch)
return loaded
| [
"def",
"load_kvm",
"(",
")",
":",
"kvm_arch",
"=",
"get_kvm_arch",
"(",
")",
"def",
"load_module",
"(",
"mod",
"=",
"'kvm'",
")",
":",
"return",
"utils",
".",
"system",
"(",
"(",
"'modprobe %s'",
"%",
"mod",
")",
")",
"loaded",
"=",
"load_module",
"(",
")",
"if",
"(",
"not",
"loaded",
")",
":",
"loaded",
"=",
"load_module",
"(",
"mod",
"=",
"kvm_arch",
")",
"return",
"loaded"
] | loads the appropriate kvm kernel modules depending on the current cpu architecture :returns: 0 on success or 1 on failure :rtype: int . | train | false |
36,473 | def encrypt_int(message, ekey, n):
if (type(message) is types.IntType):
message = long(message)
if (not (type(message) is types.LongType)):
raise TypeError('You must pass a long or int')
if ((message < 0) or (message > n)):
raise OverflowError('The message is too long')
safebit = (bit_size(n) - 2)
message += (1 << safebit)
return pow(message, ekey, n)
| [
"def",
"encrypt_int",
"(",
"message",
",",
"ekey",
",",
"n",
")",
":",
"if",
"(",
"type",
"(",
"message",
")",
"is",
"types",
".",
"IntType",
")",
":",
"message",
"=",
"long",
"(",
"message",
")",
"if",
"(",
"not",
"(",
"type",
"(",
"message",
")",
"is",
"types",
".",
"LongType",
")",
")",
":",
"raise",
"TypeError",
"(",
"'You must pass a long or int'",
")",
"if",
"(",
"(",
"message",
"<",
"0",
")",
"or",
"(",
"message",
">",
"n",
")",
")",
":",
"raise",
"OverflowError",
"(",
"'The message is too long'",
")",
"safebit",
"=",
"(",
"bit_size",
"(",
"n",
")",
"-",
"2",
")",
"message",
"+=",
"(",
"1",
"<<",
"safebit",
")",
"return",
"pow",
"(",
"message",
",",
"ekey",
",",
"n",
")"
] | encrypts a message using encryption key ekey . | train | false |
36,474 | def random_mpd(distmat, n, iters):
means = []
indices = arange(distmat.shape[0])
for i in range(iters):
shuffle(indices)
means.append(mpd(reduce_mtx(distmat, indices[:n])))
return (mean(means), std(means))
| [
"def",
"random_mpd",
"(",
"distmat",
",",
"n",
",",
"iters",
")",
":",
"means",
"=",
"[",
"]",
"indices",
"=",
"arange",
"(",
"distmat",
".",
"shape",
"[",
"0",
"]",
")",
"for",
"i",
"in",
"range",
"(",
"iters",
")",
":",
"shuffle",
"(",
"indices",
")",
"means",
".",
"append",
"(",
"mpd",
"(",
"reduce_mtx",
"(",
"distmat",
",",
"indices",
"[",
":",
"n",
"]",
")",
")",
")",
"return",
"(",
"mean",
"(",
"means",
")",
",",
"std",
"(",
"means",
")",
")"
] | calc mean . | train | false |
36,475 | def set_namespace(namespace):
if (namespace is None):
os.environ.pop(_ENV_CURRENT_NAMESPACE, None)
else:
validate_namespace(namespace)
os.environ[_ENV_CURRENT_NAMESPACE] = namespace
| [
"def",
"set_namespace",
"(",
"namespace",
")",
":",
"if",
"(",
"namespace",
"is",
"None",
")",
":",
"os",
".",
"environ",
".",
"pop",
"(",
"_ENV_CURRENT_NAMESPACE",
",",
"None",
")",
"else",
":",
"validate_namespace",
"(",
"namespace",
")",
"os",
".",
"environ",
"[",
"_ENV_CURRENT_NAMESPACE",
"]",
"=",
"namespace"
] | set the default namespace for the current http request . | train | false |
36,476 | @core_helper
def list_dict_filter(list_, search_field, output_field, value):
for item in list_:
if (item.get(search_field) == value):
return item.get(output_field, value)
return value
| [
"@",
"core_helper",
"def",
"list_dict_filter",
"(",
"list_",
",",
"search_field",
",",
"output_field",
",",
"value",
")",
":",
"for",
"item",
"in",
"list_",
":",
"if",
"(",
"item",
".",
"get",
"(",
"search_field",
")",
"==",
"value",
")",
":",
"return",
"item",
".",
"get",
"(",
"output_field",
",",
"value",
")",
"return",
"value"
] | takes a list of dicts and returns the value of a given key if the item has a matching value for a supplied key . | train | false |
36,477 | def get_access_token(scopes, service_account_id=None):
memcache_key = (_MEMCACHE_KEY_PREFIX + str(scopes))
if service_account_id:
memcache_key += (',%d' % service_account_id)
memcache_value = memcache.get(memcache_key, namespace=_MEMCACHE_NAMESPACE)
if memcache_value:
(access_token, expires_at) = memcache_value
else:
(access_token, expires_at) = get_access_token_uncached(scopes, service_account_id=service_account_id)
memcache.add(memcache_key, (access_token, expires_at), (expires_at - 300), namespace=_MEMCACHE_NAMESPACE)
return (access_token, expires_at)
| [
"def",
"get_access_token",
"(",
"scopes",
",",
"service_account_id",
"=",
"None",
")",
":",
"memcache_key",
"=",
"(",
"_MEMCACHE_KEY_PREFIX",
"+",
"str",
"(",
"scopes",
")",
")",
"if",
"service_account_id",
":",
"memcache_key",
"+=",
"(",
"',%d'",
"%",
"service_account_id",
")",
"memcache_value",
"=",
"memcache",
".",
"get",
"(",
"memcache_key",
",",
"namespace",
"=",
"_MEMCACHE_NAMESPACE",
")",
"if",
"memcache_value",
":",
"(",
"access_token",
",",
"expires_at",
")",
"=",
"memcache_value",
"else",
":",
"(",
"access_token",
",",
"expires_at",
")",
"=",
"get_access_token_uncached",
"(",
"scopes",
",",
"service_account_id",
"=",
"service_account_id",
")",
"memcache",
".",
"add",
"(",
"memcache_key",
",",
"(",
"access_token",
",",
"expires_at",
")",
",",
"(",
"expires_at",
"-",
"300",
")",
",",
"namespace",
"=",
"_MEMCACHE_NAMESPACE",
")",
"return",
"(",
"access_token",
",",
"expires_at",
")"
] | get the access token using the provided credentials arguments: username : a string containing the username to log in password : a string containing the password for the username returns: str: authentication token . | train | false |
36,478 | def depend_check(deps_name, *args):
missing = []
for dependency in args:
try:
__import__(dependency)
except ImportError:
missing.append(dependency)
print(('-' * 80))
if missing:
print((red('You are missing the following %s dependencies:') % deps_name))
for dep in missing:
name = pkg_info_dict.get(dep, dep)
print(' * ', name)
print()
else:
print((blue('All %s dependencies installed! You are good to go!\n') % deps_name))
| [
"def",
"depend_check",
"(",
"deps_name",
",",
"*",
"args",
")",
":",
"missing",
"=",
"[",
"]",
"for",
"dependency",
"in",
"args",
":",
"try",
":",
"__import__",
"(",
"dependency",
")",
"except",
"ImportError",
":",
"missing",
".",
"append",
"(",
"dependency",
")",
"print",
"(",
"(",
"'-'",
"*",
"80",
")",
")",
"if",
"missing",
":",
"print",
"(",
"(",
"red",
"(",
"'You are missing the following %s dependencies:'",
")",
"%",
"deps_name",
")",
")",
"for",
"dep",
"in",
"missing",
":",
"name",
"=",
"pkg_info_dict",
".",
"get",
"(",
"dep",
",",
"dep",
")",
"print",
"(",
"' * '",
",",
"name",
")",
"print",
"(",
")",
"else",
":",
"print",
"(",
"(",
"blue",
"(",
"'All %s dependencies installed! You are good to go!\\n'",
")",
"%",
"deps_name",
")",
")"
] | make sure a given dependency is installed . | train | false |
36,479 | def rewrite_local_config(mobsf_home):
CONFIG['MobSF']['subdir_tools'] = (mobsf_home + '\\StaticAnalyzer\\tools\\windows\\')
CONFIG['MobSF']['dir'] = mobsf_home
with open(os.path.join(CONFIG_PATH, CONFIG_FILE), 'w') as configfile:
CONFIG.write(configfile)
| [
"def",
"rewrite_local_config",
"(",
"mobsf_home",
")",
":",
"CONFIG",
"[",
"'MobSF'",
"]",
"[",
"'subdir_tools'",
"]",
"=",
"(",
"mobsf_home",
"+",
"'\\\\StaticAnalyzer\\\\tools\\\\windows\\\\'",
")",
"CONFIG",
"[",
"'MobSF'",
"]",
"[",
"'dir'",
"]",
"=",
"mobsf_home",
"with",
"open",
"(",
"os",
".",
"path",
".",
"join",
"(",
"CONFIG_PATH",
",",
"CONFIG_FILE",
")",
",",
"'w'",
")",
"as",
"configfile",
":",
"CONFIG",
".",
"write",
"(",
"configfile",
")"
] | for local installation some config-vars need to be rewritten . | train | false |
36,480 | def _screen_docrevise(docstr):
import re
if (docstr is None):
return None
screenname = _CFG['examplescreen']
newdocstr = docstr.replace(('%s.' % screenname), '')
parexp = re.compile((' \\(.+ %s\\):' % screenname))
newdocstr = parexp.sub(':', newdocstr)
return newdocstr
| [
"def",
"_screen_docrevise",
"(",
"docstr",
")",
":",
"import",
"re",
"if",
"(",
"docstr",
"is",
"None",
")",
":",
"return",
"None",
"screenname",
"=",
"_CFG",
"[",
"'examplescreen'",
"]",
"newdocstr",
"=",
"docstr",
".",
"replace",
"(",
"(",
"'%s.'",
"%",
"screenname",
")",
",",
"''",
")",
"parexp",
"=",
"re",
".",
"compile",
"(",
"(",
"' \\\\(.+ %s\\\\):'",
"%",
"screenname",
")",
")",
"newdocstr",
"=",
"parexp",
".",
"sub",
"(",
"':'",
",",
"newdocstr",
")",
"return",
"newdocstr"
] | to reduce docstrings from turtlescreen class for functions . | train | false |
36,482 | def is_bar(bar_containers, **props):
for container in bar_containers:
if (props['mplobj'] in container):
return True
return False
| [
"def",
"is_bar",
"(",
"bar_containers",
",",
"**",
"props",
")",
":",
"for",
"container",
"in",
"bar_containers",
":",
"if",
"(",
"props",
"[",
"'mplobj'",
"]",
"in",
"container",
")",
":",
"return",
"True",
"return",
"False"
] | a test to decide whether a path is a bar from a vertical bar chart . | train | false |
36,483 | def _find_process_name(port_number):
for connection in psutil.net_connections():
if (connection.laddr[1] == port_number):
return psutil.Process(connection.pid).name()
return None
| [
"def",
"_find_process_name",
"(",
"port_number",
")",
":",
"for",
"connection",
"in",
"psutil",
".",
"net_connections",
"(",
")",
":",
"if",
"(",
"connection",
".",
"laddr",
"[",
"1",
"]",
"==",
"port_number",
")",
":",
"return",
"psutil",
".",
"Process",
"(",
"connection",
".",
"pid",
")",
".",
"name",
"(",
")",
"return",
"None"
] | get the name of the process using the given port number . | train | false |
36,484 | def _contextkey(jail=None, chroot=None, root=None, prefix='pkg.list_pkgs'):
if jail:
return (str(prefix) + '.jail_{0}'.format(jail))
elif chroot:
return (str(prefix) + '.chroot_{0}'.format(chroot))
elif root:
return (str(prefix) + '.root_{0}'.format(root))
return prefix
| [
"def",
"_contextkey",
"(",
"jail",
"=",
"None",
",",
"chroot",
"=",
"None",
",",
"root",
"=",
"None",
",",
"prefix",
"=",
"'pkg.list_pkgs'",
")",
":",
"if",
"jail",
":",
"return",
"(",
"str",
"(",
"prefix",
")",
"+",
"'.jail_{0}'",
".",
"format",
"(",
"jail",
")",
")",
"elif",
"chroot",
":",
"return",
"(",
"str",
"(",
"prefix",
")",
"+",
"'.chroot_{0}'",
".",
"format",
"(",
"chroot",
")",
")",
"elif",
"root",
":",
"return",
"(",
"str",
"(",
"prefix",
")",
"+",
"'.root_{0}'",
".",
"format",
"(",
"root",
")",
")",
"return",
"prefix"
] | as this module is designed to manipulate packages in jails and chroots . | train | false |
36,485 | def docker_build_publish(parse, xml_parent, data):
db = XML.SubElement(xml_parent, 'com.cloudbees.dockerpublish.DockerBuilder')
db.set('plugin', 'docker-build-publish')
mapping = [('repo-name', 'repoName', None), ('repo-tag', 'repoTag', ''), ('no-cache', 'noCache', False), ('no-force-pull', 'noForcePull', False), ('skip-build', 'skipBuild', False), ('skip-decorate', 'skipDecorate', False), ('skip-tag-latest', 'skipTagLatest', False), ('skip-push', 'skipPush', False), ('file-path', 'dockerfilePath', ''), ('build-context', 'buildContext', '')]
convert_mapping_to_xml(db, data, mapping, fail_required=True)
| [
"def",
"docker_build_publish",
"(",
"parse",
",",
"xml_parent",
",",
"data",
")",
":",
"db",
"=",
"XML",
".",
"SubElement",
"(",
"xml_parent",
",",
"'com.cloudbees.dockerpublish.DockerBuilder'",
")",
"db",
".",
"set",
"(",
"'plugin'",
",",
"'docker-build-publish'",
")",
"mapping",
"=",
"[",
"(",
"'repo-name'",
",",
"'repoName'",
",",
"None",
")",
",",
"(",
"'repo-tag'",
",",
"'repoTag'",
",",
"''",
")",
",",
"(",
"'no-cache'",
",",
"'noCache'",
",",
"False",
")",
",",
"(",
"'no-force-pull'",
",",
"'noForcePull'",
",",
"False",
")",
",",
"(",
"'skip-build'",
",",
"'skipBuild'",
",",
"False",
")",
",",
"(",
"'skip-decorate'",
",",
"'skipDecorate'",
",",
"False",
")",
",",
"(",
"'skip-tag-latest'",
",",
"'skipTagLatest'",
",",
"False",
")",
",",
"(",
"'skip-push'",
",",
"'skipPush'",
",",
"False",
")",
",",
"(",
"'file-path'",
",",
"'dockerfilePath'",
",",
"''",
")",
",",
"(",
"'build-context'",
",",
"'buildContext'",
",",
"''",
")",
"]",
"convert_mapping_to_xml",
"(",
"db",
",",
"data",
",",
"mapping",
",",
"fail_required",
"=",
"True",
")"
] | yaml: docker-build-publish requires the jenkins :jenkins-wikidocker build publish plugin <docker+build+publish+plugin> . | train | false |
36,486 | def autoimport():
word = env.var('a:word')
if (not word):
env.error('Should be word under cursor.')
return False
with RopeContext() as ctx:
if (not ctx.importer.names):
ctx.generate_autoimport_cache()
modules = ctx.importer.get_modules(word)
if (not modules):
env.message(('Global name %s not found.' % word))
return False
if (len(modules) == 1):
_insert_import(word, modules[0], ctx)
else:
module = env.user_input_choices('Which module to import:', *modules)
_insert_import(word, module, ctx)
return True
| [
"def",
"autoimport",
"(",
")",
":",
"word",
"=",
"env",
".",
"var",
"(",
"'a:word'",
")",
"if",
"(",
"not",
"word",
")",
":",
"env",
".",
"error",
"(",
"'Should be word under cursor.'",
")",
"return",
"False",
"with",
"RopeContext",
"(",
")",
"as",
"ctx",
":",
"if",
"(",
"not",
"ctx",
".",
"importer",
".",
"names",
")",
":",
"ctx",
".",
"generate_autoimport_cache",
"(",
")",
"modules",
"=",
"ctx",
".",
"importer",
".",
"get_modules",
"(",
"word",
")",
"if",
"(",
"not",
"modules",
")",
":",
"env",
".",
"message",
"(",
"(",
"'Global name %s not found.'",
"%",
"word",
")",
")",
"return",
"False",
"if",
"(",
"len",
"(",
"modules",
")",
"==",
"1",
")",
":",
"_insert_import",
"(",
"word",
",",
"modules",
"[",
"0",
"]",
",",
"ctx",
")",
"else",
":",
"module",
"=",
"env",
".",
"user_input_choices",
"(",
"'Which module to import:'",
",",
"*",
"modules",
")",
"_insert_import",
"(",
"word",
",",
"module",
",",
"ctx",
")",
"return",
"True"
] | autoimport modules . | train | false |
36,488 | def _CreateCommentsFromPrefix(comment_prefix, comment_lineno, comment_column, standalone=False):
comments = []
lines = comment_prefix.split('\n')
index = 0
while (index < len(lines)):
comment_block = []
while ((index < len(lines)) and lines[index].lstrip().startswith('#')):
comment_block.append(lines[index])
index += 1
if comment_block:
new_lineno = ((comment_lineno + index) - 1)
comment_block[0] = comment_block[0].lstrip()
comment_block[(-1)] = comment_block[(-1)].rstrip('\n')
comment_leaf = pytree.Leaf(type=token.COMMENT, value='\n'.join(comment_block), context=('', (new_lineno, comment_column)))
comment_node = (comment_leaf if (not standalone) else pytree.Node(pygram.python_symbols.simple_stmt, [comment_leaf]))
comments.append(comment_node)
while ((index < len(lines)) and (not lines[index].lstrip())):
index += 1
return comments
| [
"def",
"_CreateCommentsFromPrefix",
"(",
"comment_prefix",
",",
"comment_lineno",
",",
"comment_column",
",",
"standalone",
"=",
"False",
")",
":",
"comments",
"=",
"[",
"]",
"lines",
"=",
"comment_prefix",
".",
"split",
"(",
"'\\n'",
")",
"index",
"=",
"0",
"while",
"(",
"index",
"<",
"len",
"(",
"lines",
")",
")",
":",
"comment_block",
"=",
"[",
"]",
"while",
"(",
"(",
"index",
"<",
"len",
"(",
"lines",
")",
")",
"and",
"lines",
"[",
"index",
"]",
".",
"lstrip",
"(",
")",
".",
"startswith",
"(",
"'#'",
")",
")",
":",
"comment_block",
".",
"append",
"(",
"lines",
"[",
"index",
"]",
")",
"index",
"+=",
"1",
"if",
"comment_block",
":",
"new_lineno",
"=",
"(",
"(",
"comment_lineno",
"+",
"index",
")",
"-",
"1",
")",
"comment_block",
"[",
"0",
"]",
"=",
"comment_block",
"[",
"0",
"]",
".",
"lstrip",
"(",
")",
"comment_block",
"[",
"(",
"-",
"1",
")",
"]",
"=",
"comment_block",
"[",
"(",
"-",
"1",
")",
"]",
".",
"rstrip",
"(",
"'\\n'",
")",
"comment_leaf",
"=",
"pytree",
".",
"Leaf",
"(",
"type",
"=",
"token",
".",
"COMMENT",
",",
"value",
"=",
"'\\n'",
".",
"join",
"(",
"comment_block",
")",
",",
"context",
"=",
"(",
"''",
",",
"(",
"new_lineno",
",",
"comment_column",
")",
")",
")",
"comment_node",
"=",
"(",
"comment_leaf",
"if",
"(",
"not",
"standalone",
")",
"else",
"pytree",
".",
"Node",
"(",
"pygram",
".",
"python_symbols",
".",
"simple_stmt",
",",
"[",
"comment_leaf",
"]",
")",
")",
"comments",
".",
"append",
"(",
"comment_node",
")",
"while",
"(",
"(",
"index",
"<",
"len",
"(",
"lines",
")",
")",
"and",
"(",
"not",
"lines",
"[",
"index",
"]",
".",
"lstrip",
"(",
")",
")",
")",
":",
"index",
"+=",
"1",
"return",
"comments"
] | create pytree nodes to represent the given comment prefix . | train | false |
36,489 | def print_top_words_per_topic(data, top_n=10, do_print=True):
msgs = []
lists = []
for (j, topic_to_word) in enumerate(data['topic_term_dists']):
top = np.argsort(topic_to_word)[::(-1)][:top_n]
prefix = ('Top words in topic %i ' % j)
top_words = [data['vocab'][i].strip().replace(' ', '_') for i in top]
msg = ' '.join(top_words)
if do_print:
print (prefix + msg)
lists.append(top_words)
return lists
| [
"def",
"print_top_words_per_topic",
"(",
"data",
",",
"top_n",
"=",
"10",
",",
"do_print",
"=",
"True",
")",
":",
"msgs",
"=",
"[",
"]",
"lists",
"=",
"[",
"]",
"for",
"(",
"j",
",",
"topic_to_word",
")",
"in",
"enumerate",
"(",
"data",
"[",
"'topic_term_dists'",
"]",
")",
":",
"top",
"=",
"np",
".",
"argsort",
"(",
"topic_to_word",
")",
"[",
":",
":",
"(",
"-",
"1",
")",
"]",
"[",
":",
"top_n",
"]",
"prefix",
"=",
"(",
"'Top words in topic %i '",
"%",
"j",
")",
"top_words",
"=",
"[",
"data",
"[",
"'vocab'",
"]",
"[",
"i",
"]",
".",
"strip",
"(",
")",
".",
"replace",
"(",
"' '",
",",
"'_'",
")",
"for",
"i",
"in",
"top",
"]",
"msg",
"=",
"' '",
".",
"join",
"(",
"top_words",
")",
"if",
"do_print",
":",
"print",
"(",
"prefix",
"+",
"msg",
")",
"lists",
".",
"append",
"(",
"top_words",
")",
"return",
"lists"
] | given a pyldavis data array . | train | false |
36,491 | def create_binding(site, hostheader='', ipaddress='*', port=80, protocol='http', sslflags=0):
pscmd = list()
protocol = str(protocol).lower()
sslflags = int(sslflags)
name = _get_binding_info(hostheader, ipaddress, port)
if (protocol not in _VALID_PROTOCOLS):
message = "Invalid protocol '{0}' specified. Valid formats: {1}".format(protocol, _VALID_PROTOCOLS)
raise SaltInvocationError(message)
if (sslflags not in _VALID_SSL_FLAGS):
message = "Invalid sslflags '{0}' specified. Valid sslflags range: {1}..{2}".format(sslflags, _VALID_SSL_FLAGS[0], _VALID_SSL_FLAGS[(-1)])
raise SaltInvocationError(message)
current_bindings = list_bindings(site)
if (name in current_bindings):
_LOG.debug('Binding already present: %s', name)
return True
pscmd.append("New-WebBinding -Name '{0}' -HostHeader '{1}'".format(site, hostheader))
pscmd.append(" -IpAddress '{0}' -Port '{1}'".format(ipaddress, port))
pscmd.append(" -Protocol '{0}' -SslFlags {1}".format(protocol, sslflags))
cmd_ret = _srvmgr(str().join(pscmd))
if (cmd_ret['retcode'] == 0):
new_bindings = list_bindings(site)
if (name in new_bindings):
_LOG.debug('Binding created successfully: %s', name)
return True
_LOG.error('Unable to create binding: %s', name)
return False
| [
"def",
"create_binding",
"(",
"site",
",",
"hostheader",
"=",
"''",
",",
"ipaddress",
"=",
"'*'",
",",
"port",
"=",
"80",
",",
"protocol",
"=",
"'http'",
",",
"sslflags",
"=",
"0",
")",
":",
"pscmd",
"=",
"list",
"(",
")",
"protocol",
"=",
"str",
"(",
"protocol",
")",
".",
"lower",
"(",
")",
"sslflags",
"=",
"int",
"(",
"sslflags",
")",
"name",
"=",
"_get_binding_info",
"(",
"hostheader",
",",
"ipaddress",
",",
"port",
")",
"if",
"(",
"protocol",
"not",
"in",
"_VALID_PROTOCOLS",
")",
":",
"message",
"=",
"\"Invalid protocol '{0}' specified. Valid formats: {1}\"",
".",
"format",
"(",
"protocol",
",",
"_VALID_PROTOCOLS",
")",
"raise",
"SaltInvocationError",
"(",
"message",
")",
"if",
"(",
"sslflags",
"not",
"in",
"_VALID_SSL_FLAGS",
")",
":",
"message",
"=",
"\"Invalid sslflags '{0}' specified. Valid sslflags range: {1}..{2}\"",
".",
"format",
"(",
"sslflags",
",",
"_VALID_SSL_FLAGS",
"[",
"0",
"]",
",",
"_VALID_SSL_FLAGS",
"[",
"(",
"-",
"1",
")",
"]",
")",
"raise",
"SaltInvocationError",
"(",
"message",
")",
"current_bindings",
"=",
"list_bindings",
"(",
"site",
")",
"if",
"(",
"name",
"in",
"current_bindings",
")",
":",
"_LOG",
".",
"debug",
"(",
"'Binding already present: %s'",
",",
"name",
")",
"return",
"True",
"pscmd",
".",
"append",
"(",
"\"New-WebBinding -Name '{0}' -HostHeader '{1}'\"",
".",
"format",
"(",
"site",
",",
"hostheader",
")",
")",
"pscmd",
".",
"append",
"(",
"\" -IpAddress '{0}' -Port '{1}'\"",
".",
"format",
"(",
"ipaddress",
",",
"port",
")",
")",
"pscmd",
".",
"append",
"(",
"\" -Protocol '{0}' -SslFlags {1}\"",
".",
"format",
"(",
"protocol",
",",
"sslflags",
")",
")",
"cmd_ret",
"=",
"_srvmgr",
"(",
"str",
"(",
")",
".",
"join",
"(",
"pscmd",
")",
")",
"if",
"(",
"cmd_ret",
"[",
"'retcode'",
"]",
"==",
"0",
")",
":",
"new_bindings",
"=",
"list_bindings",
"(",
"site",
")",
"if",
"(",
"name",
"in",
"new_bindings",
")",
":",
"_LOG",
".",
"debug",
"(",
"'Binding created successfully: %s'",
",",
"name",
")",
"return",
"True",
"_LOG",
".",
"error",
"(",
"'Unable to create binding: %s'",
",",
"name",
")",
"return",
"False"
] | create an iis binding . | train | false |
36,492 | def dev_from_pcapname(pcap_name):
return IFACES.dev_from_pcapname(pcap_name)
| [
"def",
"dev_from_pcapname",
"(",
"pcap_name",
")",
":",
"return",
"IFACES",
".",
"dev_from_pcapname",
"(",
"pcap_name",
")"
] | return libdnet/scapy device name for given pypcap device name . | train | false |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.