id_within_dataset
int64 1
55.5k
| snippet
stringlengths 19
14.2k
| tokens
listlengths 6
1.63k
| nl
stringlengths 6
352
| split_within_dataset
stringclasses 1
value | is_duplicated
bool 2
classes |
|---|---|---|---|---|---|
48,303
|
def vm_running(name):
name = name.lower()
ret = {'name': name, 'changes': {}, 'result': None, 'comment': ''}
if (name in __salt__['vmadm.list'](order='hostname', search='state=running')):
ret['result'] = True
ret['comment'] = 'vm {0} already running'.format(name)
else:
ret['result'] = (True if __opts__['test'] else __salt__['vmadm.start'](name, key='hostname'))
if ((not isinstance(ret['result'], bool)) and ret['result'].get('Error')):
ret['result'] = False
ret['comment'] = 'failed to start {0}'.format(name)
else:
ret['changes'][name] = 'running'
ret['comment'] = 'vm {0} started'.format(name)
return ret
|
[
"def",
"vm_running",
"(",
"name",
")",
":",
"name",
"=",
"name",
".",
"lower",
"(",
")",
"ret",
"=",
"{",
"'name'",
":",
"name",
",",
"'changes'",
":",
"{",
"}",
",",
"'result'",
":",
"None",
",",
"'comment'",
":",
"''",
"}",
"if",
"(",
"name",
"in",
"__salt__",
"[",
"'vmadm.list'",
"]",
"(",
"order",
"=",
"'hostname'",
",",
"search",
"=",
"'state=running'",
")",
")",
":",
"ret",
"[",
"'result'",
"]",
"=",
"True",
"ret",
"[",
"'comment'",
"]",
"=",
"'vm {0} already running'",
".",
"format",
"(",
"name",
")",
"else",
":",
"ret",
"[",
"'result'",
"]",
"=",
"(",
"True",
"if",
"__opts__",
"[",
"'test'",
"]",
"else",
"__salt__",
"[",
"'vmadm.start'",
"]",
"(",
"name",
",",
"key",
"=",
"'hostname'",
")",
")",
"if",
"(",
"(",
"not",
"isinstance",
"(",
"ret",
"[",
"'result'",
"]",
",",
"bool",
")",
")",
"and",
"ret",
"[",
"'result'",
"]",
".",
"get",
"(",
"'Error'",
")",
")",
":",
"ret",
"[",
"'result'",
"]",
"=",
"False",
"ret",
"[",
"'comment'",
"]",
"=",
"'failed to start {0}'",
".",
"format",
"(",
"name",
")",
"else",
":",
"ret",
"[",
"'changes'",
"]",
"[",
"name",
"]",
"=",
"'running'",
"ret",
"[",
"'comment'",
"]",
"=",
"'vm {0} started'",
".",
"format",
"(",
"name",
")",
"return",
"ret"
] |
ensure vm is in the running state on the computenode name : string hostname of vm .
|
train
| true
|
48,304
|
def test_sg_filter_trivial():
x = np.array([1.0])
y = savgol_filter(x, 1, 0)
assert_equal(y, [1.0])
x = np.array([3.0])
y = savgol_filter(x, 3, 1, mode='constant')
assert_almost_equal(y, [1.0], decimal=15)
x = np.array([3.0])
y = savgol_filter(x, 3, 1, mode='nearest')
assert_almost_equal(y, [3.0], decimal=15)
x = np.array(([1.0] * 3))
y = savgol_filter(x, 3, 1, mode='wrap')
assert_almost_equal(y, [1.0, 1.0, 1.0], decimal=15)
|
[
"def",
"test_sg_filter_trivial",
"(",
")",
":",
"x",
"=",
"np",
".",
"array",
"(",
"[",
"1.0",
"]",
")",
"y",
"=",
"savgol_filter",
"(",
"x",
",",
"1",
",",
"0",
")",
"assert_equal",
"(",
"y",
",",
"[",
"1.0",
"]",
")",
"x",
"=",
"np",
".",
"array",
"(",
"[",
"3.0",
"]",
")",
"y",
"=",
"savgol_filter",
"(",
"x",
",",
"3",
",",
"1",
",",
"mode",
"=",
"'constant'",
")",
"assert_almost_equal",
"(",
"y",
",",
"[",
"1.0",
"]",
",",
"decimal",
"=",
"15",
")",
"x",
"=",
"np",
".",
"array",
"(",
"[",
"3.0",
"]",
")",
"y",
"=",
"savgol_filter",
"(",
"x",
",",
"3",
",",
"1",
",",
"mode",
"=",
"'nearest'",
")",
"assert_almost_equal",
"(",
"y",
",",
"[",
"3.0",
"]",
",",
"decimal",
"=",
"15",
")",
"x",
"=",
"np",
".",
"array",
"(",
"(",
"[",
"1.0",
"]",
"*",
"3",
")",
")",
"y",
"=",
"savgol_filter",
"(",
"x",
",",
"3",
",",
"1",
",",
"mode",
"=",
"'wrap'",
")",
"assert_almost_equal",
"(",
"y",
",",
"[",
"1.0",
",",
"1.0",
",",
"1.0",
"]",
",",
"decimal",
"=",
"15",
")"
] |
test some trivial edge cases for savgol_filter() .
|
train
| false
|
48,307
|
def faces(G):
f = set()
for (v1, edges) in G.items():
for v2 in edges:
for v3 in G[v2]:
if (v1 == v3):
continue
if (v1 in G[v3]):
f.add(frozenset([v1, v2, v3]))
else:
for v4 in G[v3]:
if (v4 == v2):
continue
if (v1 in G[v4]):
f.add(frozenset([v1, v2, v3, v4]))
else:
for v5 in G[v4]:
if ((v5 == v3) or (v5 == v2)):
continue
if (v1 in G[v5]):
f.add(frozenset([v1, v2, v3, v4, v5]))
return f
|
[
"def",
"faces",
"(",
"G",
")",
":",
"f",
"=",
"set",
"(",
")",
"for",
"(",
"v1",
",",
"edges",
")",
"in",
"G",
".",
"items",
"(",
")",
":",
"for",
"v2",
"in",
"edges",
":",
"for",
"v3",
"in",
"G",
"[",
"v2",
"]",
":",
"if",
"(",
"v1",
"==",
"v3",
")",
":",
"continue",
"if",
"(",
"v1",
"in",
"G",
"[",
"v3",
"]",
")",
":",
"f",
".",
"add",
"(",
"frozenset",
"(",
"[",
"v1",
",",
"v2",
",",
"v3",
"]",
")",
")",
"else",
":",
"for",
"v4",
"in",
"G",
"[",
"v3",
"]",
":",
"if",
"(",
"v4",
"==",
"v2",
")",
":",
"continue",
"if",
"(",
"v1",
"in",
"G",
"[",
"v4",
"]",
")",
":",
"f",
".",
"add",
"(",
"frozenset",
"(",
"[",
"v1",
",",
"v2",
",",
"v3",
",",
"v4",
"]",
")",
")",
"else",
":",
"for",
"v5",
"in",
"G",
"[",
"v4",
"]",
":",
"if",
"(",
"(",
"v5",
"==",
"v3",
")",
"or",
"(",
"v5",
"==",
"v2",
")",
")",
":",
"continue",
"if",
"(",
"v1",
"in",
"G",
"[",
"v5",
"]",
")",
":",
"f",
".",
"add",
"(",
"frozenset",
"(",
"[",
"v1",
",",
"v2",
",",
"v3",
",",
"v4",
",",
"v5",
"]",
")",
")",
"return",
"f"
] |
return a set of faces in g .
|
train
| false
|
48,309
|
def jitclass(spec):
def wrap(cls):
if config.DISABLE_JIT:
return cls
else:
return register_class_type(cls, spec, types.ClassType, ClassBuilder)
return wrap
|
[
"def",
"jitclass",
"(",
"spec",
")",
":",
"def",
"wrap",
"(",
"cls",
")",
":",
"if",
"config",
".",
"DISABLE_JIT",
":",
"return",
"cls",
"else",
":",
"return",
"register_class_type",
"(",
"cls",
",",
"spec",
",",
"types",
".",
"ClassType",
",",
"ClassBuilder",
")",
"return",
"wrap"
] |
a decorator for creating a jitclass .
|
train
| false
|
48,310
|
def run_in_subprocess_with_hash_randomization(function, function_args=(), function_kwargs={}, command=sys.executable, module='sympy.utilities.runtests', force=False):
p = subprocess.Popen([command, '-RV'], stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
p.communicate()
if (p.returncode != 0):
return False
hash_seed = os.getenv('PYTHONHASHSEED')
if (not hash_seed):
os.environ['PYTHONHASHSEED'] = str(random.randrange((2 ** 32)))
elif (not force):
return False
commandstring = ('import sys; from %s import %s;sys.exit(%s(*%s, **%s))' % (module, function, function, repr(function_args), repr(function_kwargs)))
try:
p = subprocess.Popen([command, '-R', '-c', commandstring])
p.communicate()
except KeyboardInterrupt:
p.wait()
finally:
if (hash_seed is None):
del os.environ['PYTHONHASHSEED']
else:
os.environ['PYTHONHASHSEED'] = hash_seed
return p.returncode
|
[
"def",
"run_in_subprocess_with_hash_randomization",
"(",
"function",
",",
"function_args",
"=",
"(",
")",
",",
"function_kwargs",
"=",
"{",
"}",
",",
"command",
"=",
"sys",
".",
"executable",
",",
"module",
"=",
"'sympy.utilities.runtests'",
",",
"force",
"=",
"False",
")",
":",
"p",
"=",
"subprocess",
".",
"Popen",
"(",
"[",
"command",
",",
"'-RV'",
"]",
",",
"stdout",
"=",
"subprocess",
".",
"PIPE",
",",
"stderr",
"=",
"subprocess",
".",
"STDOUT",
")",
"p",
".",
"communicate",
"(",
")",
"if",
"(",
"p",
".",
"returncode",
"!=",
"0",
")",
":",
"return",
"False",
"hash_seed",
"=",
"os",
".",
"getenv",
"(",
"'PYTHONHASHSEED'",
")",
"if",
"(",
"not",
"hash_seed",
")",
":",
"os",
".",
"environ",
"[",
"'PYTHONHASHSEED'",
"]",
"=",
"str",
"(",
"random",
".",
"randrange",
"(",
"(",
"2",
"**",
"32",
")",
")",
")",
"elif",
"(",
"not",
"force",
")",
":",
"return",
"False",
"commandstring",
"=",
"(",
"'import sys; from %s import %s;sys.exit(%s(*%s, **%s))'",
"%",
"(",
"module",
",",
"function",
",",
"function",
",",
"repr",
"(",
"function_args",
")",
",",
"repr",
"(",
"function_kwargs",
")",
")",
")",
"try",
":",
"p",
"=",
"subprocess",
".",
"Popen",
"(",
"[",
"command",
",",
"'-R'",
",",
"'-c'",
",",
"commandstring",
"]",
")",
"p",
".",
"communicate",
"(",
")",
"except",
"KeyboardInterrupt",
":",
"p",
".",
"wait",
"(",
")",
"finally",
":",
"if",
"(",
"hash_seed",
"is",
"None",
")",
":",
"del",
"os",
".",
"environ",
"[",
"'PYTHONHASHSEED'",
"]",
"else",
":",
"os",
".",
"environ",
"[",
"'PYTHONHASHSEED'",
"]",
"=",
"hash_seed",
"return",
"p",
".",
"returncode"
] |
run a function in a python subprocess with hash randomization enabled .
|
train
| false
|
48,311
|
def failhard(role):
raise FileserverConfigError('Failed to load {0}'.format(role))
|
[
"def",
"failhard",
"(",
"role",
")",
":",
"raise",
"FileserverConfigError",
"(",
"'Failed to load {0}'",
".",
"format",
"(",
"role",
")",
")"
] |
fatal configuration issue .
|
train
| false
|
48,312
|
def email_notification(version, build, email):
log.debug(LOG_TEMPLATE.format(project=version.project.slug, version=version.slug, msg=('sending email to: %s' % email)))
context = {'version': version, 'project': version.project, 'build': build, 'build_url': 'https://{0}{1}'.format(getattr(settings, 'PRODUCTION_DOMAIN', 'readthedocs.org'), build.get_absolute_url()), 'unsub_url': 'https://{0}{1}'.format(getattr(settings, 'PRODUCTION_DOMAIN', 'readthedocs.org'), reverse('projects_notifications', args=[version.project.slug]))}
if build.commit:
title = _('Failed: {project.name} ({commit})').format(commit=build.commit[:8], **context)
else:
title = _('Failed: {project.name} ({version.verbose_name})').format(**context)
send_email(email, title, template='projects/email/build_failed.txt', template_html='projects/email/build_failed.html', context=context)
|
[
"def",
"email_notification",
"(",
"version",
",",
"build",
",",
"email",
")",
":",
"log",
".",
"debug",
"(",
"LOG_TEMPLATE",
".",
"format",
"(",
"project",
"=",
"version",
".",
"project",
".",
"slug",
",",
"version",
"=",
"version",
".",
"slug",
",",
"msg",
"=",
"(",
"'sending email to: %s'",
"%",
"email",
")",
")",
")",
"context",
"=",
"{",
"'version'",
":",
"version",
",",
"'project'",
":",
"version",
".",
"project",
",",
"'build'",
":",
"build",
",",
"'build_url'",
":",
"'https://{0}{1}'",
".",
"format",
"(",
"getattr",
"(",
"settings",
",",
"'PRODUCTION_DOMAIN'",
",",
"'readthedocs.org'",
")",
",",
"build",
".",
"get_absolute_url",
"(",
")",
")",
",",
"'unsub_url'",
":",
"'https://{0}{1}'",
".",
"format",
"(",
"getattr",
"(",
"settings",
",",
"'PRODUCTION_DOMAIN'",
",",
"'readthedocs.org'",
")",
",",
"reverse",
"(",
"'projects_notifications'",
",",
"args",
"=",
"[",
"version",
".",
"project",
".",
"slug",
"]",
")",
")",
"}",
"if",
"build",
".",
"commit",
":",
"title",
"=",
"_",
"(",
"'Failed: {project.name} ({commit})'",
")",
".",
"format",
"(",
"commit",
"=",
"build",
".",
"commit",
"[",
":",
"8",
"]",
",",
"**",
"context",
")",
"else",
":",
"title",
"=",
"_",
"(",
"'Failed: {project.name} ({version.verbose_name})'",
")",
".",
"format",
"(",
"**",
"context",
")",
"send_email",
"(",
"email",
",",
"title",
",",
"template",
"=",
"'projects/email/build_failed.txt'",
",",
"template_html",
"=",
"'projects/email/build_failed.html'",
",",
"context",
"=",
"context",
")"
] |
send email notifications for build failure .
|
train
| false
|
48,313
|
def head_pos_to_trans_rot_t(quats):
t = quats[..., 0].copy()
rotation = quat_to_rot(quats[..., 1:4])
translation = quats[..., 4:7].copy()
return (translation, rotation, t)
|
[
"def",
"head_pos_to_trans_rot_t",
"(",
"quats",
")",
":",
"t",
"=",
"quats",
"[",
"...",
",",
"0",
"]",
".",
"copy",
"(",
")",
"rotation",
"=",
"quat_to_rot",
"(",
"quats",
"[",
"...",
",",
"1",
":",
"4",
"]",
")",
"translation",
"=",
"quats",
"[",
"...",
",",
"4",
":",
"7",
"]",
".",
"copy",
"(",
")",
"return",
"(",
"translation",
",",
"rotation",
",",
"t",
")"
] |
convert maxfilter-formatted head position quaternions .
|
train
| false
|
48,314
|
@utils.arg('server', metavar='<server>', help=_('Name or ID of server.'))
@utils.arg('--console-type', default='serial', help=_('Type of serial console, default="serial".'))
def do_get_serial_console(cs, args):
if (args.console_type not in ('serial',)):
raise exceptions.CommandError(_("Invalid parameter value for 'console_type', currently supported 'serial'."))
server = _find_server(cs, args.server)
data = server.get_serial_console(args.console_type)
print_console(cs, data)
|
[
"@",
"utils",
".",
"arg",
"(",
"'server'",
",",
"metavar",
"=",
"'<server>'",
",",
"help",
"=",
"_",
"(",
"'Name or ID of server.'",
")",
")",
"@",
"utils",
".",
"arg",
"(",
"'--console-type'",
",",
"default",
"=",
"'serial'",
",",
"help",
"=",
"_",
"(",
"'Type of serial console, default=\"serial\".'",
")",
")",
"def",
"do_get_serial_console",
"(",
"cs",
",",
"args",
")",
":",
"if",
"(",
"args",
".",
"console_type",
"not",
"in",
"(",
"'serial'",
",",
")",
")",
":",
"raise",
"exceptions",
".",
"CommandError",
"(",
"_",
"(",
"\"Invalid parameter value for 'console_type', currently supported 'serial'.\"",
")",
")",
"server",
"=",
"_find_server",
"(",
"cs",
",",
"args",
".",
"server",
")",
"data",
"=",
"server",
".",
"get_serial_console",
"(",
"args",
".",
"console_type",
")",
"print_console",
"(",
"cs",
",",
"data",
")"
] |
get a serial console to a server .
|
train
| false
|
48,316
|
@utils.arg('--fields', default=None, metavar='<fields>', help=_('Comma-separated list of fields to display. Use the show command to see which fields are available.'))
@deprecated_network
def do_network_list(cs, args):
network_list = cs.networks.list()
columns = ['ID', 'Label', 'Cidr']
columns += _get_list_table_columns_and_formatters(args.fields, network_list, exclude_fields=(c.lower() for c in columns))[0]
utils.print_list(network_list, columns)
|
[
"@",
"utils",
".",
"arg",
"(",
"'--fields'",
",",
"default",
"=",
"None",
",",
"metavar",
"=",
"'<fields>'",
",",
"help",
"=",
"_",
"(",
"'Comma-separated list of fields to display. Use the show command to see which fields are available.'",
")",
")",
"@",
"deprecated_network",
"def",
"do_network_list",
"(",
"cs",
",",
"args",
")",
":",
"network_list",
"=",
"cs",
".",
"networks",
".",
"list",
"(",
")",
"columns",
"=",
"[",
"'ID'",
",",
"'Label'",
",",
"'Cidr'",
"]",
"columns",
"+=",
"_get_list_table_columns_and_formatters",
"(",
"args",
".",
"fields",
",",
"network_list",
",",
"exclude_fields",
"=",
"(",
"c",
".",
"lower",
"(",
")",
"for",
"c",
"in",
"columns",
")",
")",
"[",
"0",
"]",
"utils",
".",
"print_list",
"(",
"network_list",
",",
"columns",
")"
] |
print a list of available networks .
|
train
| false
|
48,317
|
def tdecode(data, key, decode=base64.b64decode, salt_length=16):
if decode:
data = decode(data)
salt = data[:salt_length]
return crypt(data[salt_length:], sha1((key + salt)).digest())
|
[
"def",
"tdecode",
"(",
"data",
",",
"key",
",",
"decode",
"=",
"base64",
".",
"b64decode",
",",
"salt_length",
"=",
"16",
")",
":",
"if",
"decode",
":",
"data",
"=",
"decode",
"(",
"data",
")",
"salt",
"=",
"data",
"[",
":",
"salt_length",
"]",
"return",
"crypt",
"(",
"data",
"[",
"salt_length",
":",
"]",
",",
"sha1",
"(",
"(",
"key",
"+",
"salt",
")",
")",
".",
"digest",
"(",
")",
")"
] |
rc4 decryption of encoded data .
|
train
| false
|
48,318
|
def randomPolicy(Ts):
numA = len(Ts)
dim = len(Ts[0])
return ((ones((dim, numA)) / float(numA)), mean(array(Ts), axis=0))
|
[
"def",
"randomPolicy",
"(",
"Ts",
")",
":",
"numA",
"=",
"len",
"(",
"Ts",
")",
"dim",
"=",
"len",
"(",
"Ts",
"[",
"0",
"]",
")",
"return",
"(",
"(",
"ones",
"(",
"(",
"dim",
",",
"numA",
")",
")",
"/",
"float",
"(",
"numA",
")",
")",
",",
"mean",
"(",
"array",
"(",
"Ts",
")",
",",
"axis",
"=",
"0",
")",
")"
] |
each action is equally likely .
|
train
| false
|
48,319
|
def raise_invalid(request, location='body', name=None, description=None, **kwargs):
request.errors.add(location, name, description, **kwargs)
response = json_error_handler(request)
raise response
|
[
"def",
"raise_invalid",
"(",
"request",
",",
"location",
"=",
"'body'",
",",
"name",
"=",
"None",
",",
"description",
"=",
"None",
",",
"**",
"kwargs",
")",
":",
"request",
".",
"errors",
".",
"add",
"(",
"location",
",",
"name",
",",
"description",
",",
"**",
"kwargs",
")",
"response",
"=",
"json_error_handler",
"(",
"request",
")",
"raise",
"response"
] |
helper to raise a validation error .
|
train
| false
|
48,320
|
@rule(u'.*')
@priority(u'low')
def collectlines(bot, trigger):
if trigger.is_privmsg:
return
if (trigger.sender not in bot.memory[u'find_lines']):
bot.memory[u'find_lines'][trigger.sender] = SopelMemory()
if (Identifier(trigger.nick) not in bot.memory[u'find_lines'][trigger.sender]):
bot.memory[u'find_lines'][trigger.sender][Identifier(trigger.nick)] = list()
templist = bot.memory[u'find_lines'][trigger.sender][Identifier(trigger.nick)]
line = trigger.group()
if line.startswith(u's/'):
return
elif line.startswith(u'\x01ACTION'):
line = line[:(-1)]
templist.append(line)
else:
templist.append(line)
del templist[:(-10)]
bot.memory[u'find_lines'][trigger.sender][Identifier(trigger.nick)] = templist
|
[
"@",
"rule",
"(",
"u'.*'",
")",
"@",
"priority",
"(",
"u'low'",
")",
"def",
"collectlines",
"(",
"bot",
",",
"trigger",
")",
":",
"if",
"trigger",
".",
"is_privmsg",
":",
"return",
"if",
"(",
"trigger",
".",
"sender",
"not",
"in",
"bot",
".",
"memory",
"[",
"u'find_lines'",
"]",
")",
":",
"bot",
".",
"memory",
"[",
"u'find_lines'",
"]",
"[",
"trigger",
".",
"sender",
"]",
"=",
"SopelMemory",
"(",
")",
"if",
"(",
"Identifier",
"(",
"trigger",
".",
"nick",
")",
"not",
"in",
"bot",
".",
"memory",
"[",
"u'find_lines'",
"]",
"[",
"trigger",
".",
"sender",
"]",
")",
":",
"bot",
".",
"memory",
"[",
"u'find_lines'",
"]",
"[",
"trigger",
".",
"sender",
"]",
"[",
"Identifier",
"(",
"trigger",
".",
"nick",
")",
"]",
"=",
"list",
"(",
")",
"templist",
"=",
"bot",
".",
"memory",
"[",
"u'find_lines'",
"]",
"[",
"trigger",
".",
"sender",
"]",
"[",
"Identifier",
"(",
"trigger",
".",
"nick",
")",
"]",
"line",
"=",
"trigger",
".",
"group",
"(",
")",
"if",
"line",
".",
"startswith",
"(",
"u's/'",
")",
":",
"return",
"elif",
"line",
".",
"startswith",
"(",
"u'\\x01ACTION'",
")",
":",
"line",
"=",
"line",
"[",
":",
"(",
"-",
"1",
")",
"]",
"templist",
".",
"append",
"(",
"line",
")",
"else",
":",
"templist",
".",
"append",
"(",
"line",
")",
"del",
"templist",
"[",
":",
"(",
"-",
"10",
")",
"]",
"bot",
".",
"memory",
"[",
"u'find_lines'",
"]",
"[",
"trigger",
".",
"sender",
"]",
"[",
"Identifier",
"(",
"trigger",
".",
"nick",
")",
"]",
"=",
"templist"
] |
create a temporary log of what people say .
|
train
| false
|
48,321
|
def parseLineReplace(firstWordTable, line, output):
firstWord = gcodec.getFirstWordFromLine(line)
if (firstWord in firstWordTable):
line = firstWordTable[firstWord]
gcodec.addLineAndNewlineIfNecessary(line, output)
|
[
"def",
"parseLineReplace",
"(",
"firstWordTable",
",",
"line",
",",
"output",
")",
":",
"firstWord",
"=",
"gcodec",
".",
"getFirstWordFromLine",
"(",
"line",
")",
"if",
"(",
"firstWord",
"in",
"firstWordTable",
")",
":",
"line",
"=",
"firstWordTable",
"[",
"firstWord",
"]",
"gcodec",
".",
"addLineAndNewlineIfNecessary",
"(",
"line",
",",
"output",
")"
] |
parse the line and replace it if the first word of the line is in the first word table .
|
train
| false
|
48,322
|
def _parse_tformat(tform):
try:
(repeat, format, option) = TFORMAT_RE.match(tform.strip()).groups()
except Exception:
raise VerifyError('Format {!r} is not recognized.'.format(tform))
if (repeat == ''):
repeat = 1
else:
repeat = int(repeat)
return (repeat, format.upper(), option)
|
[
"def",
"_parse_tformat",
"(",
"tform",
")",
":",
"try",
":",
"(",
"repeat",
",",
"format",
",",
"option",
")",
"=",
"TFORMAT_RE",
".",
"match",
"(",
"tform",
".",
"strip",
"(",
")",
")",
".",
"groups",
"(",
")",
"except",
"Exception",
":",
"raise",
"VerifyError",
"(",
"'Format {!r} is not recognized.'",
".",
"format",
"(",
"tform",
")",
")",
"if",
"(",
"repeat",
"==",
"''",
")",
":",
"repeat",
"=",
"1",
"else",
":",
"repeat",
"=",
"int",
"(",
"repeat",
")",
"return",
"(",
"repeat",
",",
"format",
".",
"upper",
"(",
")",
",",
"option",
")"
] |
parse tformn keyword for a binary table into a tuple .
|
train
| false
|
48,323
|
def _retrieve_ntp_peers():
return __salt__['ntp.peers']()
|
[
"def",
"_retrieve_ntp_peers",
"(",
")",
":",
"return",
"__salt__",
"[",
"'ntp.peers'",
"]",
"(",
")"
] |
retrieves configured ntp peers .
|
train
| false
|
48,324
|
def author(name):
return ('from:%s' % name)
|
[
"def",
"author",
"(",
"name",
")",
":",
"return",
"(",
"'from:%s'",
"%",
"name",
")"
] |
returns a twitter query-by-author-name that can be passed to twitter .
|
train
| false
|
48,326
|
def _generate_sample_indices(random_state, n_samples):
random_instance = check_random_state(random_state)
sample_indices = random_instance.randint(0, n_samples, n_samples)
return sample_indices
|
[
"def",
"_generate_sample_indices",
"(",
"random_state",
",",
"n_samples",
")",
":",
"random_instance",
"=",
"check_random_state",
"(",
"random_state",
")",
"sample_indices",
"=",
"random_instance",
".",
"randint",
"(",
"0",
",",
"n_samples",
",",
"n_samples",
")",
"return",
"sample_indices"
] |
private function used to _parallel_build_trees function .
|
train
| false
|
48,327
|
def uidFromString(uidString):
try:
return int(uidString)
except ValueError:
if (pwd is None):
raise
return pwd.getpwnam(uidString)[2]
|
[
"def",
"uidFromString",
"(",
"uidString",
")",
":",
"try",
":",
"return",
"int",
"(",
"uidString",
")",
"except",
"ValueError",
":",
"if",
"(",
"pwd",
"is",
"None",
")",
":",
"raise",
"return",
"pwd",
".",
"getpwnam",
"(",
"uidString",
")",
"[",
"2",
"]"
] |
convert a user identifier .
|
train
| false
|
48,328
|
def has_scope(context=None):
if (not booted(context)):
return False
_sd_version = version(context)
if (_sd_version is None):
return False
return (_sd_version >= 205)
|
[
"def",
"has_scope",
"(",
"context",
"=",
"None",
")",
":",
"if",
"(",
"not",
"booted",
"(",
"context",
")",
")",
":",
"return",
"False",
"_sd_version",
"=",
"version",
"(",
"context",
")",
"if",
"(",
"_sd_version",
"is",
"None",
")",
":",
"return",
"False",
"return",
"(",
"_sd_version",
">=",
"205",
")"
] |
scopes were introduced in systemd 205 .
|
train
| true
|
48,330
|
def test_api_fixture(hug_api):
assert isinstance(hug_api, hug.API)
assert (hug_api != api)
|
[
"def",
"test_api_fixture",
"(",
"hug_api",
")",
":",
"assert",
"isinstance",
"(",
"hug_api",
",",
"hug",
".",
"API",
")",
"assert",
"(",
"hug_api",
"!=",
"api",
")"
] |
ensure its possible to dynamically insert a new hug api on demand .
|
train
| false
|
48,332
|
def assert_instance_of(expected, actual, msg=None):
assert isinstance(actual, expected), msg
|
[
"def",
"assert_instance_of",
"(",
"expected",
",",
"actual",
",",
"msg",
"=",
"None",
")",
":",
"assert",
"isinstance",
"(",
"actual",
",",
"expected",
")",
",",
"msg"
] |
verify that object is an instance of expected .
|
train
| false
|
48,333
|
def _create_transformations(all_tokens, seen_ts):
for (parent, token) in all_tokens:
if isinstance(token, TransformationToken):
if (token.number not in seen_ts):
raise RuntimeError(('Tabstop %i is not known but is used by a Transformation' % token.number))
Transformation(parent, seen_ts[token.number], token)
|
[
"def",
"_create_transformations",
"(",
"all_tokens",
",",
"seen_ts",
")",
":",
"for",
"(",
"parent",
",",
"token",
")",
"in",
"all_tokens",
":",
"if",
"isinstance",
"(",
"token",
",",
"TransformationToken",
")",
":",
"if",
"(",
"token",
".",
"number",
"not",
"in",
"seen_ts",
")",
":",
"raise",
"RuntimeError",
"(",
"(",
"'Tabstop %i is not known but is used by a Transformation'",
"%",
"token",
".",
"number",
")",
")",
"Transformation",
"(",
"parent",
",",
"seen_ts",
"[",
"token",
".",
"number",
"]",
",",
"token",
")"
] |
create the objects that need to know about tabstops .
|
train
| false
|
48,334
|
def java_library(name, srcs=[], deps=[], resources=[], source_encoding=None, warnings=None, prebuilt=False, binary_jar='', exported_deps=[], provided_deps=[], **kwargs):
target = JavaLibrary(name, srcs, deps, resources, source_encoding, warnings, prebuilt, binary_jar, exported_deps, provided_deps, kwargs)
blade.blade.register_target(target)
|
[
"def",
"java_library",
"(",
"name",
",",
"srcs",
"=",
"[",
"]",
",",
"deps",
"=",
"[",
"]",
",",
"resources",
"=",
"[",
"]",
",",
"source_encoding",
"=",
"None",
",",
"warnings",
"=",
"None",
",",
"prebuilt",
"=",
"False",
",",
"binary_jar",
"=",
"''",
",",
"exported_deps",
"=",
"[",
"]",
",",
"provided_deps",
"=",
"[",
"]",
",",
"**",
"kwargs",
")",
":",
"target",
"=",
"JavaLibrary",
"(",
"name",
",",
"srcs",
",",
"deps",
",",
"resources",
",",
"source_encoding",
",",
"warnings",
",",
"prebuilt",
",",
"binary_jar",
",",
"exported_deps",
",",
"provided_deps",
",",
"kwargs",
")",
"blade",
".",
"blade",
".",
"register_target",
"(",
"target",
")"
] |
define java_library target .
|
train
| false
|
48,335
|
def trim_dir(directory):
def access_time(f):
return os.stat(os.path.join(directory, f)).st_atime
files = sorted(os.listdir(directory), key=access_time)
file_name = os.path.join(directory, files[0])
log.debug(u'removing least accessed file: %s', file_name)
os.remove(file_name)
|
[
"def",
"trim_dir",
"(",
"directory",
")",
":",
"def",
"access_time",
"(",
"f",
")",
":",
"return",
"os",
".",
"stat",
"(",
"os",
".",
"path",
".",
"join",
"(",
"directory",
",",
"f",
")",
")",
".",
"st_atime",
"files",
"=",
"sorted",
"(",
"os",
".",
"listdir",
"(",
"directory",
")",
",",
"key",
"=",
"access_time",
")",
"file_name",
"=",
"os",
".",
"path",
".",
"join",
"(",
"directory",
",",
"files",
"[",
"0",
"]",
")",
"log",
".",
"debug",
"(",
"u'removing least accessed file: %s'",
",",
"file_name",
")",
"os",
".",
"remove",
"(",
"file_name",
")"
] |
removed the least accessed file on a given dir .
|
train
| false
|
48,336
|
def find_asgs(conn, module, name=None, tags=None):
try:
asgs_paginator = conn.get_paginator('describe_auto_scaling_groups')
asgs = asgs_paginator.paginate().build_full_result()
except ClientError as e:
module.fail_json(msg=e.message, **camel_dict_to_snake_dict(e.response))
matched_asgs = []
if (name is not None):
name_prog = re.compile(('^' + name))
for asg in asgs['AutoScalingGroups']:
if name:
matched_name = name_prog.search(asg['AutoScalingGroupName'])
else:
matched_name = True
if tags:
matched_tags = match_asg_tags(tags, asg)
else:
matched_tags = True
if (matched_name and matched_tags):
matched_asgs.append(camel_dict_to_snake_dict(asg))
return matched_asgs
|
[
"def",
"find_asgs",
"(",
"conn",
",",
"module",
",",
"name",
"=",
"None",
",",
"tags",
"=",
"None",
")",
":",
"try",
":",
"asgs_paginator",
"=",
"conn",
".",
"get_paginator",
"(",
"'describe_auto_scaling_groups'",
")",
"asgs",
"=",
"asgs_paginator",
".",
"paginate",
"(",
")",
".",
"build_full_result",
"(",
")",
"except",
"ClientError",
"as",
"e",
":",
"module",
".",
"fail_json",
"(",
"msg",
"=",
"e",
".",
"message",
",",
"**",
"camel_dict_to_snake_dict",
"(",
"e",
".",
"response",
")",
")",
"matched_asgs",
"=",
"[",
"]",
"if",
"(",
"name",
"is",
"not",
"None",
")",
":",
"name_prog",
"=",
"re",
".",
"compile",
"(",
"(",
"'^'",
"+",
"name",
")",
")",
"for",
"asg",
"in",
"asgs",
"[",
"'AutoScalingGroups'",
"]",
":",
"if",
"name",
":",
"matched_name",
"=",
"name_prog",
".",
"search",
"(",
"asg",
"[",
"'AutoScalingGroupName'",
"]",
")",
"else",
":",
"matched_name",
"=",
"True",
"if",
"tags",
":",
"matched_tags",
"=",
"match_asg_tags",
"(",
"tags",
",",
"asg",
")",
"else",
":",
"matched_tags",
"=",
"True",
"if",
"(",
"matched_name",
"and",
"matched_tags",
")",
":",
"matched_asgs",
".",
"append",
"(",
"camel_dict_to_snake_dict",
"(",
"asg",
")",
")",
"return",
"matched_asgs"
] |
args: conn : valid boto3 asg client .
|
train
| false
|
48,337
|
def libvlc_media_player_pause(p_mi):
f = (_Cfunctions.get('libvlc_media_player_pause', None) or _Cfunction('libvlc_media_player_pause', ((1,),), None, None, MediaPlayer))
return f(p_mi)
|
[
"def",
"libvlc_media_player_pause",
"(",
"p_mi",
")",
":",
"f",
"=",
"(",
"_Cfunctions",
".",
"get",
"(",
"'libvlc_media_player_pause'",
",",
"None",
")",
"or",
"_Cfunction",
"(",
"'libvlc_media_player_pause'",
",",
"(",
"(",
"1",
",",
")",
",",
")",
",",
"None",
",",
"None",
",",
"MediaPlayer",
")",
")",
"return",
"f",
"(",
"p_mi",
")"
] |
toggle pause .
|
train
| false
|
48,338
|
def center_and_norm(x, axis=(-1)):
x = np.rollaxis(x, axis)
x -= x.mean(axis=0)
x /= x.std(axis=0)
|
[
"def",
"center_and_norm",
"(",
"x",
",",
"axis",
"=",
"(",
"-",
"1",
")",
")",
":",
"x",
"=",
"np",
".",
"rollaxis",
"(",
"x",
",",
"axis",
")",
"x",
"-=",
"x",
".",
"mean",
"(",
"axis",
"=",
"0",
")",
"x",
"/=",
"x",
".",
"std",
"(",
"axis",
"=",
"0",
")"
] |
centers and norms x **in place** parameters x: ndarray array with an axis of observations measured on random variables .
|
train
| false
|
48,340
|
def htmlsafe_dumps(obj, **kwargs):
rv = dumps(obj, **kwargs).replace(u'<', u'\\u003c').replace(u'>', u'\\u003e').replace(u'&', u'\\u0026').replace(u"'", u'\\u0027')
if (not _slash_escape):
rv = rv.replace('\\/', '/')
return rv
|
[
"def",
"htmlsafe_dumps",
"(",
"obj",
",",
"**",
"kwargs",
")",
":",
"rv",
"=",
"dumps",
"(",
"obj",
",",
"**",
"kwargs",
")",
".",
"replace",
"(",
"u'<'",
",",
"u'\\\\u003c'",
")",
".",
"replace",
"(",
"u'>'",
",",
"u'\\\\u003e'",
")",
".",
"replace",
"(",
"u'&'",
",",
"u'\\\\u0026'",
")",
".",
"replace",
"(",
"u\"'\"",
",",
"u'\\\\u0027'",
")",
"if",
"(",
"not",
"_slash_escape",
")",
":",
"rv",
"=",
"rv",
".",
"replace",
"(",
"'\\\\/'",
",",
"'/'",
")",
"return",
"rv"
] |
works exactly like :func:dumps but is safe for use in <script> tags .
|
train
| true
|
48,341
|
def get_dict(*keys, **extras):
_keys = ('url', 'args', 'form', 'data', 'origin', 'headers', 'files', 'json')
assert all(map(_keys.__contains__, keys))
data = request.data
form = semiflatten(request.form)
try:
_json = json.loads(data.decode('utf-8'))
except (ValueError, TypeError):
_json = None
d = dict(url=get_url(request), args=semiflatten(request.args), form=form, data=json_safe(data), origin=request.headers.get('X-Forwarded-For', request.remote_addr), headers=get_headers(), files=get_files(), json=_json)
out_d = dict()
for key in keys:
out_d[key] = d.get(key)
out_d.update(extras)
return out_d
|
[
"def",
"get_dict",
"(",
"*",
"keys",
",",
"**",
"extras",
")",
":",
"_keys",
"=",
"(",
"'url'",
",",
"'args'",
",",
"'form'",
",",
"'data'",
",",
"'origin'",
",",
"'headers'",
",",
"'files'",
",",
"'json'",
")",
"assert",
"all",
"(",
"map",
"(",
"_keys",
".",
"__contains__",
",",
"keys",
")",
")",
"data",
"=",
"request",
".",
"data",
"form",
"=",
"semiflatten",
"(",
"request",
".",
"form",
")",
"try",
":",
"_json",
"=",
"json",
".",
"loads",
"(",
"data",
".",
"decode",
"(",
"'utf-8'",
")",
")",
"except",
"(",
"ValueError",
",",
"TypeError",
")",
":",
"_json",
"=",
"None",
"d",
"=",
"dict",
"(",
"url",
"=",
"get_url",
"(",
"request",
")",
",",
"args",
"=",
"semiflatten",
"(",
"request",
".",
"args",
")",
",",
"form",
"=",
"form",
",",
"data",
"=",
"json_safe",
"(",
"data",
")",
",",
"origin",
"=",
"request",
".",
"headers",
".",
"get",
"(",
"'X-Forwarded-For'",
",",
"request",
".",
"remote_addr",
")",
",",
"headers",
"=",
"get_headers",
"(",
")",
",",
"files",
"=",
"get_files",
"(",
")",
",",
"json",
"=",
"_json",
")",
"out_d",
"=",
"dict",
"(",
")",
"for",
"key",
"in",
"keys",
":",
"out_d",
"[",
"key",
"]",
"=",
"d",
".",
"get",
"(",
"key",
")",
"out_d",
".",
"update",
"(",
"extras",
")",
"return",
"out_d"
] |
returns request dict of given keys .
|
train
| true
|
48,342
|
def process_result(method, result):
if (method.post_process_func is not None):
result = method.post_process_func(result)
if method.boolean:
if (result in [1, '1']):
result = True
elif (result in [0, '0']):
result = False
return result
|
[
"def",
"process_result",
"(",
"method",
",",
"result",
")",
":",
"if",
"(",
"method",
".",
"post_process_func",
"is",
"not",
"None",
")",
":",
"result",
"=",
"method",
".",
"post_process_func",
"(",
"result",
")",
"if",
"method",
".",
"boolean",
":",
"if",
"(",
"result",
"in",
"[",
"1",
",",
"'1'",
"]",
")",
":",
"result",
"=",
"True",
"elif",
"(",
"result",
"in",
"[",
"0",
",",
"'0'",
"]",
")",
":",
"result",
"=",
"False",
"return",
"result"
] |
process given c{b{result}} based on flags set in c{b{method}} .
|
train
| false
|
48,343
|
def threshold_yen(image, nbins=256):
(hist, bin_centers) = histogram(image.ravel(), nbins)
if (bin_centers.size == 1):
return bin_centers[0]
pmf = (hist.astype(np.float32) / hist.sum())
P1 = np.cumsum(pmf)
P1_sq = np.cumsum((pmf ** 2))
P2_sq = np.cumsum((pmf[::(-1)] ** 2))[::(-1)]
crit = np.log((((P1_sq[:(-1)] * P2_sq[1:]) ** (-1)) * ((P1[:(-1)] * (1.0 - P1[:(-1)])) ** 2)))
return bin_centers[crit.argmax()]
|
[
"def",
"threshold_yen",
"(",
"image",
",",
"nbins",
"=",
"256",
")",
":",
"(",
"hist",
",",
"bin_centers",
")",
"=",
"histogram",
"(",
"image",
".",
"ravel",
"(",
")",
",",
"nbins",
")",
"if",
"(",
"bin_centers",
".",
"size",
"==",
"1",
")",
":",
"return",
"bin_centers",
"[",
"0",
"]",
"pmf",
"=",
"(",
"hist",
".",
"astype",
"(",
"np",
".",
"float32",
")",
"/",
"hist",
".",
"sum",
"(",
")",
")",
"P1",
"=",
"np",
".",
"cumsum",
"(",
"pmf",
")",
"P1_sq",
"=",
"np",
".",
"cumsum",
"(",
"(",
"pmf",
"**",
"2",
")",
")",
"P2_sq",
"=",
"np",
".",
"cumsum",
"(",
"(",
"pmf",
"[",
":",
":",
"(",
"-",
"1",
")",
"]",
"**",
"2",
")",
")",
"[",
":",
":",
"(",
"-",
"1",
")",
"]",
"crit",
"=",
"np",
".",
"log",
"(",
"(",
"(",
"(",
"P1_sq",
"[",
":",
"(",
"-",
"1",
")",
"]",
"*",
"P2_sq",
"[",
"1",
":",
"]",
")",
"**",
"(",
"-",
"1",
")",
")",
"*",
"(",
"(",
"P1",
"[",
":",
"(",
"-",
"1",
")",
"]",
"*",
"(",
"1.0",
"-",
"P1",
"[",
":",
"(",
"-",
"1",
")",
"]",
")",
")",
"**",
"2",
")",
")",
")",
"return",
"bin_centers",
"[",
"crit",
".",
"argmax",
"(",
")",
"]"
] |
return threshold value based on yens method .
|
train
| false
|
48,344
|
def write_worksheet_hyperlinks(doc, worksheet):
write_hyperlinks = False
for cell in worksheet.get_cell_collection():
if (cell.hyperlink_rel_id is not None):
write_hyperlinks = True
break
if write_hyperlinks:
start_tag(doc, 'hyperlinks')
for cell in worksheet.get_cell_collection():
if (cell.hyperlink_rel_id is not None):
attrs = {'display': cell.hyperlink, 'ref': cell.get_coordinate(), 'r:id': cell.hyperlink_rel_id}
tag(doc, 'hyperlink', attrs)
end_tag(doc, 'hyperlinks')
|
[
"def",
"write_worksheet_hyperlinks",
"(",
"doc",
",",
"worksheet",
")",
":",
"write_hyperlinks",
"=",
"False",
"for",
"cell",
"in",
"worksheet",
".",
"get_cell_collection",
"(",
")",
":",
"if",
"(",
"cell",
".",
"hyperlink_rel_id",
"is",
"not",
"None",
")",
":",
"write_hyperlinks",
"=",
"True",
"break",
"if",
"write_hyperlinks",
":",
"start_tag",
"(",
"doc",
",",
"'hyperlinks'",
")",
"for",
"cell",
"in",
"worksheet",
".",
"get_cell_collection",
"(",
")",
":",
"if",
"(",
"cell",
".",
"hyperlink_rel_id",
"is",
"not",
"None",
")",
":",
"attrs",
"=",
"{",
"'display'",
":",
"cell",
".",
"hyperlink",
",",
"'ref'",
":",
"cell",
".",
"get_coordinate",
"(",
")",
",",
"'r:id'",
":",
"cell",
".",
"hyperlink_rel_id",
"}",
"tag",
"(",
"doc",
",",
"'hyperlink'",
",",
"attrs",
")",
"end_tag",
"(",
"doc",
",",
"'hyperlinks'",
")"
] |
write worksheet hyperlinks to xml .
|
train
| false
|
48,345
|
def set_remove_subs(ssli):
part = []
for s in sorted(list(set(ssli)), key=(lambda x: len(set(x))))[::(-1)]:
if (not any((set(s).issubset(set(t)) for t in part))):
part.append(s)
return part
|
[
"def",
"set_remove_subs",
"(",
"ssli",
")",
":",
"part",
"=",
"[",
"]",
"for",
"s",
"in",
"sorted",
"(",
"list",
"(",
"set",
"(",
"ssli",
")",
")",
",",
"key",
"=",
"(",
"lambda",
"x",
":",
"len",
"(",
"set",
"(",
"x",
")",
")",
")",
")",
"[",
":",
":",
"(",
"-",
"1",
")",
"]",
":",
"if",
"(",
"not",
"any",
"(",
"(",
"set",
"(",
"s",
")",
".",
"issubset",
"(",
"set",
"(",
"t",
")",
")",
"for",
"t",
"in",
"part",
")",
")",
")",
":",
"part",
".",
"append",
"(",
"s",
")",
"return",
"part"
] |
remove sets that are subsets of another set from a list of tuples parameters ssli : list of tuples each tuple is considered as a set returns part : list of tuples new list with subset tuples removed .
|
train
| false
|
48,346
|
def for_all_dtypes_combination(names=('dtyes',), no_float16=False, no_bool=False, full=None):
types = _make_all_dtypes(no_float16, no_bool)
return for_dtypes_combination(types, names, full)
|
[
"def",
"for_all_dtypes_combination",
"(",
"names",
"=",
"(",
"'dtyes'",
",",
")",
",",
"no_float16",
"=",
"False",
",",
"no_bool",
"=",
"False",
",",
"full",
"=",
"None",
")",
":",
"types",
"=",
"_make_all_dtypes",
"(",
"no_float16",
",",
"no_bool",
")",
"return",
"for_dtypes_combination",
"(",
"types",
",",
"names",
",",
"full",
")"
] |
decorator that checks the fixture with a product set of all dtypes .
|
train
| false
|
48,347
|
def send_video_status_update(updates):
for update in updates:
update_video_status(update.get('edxVideoId'), update.get('status'))
LOGGER.info('VIDEOS: Video status update with id [%s], status [%s] and message [%s]', update.get('edxVideoId'), update.get('status'), update.get('message'))
return JsonResponse()
|
[
"def",
"send_video_status_update",
"(",
"updates",
")",
":",
"for",
"update",
"in",
"updates",
":",
"update_video_status",
"(",
"update",
".",
"get",
"(",
"'edxVideoId'",
")",
",",
"update",
".",
"get",
"(",
"'status'",
")",
")",
"LOGGER",
".",
"info",
"(",
"'VIDEOS: Video status update with id [%s], status [%s] and message [%s]'",
",",
"update",
".",
"get",
"(",
"'edxVideoId'",
")",
",",
"update",
".",
"get",
"(",
"'status'",
")",
",",
"update",
".",
"get",
"(",
"'message'",
")",
")",
"return",
"JsonResponse",
"(",
")"
] |
update video status in edx-val .
|
train
| false
|
48,349
|
def __main__(argv=None):
import sys as _sys
if (not argv):
argv = _sys.argv
exitcode = None
try:
unittest.main(argv=argv, defaultTest='suite')
except SystemExit as exc:
exitcode = exc.code
return exitcode
|
[
"def",
"__main__",
"(",
"argv",
"=",
"None",
")",
":",
"import",
"sys",
"as",
"_sys",
"if",
"(",
"not",
"argv",
")",
":",
"argv",
"=",
"_sys",
".",
"argv",
"exitcode",
"=",
"None",
"try",
":",
"unittest",
".",
"main",
"(",
"argv",
"=",
"argv",
",",
"defaultTest",
"=",
"'suite'",
")",
"except",
"SystemExit",
"as",
"exc",
":",
"exitcode",
"=",
"exc",
".",
"code",
"return",
"exitcode"
] |
mainline function for this module .
|
train
| false
|
48,350
|
def drop_views(manager, views):
check_exists(manager, views, VIEW_TYPE)
for view in views:
manager.execute(('DROP VIEW `%s`' % view))
|
[
"def",
"drop_views",
"(",
"manager",
",",
"views",
")",
":",
"check_exists",
"(",
"manager",
",",
"views",
",",
"VIEW_TYPE",
")",
"for",
"view",
"in",
"views",
":",
"manager",
".",
"execute",
"(",
"(",
"'DROP VIEW `%s`'",
"%",
"view",
")",
")"
] |
drops the specified views from the database if a specified view does not exist in the database .
|
train
| false
|
48,351
|
def flightmode_menu():
modes = mestate.mlog.flightmode_list()
ret = []
idx = 0
for (mode, t1, t2) in modes:
modestr = ('%s %us' % (mode, (t2 - t1)))
ret.append(MPMenuCheckbox(modestr, modestr, ('mode-%u' % idx)))
idx += 1
mestate.flightmode_selections.append(False)
return ret
|
[
"def",
"flightmode_menu",
"(",
")",
":",
"modes",
"=",
"mestate",
".",
"mlog",
".",
"flightmode_list",
"(",
")",
"ret",
"=",
"[",
"]",
"idx",
"=",
"0",
"for",
"(",
"mode",
",",
"t1",
",",
"t2",
")",
"in",
"modes",
":",
"modestr",
"=",
"(",
"'%s %us'",
"%",
"(",
"mode",
",",
"(",
"t2",
"-",
"t1",
")",
")",
")",
"ret",
".",
"append",
"(",
"MPMenuCheckbox",
"(",
"modestr",
",",
"modestr",
",",
"(",
"'mode-%u'",
"%",
"idx",
")",
")",
")",
"idx",
"+=",
"1",
"mestate",
".",
"flightmode_selections",
".",
"append",
"(",
"False",
")",
"return",
"ret"
] |
construct flightmode menu .
|
train
| true
|
48,352
|
def url_unquote(s, charset='utf-8', errors='replace'):
if isinstance(s, unicode):
s = s.encode(charset)
return _decode_unicode(_unquote(s), charset, errors)
|
[
"def",
"url_unquote",
"(",
"s",
",",
"charset",
"=",
"'utf-8'",
",",
"errors",
"=",
"'replace'",
")",
":",
"if",
"isinstance",
"(",
"s",
",",
"unicode",
")",
":",
"s",
"=",
"s",
".",
"encode",
"(",
"charset",
")",
"return",
"_decode_unicode",
"(",
"_unquote",
"(",
"s",
")",
",",
"charset",
",",
"errors",
")"
] |
url decode a single string with a given encoding .
|
train
| true
|
48,353
|
def image_to_base64(image):
ax = image.axes
binary_buffer = io.BytesIO()
lim = ax.axis()
ax.axis(image.get_extent())
image.write_png(binary_buffer)
ax.axis(lim)
binary_buffer.seek(0)
return base64.b64encode(binary_buffer.read()).decode('utf-8')
|
[
"def",
"image_to_base64",
"(",
"image",
")",
":",
"ax",
"=",
"image",
".",
"axes",
"binary_buffer",
"=",
"io",
".",
"BytesIO",
"(",
")",
"lim",
"=",
"ax",
".",
"axis",
"(",
")",
"ax",
".",
"axis",
"(",
"image",
".",
"get_extent",
"(",
")",
")",
"image",
".",
"write_png",
"(",
"binary_buffer",
")",
"ax",
".",
"axis",
"(",
"lim",
")",
"binary_buffer",
".",
"seek",
"(",
"0",
")",
"return",
"base64",
".",
"b64encode",
"(",
"binary_buffer",
".",
"read",
"(",
")",
")",
".",
"decode",
"(",
"'utf-8'",
")"
] |
convert a matplotlib image to a base64 png representation parameters image : matplotlib image object the image to be converted .
|
train
| true
|
48,356
|
def get_bumper_sources(video):
try:
val_profiles = ['desktop_webm', 'desktop_mp4']
val_video_urls = edxval_api.get_urls_for_profiles(video.bumper['edx_video_id'], val_profiles)
bumper_sources = filter(None, [val_video_urls[p] for p in val_profiles])
except edxval_api.ValInternalError:
log.warning('Could not retrieve information from VAL for Bumper edx Video ID: %s.', video.bumper['edx_video_id'])
return []
return bumper_sources
|
[
"def",
"get_bumper_sources",
"(",
"video",
")",
":",
"try",
":",
"val_profiles",
"=",
"[",
"'desktop_webm'",
",",
"'desktop_mp4'",
"]",
"val_video_urls",
"=",
"edxval_api",
".",
"get_urls_for_profiles",
"(",
"video",
".",
"bumper",
"[",
"'edx_video_id'",
"]",
",",
"val_profiles",
")",
"bumper_sources",
"=",
"filter",
"(",
"None",
",",
"[",
"val_video_urls",
"[",
"p",
"]",
"for",
"p",
"in",
"val_profiles",
"]",
")",
"except",
"edxval_api",
".",
"ValInternalError",
":",
"log",
".",
"warning",
"(",
"'Could not retrieve information from VAL for Bumper edx Video ID: %s.'",
",",
"video",
".",
"bumper",
"[",
"'edx_video_id'",
"]",
")",
"return",
"[",
"]",
"return",
"bumper_sources"
] |
get bumper sources from edxval .
|
train
| false
|
48,357
|
def set_symbols(pcontracts, dt_start='1980-1-1', dt_end='2100-1-1', n=None, spec_date={}):
global _simulator
_simulator = ExecuteUnit(pcontracts, dt_start, dt_end, n, spec_date)
return _simulator
|
[
"def",
"set_symbols",
"(",
"pcontracts",
",",
"dt_start",
"=",
"'1980-1-1'",
",",
"dt_end",
"=",
"'2100-1-1'",
",",
"n",
"=",
"None",
",",
"spec_date",
"=",
"{",
"}",
")",
":",
"global",
"_simulator",
"_simulator",
"=",
"ExecuteUnit",
"(",
"pcontracts",
",",
"dt_start",
",",
"dt_end",
",",
"n",
",",
"spec_date",
")",
"return",
"_simulator"
] |
args: pcontracts : list of pcontracts dt_start : start time of all pcontracts dt_end : end time of all pcontracts n : last n bars spec_date : time range for specific pcontracts .
|
train
| false
|
48,358
|
def GetIOServicesByType(service_type):
serial_port_iterator = ctypes.c_void_p()
iokit.IOServiceGetMatchingServices(kIOMasterPortDefault, iokit.IOServiceMatching(service_type.encode('mac_roman')), ctypes.byref(serial_port_iterator))
services = []
while iokit.IOIteratorIsValid(serial_port_iterator):
service = iokit.IOIteratorNext(serial_port_iterator)
if (not service):
break
services.append(service)
iokit.IOObjectRelease(serial_port_iterator)
return services
|
[
"def",
"GetIOServicesByType",
"(",
"service_type",
")",
":",
"serial_port_iterator",
"=",
"ctypes",
".",
"c_void_p",
"(",
")",
"iokit",
".",
"IOServiceGetMatchingServices",
"(",
"kIOMasterPortDefault",
",",
"iokit",
".",
"IOServiceMatching",
"(",
"service_type",
".",
"encode",
"(",
"'mac_roman'",
")",
")",
",",
"ctypes",
".",
"byref",
"(",
"serial_port_iterator",
")",
")",
"services",
"=",
"[",
"]",
"while",
"iokit",
".",
"IOIteratorIsValid",
"(",
"serial_port_iterator",
")",
":",
"service",
"=",
"iokit",
".",
"IOIteratorNext",
"(",
"serial_port_iterator",
")",
"if",
"(",
"not",
"service",
")",
":",
"break",
"services",
".",
"append",
"(",
"service",
")",
"iokit",
".",
"IOObjectRelease",
"(",
"serial_port_iterator",
")",
"return",
"services"
] |
returns iterator over specified service_type .
|
train
| false
|
48,359
|
def server_error(request, template_name='500.html'):
r = render_to_response(template_name, context_instance=RequestContext(request))
r.status_code = 500
return r
|
[
"def",
"server_error",
"(",
"request",
",",
"template_name",
"=",
"'500.html'",
")",
":",
"r",
"=",
"render_to_response",
"(",
"template_name",
",",
"context_instance",
"=",
"RequestContext",
"(",
"request",
")",
")",
"r",
".",
"status_code",
"=",
"500",
"return",
"r"
] |
a simple 500 handler so we get media .
|
train
| false
|
48,360
|
def decrypt_secret(secret, key):
decrypted_data = ''
j = 0
for i in range(0, len(secret), 8):
enc_block = secret[i:(i + 8)]
block_key = key[j:(j + 7)]
des_key = str_to_key(block_key)
des = DES.new(des_key, DES.MODE_ECB)
decrypted_data += des.decrypt(enc_block)
j += 7
if (len(key[j:(j + 7)]) < 7):
j = len(key[j:(j + 7)])
(dec_data_len,) = unpack('<L', decrypted_data[:4])
return decrypted_data[8:(8 + dec_data_len)]
|
[
"def",
"decrypt_secret",
"(",
"secret",
",",
"key",
")",
":",
"decrypted_data",
"=",
"''",
"j",
"=",
"0",
"for",
"i",
"in",
"range",
"(",
"0",
",",
"len",
"(",
"secret",
")",
",",
"8",
")",
":",
"enc_block",
"=",
"secret",
"[",
"i",
":",
"(",
"i",
"+",
"8",
")",
"]",
"block_key",
"=",
"key",
"[",
"j",
":",
"(",
"j",
"+",
"7",
")",
"]",
"des_key",
"=",
"str_to_key",
"(",
"block_key",
")",
"des",
"=",
"DES",
".",
"new",
"(",
"des_key",
",",
"DES",
".",
"MODE_ECB",
")",
"decrypted_data",
"+=",
"des",
".",
"decrypt",
"(",
"enc_block",
")",
"j",
"+=",
"7",
"if",
"(",
"len",
"(",
"key",
"[",
"j",
":",
"(",
"j",
"+",
"7",
")",
"]",
")",
"<",
"7",
")",
":",
"j",
"=",
"len",
"(",
"key",
"[",
"j",
":",
"(",
"j",
"+",
"7",
")",
"]",
")",
"(",
"dec_data_len",
",",
")",
"=",
"unpack",
"(",
"'<L'",
",",
"decrypted_data",
"[",
":",
"4",
"]",
")",
"return",
"decrypted_data",
"[",
"8",
":",
"(",
"8",
"+",
"dec_data_len",
")",
"]"
] |
python implementation of systemfunction005 .
|
train
| false
|
48,361
|
def set_saucelabs_job_status(jobid, passed=True):
config = get_saucelabs_username_and_key()
url = 'http://saucelabs.com/rest/v1/{}/jobs/{}'.format(config['username'], world.jobid)
body_content = dumps({'passed': passed})
base64string = encodestring('{}:{}'.format(config['username'], config['access-key']))[:(-1)]
headers = {'Authorization': 'Basic {}'.format(base64string)}
result = requests.put(url, data=body_content, headers=headers)
return (result.status_code == 200)
|
[
"def",
"set_saucelabs_job_status",
"(",
"jobid",
",",
"passed",
"=",
"True",
")",
":",
"config",
"=",
"get_saucelabs_username_and_key",
"(",
")",
"url",
"=",
"'http://saucelabs.com/rest/v1/{}/jobs/{}'",
".",
"format",
"(",
"config",
"[",
"'username'",
"]",
",",
"world",
".",
"jobid",
")",
"body_content",
"=",
"dumps",
"(",
"{",
"'passed'",
":",
"passed",
"}",
")",
"base64string",
"=",
"encodestring",
"(",
"'{}:{}'",
".",
"format",
"(",
"config",
"[",
"'username'",
"]",
",",
"config",
"[",
"'access-key'",
"]",
")",
")",
"[",
":",
"(",
"-",
"1",
")",
"]",
"headers",
"=",
"{",
"'Authorization'",
":",
"'Basic {}'",
".",
"format",
"(",
"base64string",
")",
"}",
"result",
"=",
"requests",
".",
"put",
"(",
"url",
",",
"data",
"=",
"body_content",
",",
"headers",
"=",
"headers",
")",
"return",
"(",
"result",
".",
"status_code",
"==",
"200",
")"
] |
sets the job status on sauce labs .
|
train
| false
|
48,362
|
def use_aio():
from txaio import aio
txaio._use_framework(aio)
|
[
"def",
"use_aio",
"(",
")",
":",
"from",
"txaio",
"import",
"aio",
"txaio",
".",
"_use_framework",
"(",
"aio",
")"
] |
monkey-patched for doc-building .
|
train
| false
|
48,363
|
def get_trigger_db_by_uid(uid):
try:
return Trigger.get_by_uid(uid)
except StackStormDBObjectNotFoundError as e:
LOG.debug('Database lookup for uid="%s" resulted in exception : %s.', uid, e, exc_info=True)
return None
|
[
"def",
"get_trigger_db_by_uid",
"(",
"uid",
")",
":",
"try",
":",
"return",
"Trigger",
".",
"get_by_uid",
"(",
"uid",
")",
"except",
"StackStormDBObjectNotFoundError",
"as",
"e",
":",
"LOG",
".",
"debug",
"(",
"'Database lookup for uid=\"%s\" resulted in exception : %s.'",
",",
"uid",
",",
"e",
",",
"exc_info",
"=",
"True",
")",
"return",
"None"
] |
returns the trigger object from db given a trigger uid .
|
train
| false
|
48,364
|
def checkIntf(intf):
config = quietRun(('ifconfig %s 2>/dev/null' % intf), shell=True)
if (not config):
error('Error:', intf, 'does not exist!\n')
exit(1)
ips = re.findall('\\d+\\.\\d+\\.\\d+\\.\\d+', config)
if ips:
error('Error:', intf, 'has an IP address,and is probably in use!\n')
exit(1)
|
[
"def",
"checkIntf",
"(",
"intf",
")",
":",
"config",
"=",
"quietRun",
"(",
"(",
"'ifconfig %s 2>/dev/null'",
"%",
"intf",
")",
",",
"shell",
"=",
"True",
")",
"if",
"(",
"not",
"config",
")",
":",
"error",
"(",
"'Error:'",
",",
"intf",
",",
"'does not exist!\\n'",
")",
"exit",
"(",
"1",
")",
"ips",
"=",
"re",
".",
"findall",
"(",
"'\\\\d+\\\\.\\\\d+\\\\.\\\\d+\\\\.\\\\d+'",
",",
"config",
")",
"if",
"ips",
":",
"error",
"(",
"'Error:'",
",",
"intf",
",",
"'has an IP address,and is probably in use!\\n'",
")",
"exit",
"(",
"1",
")"
] |
make sure intf exists and is not configured .
|
train
| false
|
48,365
|
def ContinuousRV(symbol, density, set=Interval((- oo), oo)):
pdf = Lambda(symbol, density)
dist = ContinuousDistributionHandmade(pdf, set)
return SingleContinuousPSpace(symbol, dist).value
|
[
"def",
"ContinuousRV",
"(",
"symbol",
",",
"density",
",",
"set",
"=",
"Interval",
"(",
"(",
"-",
"oo",
")",
",",
"oo",
")",
")",
":",
"pdf",
"=",
"Lambda",
"(",
"symbol",
",",
"density",
")",
"dist",
"=",
"ContinuousDistributionHandmade",
"(",
"pdf",
",",
"set",
")",
"return",
"SingleContinuousPSpace",
"(",
"symbol",
",",
"dist",
")",
".",
"value"
] |
create a continuous random variable given the following: -- a symbol -- a probability density function -- set on which the pdf is valid returns a randomsymbol .
|
train
| false
|
48,366
|
def phi(n):
assert isinstance(n, integer_types)
if (n < 3):
return 1
result = 1
ff = factorization(n)
for f in ff:
e = f[1]
if (e > 1):
result = ((result * (f[0] ** (e - 1))) * (f[0] - 1))
else:
result = (result * (f[0] - 1))
return result
|
[
"def",
"phi",
"(",
"n",
")",
":",
"assert",
"isinstance",
"(",
"n",
",",
"integer_types",
")",
"if",
"(",
"n",
"<",
"3",
")",
":",
"return",
"1",
"result",
"=",
"1",
"ff",
"=",
"factorization",
"(",
"n",
")",
"for",
"f",
"in",
"ff",
":",
"e",
"=",
"f",
"[",
"1",
"]",
"if",
"(",
"e",
">",
"1",
")",
":",
"result",
"=",
"(",
"(",
"result",
"*",
"(",
"f",
"[",
"0",
"]",
"**",
"(",
"e",
"-",
"1",
")",
")",
")",
"*",
"(",
"f",
"[",
"0",
"]",
"-",
"1",
")",
")",
"else",
":",
"result",
"=",
"(",
"result",
"*",
"(",
"f",
"[",
"0",
"]",
"-",
"1",
")",
")",
"return",
"result"
] |
return the euler totient function of n .
|
train
| true
|
48,367
|
def MakeHTTPException(code=500, msg='Error'):
response = requests.Response()
response.status_code = code
return requests.ConnectionError(msg, response=response)
|
[
"def",
"MakeHTTPException",
"(",
"code",
"=",
"500",
",",
"msg",
"=",
"'Error'",
")",
":",
"response",
"=",
"requests",
".",
"Response",
"(",
")",
"response",
".",
"status_code",
"=",
"code",
"return",
"requests",
".",
"ConnectionError",
"(",
"msg",
",",
"response",
"=",
"response",
")"
] |
a helper for creating a httperror exception .
|
train
| false
|
48,368
|
def create_chart(klass, values, compute_values=True, **kws):
_chart = klass(values, title='title', xlabel='xlabel', ylabel='ylabel', legend='top_left', xscale='linear', yscale='linear', width=800, height=600, tools=True, filename=False, server=False, notebook=False, **kws)
return _chart
|
[
"def",
"create_chart",
"(",
"klass",
",",
"values",
",",
"compute_values",
"=",
"True",
",",
"**",
"kws",
")",
":",
"_chart",
"=",
"klass",
"(",
"values",
",",
"title",
"=",
"'title'",
",",
"xlabel",
"=",
"'xlabel'",
",",
"ylabel",
"=",
"'ylabel'",
",",
"legend",
"=",
"'top_left'",
",",
"xscale",
"=",
"'linear'",
",",
"yscale",
"=",
"'linear'",
",",
"width",
"=",
"800",
",",
"height",
"=",
"600",
",",
"tools",
"=",
"True",
",",
"filename",
"=",
"False",
",",
"server",
"=",
"False",
",",
"notebook",
"=",
"False",
",",
"**",
"kws",
")",
"return",
"_chart"
] |
create a new chart class instance with values and the extra kws keyword parameters .
|
train
| false
|
48,370
|
def find_subscription_type(subscription):
subs_available = list(constants.USER_SUBSCRIPTIONS_AVAILABLE.keys())
subs_available.extend(list(constants.NODE_SUBSCRIPTIONS_AVAILABLE.keys()))
for available in subs_available:
if (available in subscription):
return available
|
[
"def",
"find_subscription_type",
"(",
"subscription",
")",
":",
"subs_available",
"=",
"list",
"(",
"constants",
".",
"USER_SUBSCRIPTIONS_AVAILABLE",
".",
"keys",
"(",
")",
")",
"subs_available",
".",
"extend",
"(",
"list",
"(",
"constants",
".",
"NODE_SUBSCRIPTIONS_AVAILABLE",
".",
"keys",
"(",
")",
")",
")",
"for",
"available",
"in",
"subs_available",
":",
"if",
"(",
"available",
"in",
"subscription",
")",
":",
"return",
"available"
] |
find subscription type string within specific subscription .
|
train
| false
|
48,371
|
def get_session_plot_options():
return copy.deepcopy(_session['plot_options'])
|
[
"def",
"get_session_plot_options",
"(",
")",
":",
"return",
"copy",
".",
"deepcopy",
"(",
"_session",
"[",
"'plot_options'",
"]",
")"
] |
returns a copy of the user supplied plot options .
|
train
| false
|
48,372
|
def _read_tagdesc(f):
tagdesc = {'offset': _read_long(f)}
if (tagdesc['offset'] == (-1)):
tagdesc['offset'] = _read_uint64(f)
tagdesc['typecode'] = _read_long(f)
tagflags = _read_long(f)
tagdesc['array'] = ((tagflags & 4) == 4)
tagdesc['structure'] = ((tagflags & 32) == 32)
tagdesc['scalar'] = (tagdesc['typecode'] in DTYPE_DICT)
return tagdesc
|
[
"def",
"_read_tagdesc",
"(",
"f",
")",
":",
"tagdesc",
"=",
"{",
"'offset'",
":",
"_read_long",
"(",
"f",
")",
"}",
"if",
"(",
"tagdesc",
"[",
"'offset'",
"]",
"==",
"(",
"-",
"1",
")",
")",
":",
"tagdesc",
"[",
"'offset'",
"]",
"=",
"_read_uint64",
"(",
"f",
")",
"tagdesc",
"[",
"'typecode'",
"]",
"=",
"_read_long",
"(",
"f",
")",
"tagflags",
"=",
"_read_long",
"(",
"f",
")",
"tagdesc",
"[",
"'array'",
"]",
"=",
"(",
"(",
"tagflags",
"&",
"4",
")",
"==",
"4",
")",
"tagdesc",
"[",
"'structure'",
"]",
"=",
"(",
"(",
"tagflags",
"&",
"32",
")",
"==",
"32",
")",
"tagdesc",
"[",
"'scalar'",
"]",
"=",
"(",
"tagdesc",
"[",
"'typecode'",
"]",
"in",
"DTYPE_DICT",
")",
"return",
"tagdesc"
] |
function to read in a tag descriptor .
|
train
| false
|
48,373
|
def get_available_extension(name, user=None, host=None, port=None, maintenance_db=None, password=None, runas=None):
return available_extensions(user=user, host=host, port=port, maintenance_db=maintenance_db, password=password, runas=runas).get(name, None)
|
[
"def",
"get_available_extension",
"(",
"name",
",",
"user",
"=",
"None",
",",
"host",
"=",
"None",
",",
"port",
"=",
"None",
",",
"maintenance_db",
"=",
"None",
",",
"password",
"=",
"None",
",",
"runas",
"=",
"None",
")",
":",
"return",
"available_extensions",
"(",
"user",
"=",
"user",
",",
"host",
"=",
"host",
",",
"port",
"=",
"port",
",",
"maintenance_db",
"=",
"maintenance_db",
",",
"password",
"=",
"password",
",",
"runas",
"=",
"runas",
")",
".",
"get",
"(",
"name",
",",
"None",
")"
] |
get info about an available postgresql extension cli example: .
|
train
| true
|
48,374
|
def Tm_staluc(s, dnac=50, saltc=50, rna=0):
warnings.warn(('Tm_staluc may be depreciated in the future. Use Tm_NN ' + 'instead.'), PendingDeprecationWarning)
if (not rna):
return Tm_NN(s, dnac1=(dnac / 2.0), dnac2=(dnac / 2.0), Na=saltc)
elif (rna == 1):
return Tm_NN(s, dnac1=(dnac / 2.0), dnac2=(dnac / 2.0), Na=saltc, nn_table=RNA_NN2)
else:
raise ValueError('rna={0} not supported'.format(rna))
|
[
"def",
"Tm_staluc",
"(",
"s",
",",
"dnac",
"=",
"50",
",",
"saltc",
"=",
"50",
",",
"rna",
"=",
"0",
")",
":",
"warnings",
".",
"warn",
"(",
"(",
"'Tm_staluc may be depreciated in the future. Use Tm_NN '",
"+",
"'instead.'",
")",
",",
"PendingDeprecationWarning",
")",
"if",
"(",
"not",
"rna",
")",
":",
"return",
"Tm_NN",
"(",
"s",
",",
"dnac1",
"=",
"(",
"dnac",
"/",
"2.0",
")",
",",
"dnac2",
"=",
"(",
"dnac",
"/",
"2.0",
")",
",",
"Na",
"=",
"saltc",
")",
"elif",
"(",
"rna",
"==",
"1",
")",
":",
"return",
"Tm_NN",
"(",
"s",
",",
"dnac1",
"=",
"(",
"dnac",
"/",
"2.0",
")",
",",
"dnac2",
"=",
"(",
"dnac",
"/",
"2.0",
")",
",",
"Na",
"=",
"saltc",
",",
"nn_table",
"=",
"RNA_NN2",
")",
"else",
":",
"raise",
"ValueError",
"(",
"'rna={0} not supported'",
".",
"format",
"(",
"rna",
")",
")"
] |
returns dna/dna tm using nearest neighbor thermodynamics .
|
train
| false
|
48,375
|
def paypal_time(time_obj=None):
if (time_obj is None):
time_obj = time.gmtime()
return time.strftime(PayPalNVP.TIMESTAMP_FORMAT, time_obj)
|
[
"def",
"paypal_time",
"(",
"time_obj",
"=",
"None",
")",
":",
"if",
"(",
"time_obj",
"is",
"None",
")",
":",
"time_obj",
"=",
"time",
".",
"gmtime",
"(",
")",
"return",
"time",
".",
"strftime",
"(",
"PayPalNVP",
".",
"TIMESTAMP_FORMAT",
",",
"time_obj",
")"
] |
returns a time suitable for paypal time fields .
|
train
| true
|
48,376
|
def md5_hexdigest(file):
if isinstance(file, compat.string_types):
with open(file, u'rb') as infile:
return _md5_hexdigest(infile)
return _md5_hexdigest(file)
|
[
"def",
"md5_hexdigest",
"(",
"file",
")",
":",
"if",
"isinstance",
"(",
"file",
",",
"compat",
".",
"string_types",
")",
":",
"with",
"open",
"(",
"file",
",",
"u'rb'",
")",
"as",
"infile",
":",
"return",
"_md5_hexdigest",
"(",
"infile",
")",
"return",
"_md5_hexdigest",
"(",
"file",
")"
] |
calculate and return the md5 checksum for a given file .
|
train
| false
|
48,377
|
def uniform_labelings_scores(score_func, n_samples, n_clusters_range, fixed_n_classes=None, n_runs=5, seed=42):
random_labels = np.random.RandomState(seed).randint
scores = np.zeros((len(n_clusters_range), n_runs))
if (fixed_n_classes is not None):
labels_a = random_labels(low=0, high=fixed_n_classes, size=n_samples)
for (i, k) in enumerate(n_clusters_range):
for j in range(n_runs):
if (fixed_n_classes is None):
labels_a = random_labels(low=0, high=k, size=n_samples)
labels_b = random_labels(low=0, high=k, size=n_samples)
scores[(i, j)] = score_func(labels_a, labels_b)
return scores
|
[
"def",
"uniform_labelings_scores",
"(",
"score_func",
",",
"n_samples",
",",
"n_clusters_range",
",",
"fixed_n_classes",
"=",
"None",
",",
"n_runs",
"=",
"5",
",",
"seed",
"=",
"42",
")",
":",
"random_labels",
"=",
"np",
".",
"random",
".",
"RandomState",
"(",
"seed",
")",
".",
"randint",
"scores",
"=",
"np",
".",
"zeros",
"(",
"(",
"len",
"(",
"n_clusters_range",
")",
",",
"n_runs",
")",
")",
"if",
"(",
"fixed_n_classes",
"is",
"not",
"None",
")",
":",
"labels_a",
"=",
"random_labels",
"(",
"low",
"=",
"0",
",",
"high",
"=",
"fixed_n_classes",
",",
"size",
"=",
"n_samples",
")",
"for",
"(",
"i",
",",
"k",
")",
"in",
"enumerate",
"(",
"n_clusters_range",
")",
":",
"for",
"j",
"in",
"range",
"(",
"n_runs",
")",
":",
"if",
"(",
"fixed_n_classes",
"is",
"None",
")",
":",
"labels_a",
"=",
"random_labels",
"(",
"low",
"=",
"0",
",",
"high",
"=",
"k",
",",
"size",
"=",
"n_samples",
")",
"labels_b",
"=",
"random_labels",
"(",
"low",
"=",
"0",
",",
"high",
"=",
"k",
",",
"size",
"=",
"n_samples",
")",
"scores",
"[",
"(",
"i",
",",
"j",
")",
"]",
"=",
"score_func",
"(",
"labels_a",
",",
"labels_b",
")",
"return",
"scores"
] |
compute score for 2 random uniform cluster labelings .
|
train
| false
|
48,378
|
def post_object(url, token, container, name, headers, http_conn=None, response_dict=None, service_token=None):
if http_conn:
(parsed, conn) = http_conn
else:
(parsed, conn) = http_connection(url)
path = ('%s/%s/%s' % (parsed.path, quote(container), quote(name)))
headers['X-Auth-Token'] = token
if service_token:
headers['X-Service-Token'] = service_token
conn.request('POST', path, '', headers)
resp = conn.getresponse()
body = resp.read()
http_log((('%s%s' % (url.replace(parsed.path, ''), path)), 'POST'), {'headers': headers}, resp, body)
store_response(resp, response_dict)
if ((resp.status < 200) or (resp.status >= 300)):
raise ClientException.from_response(resp, 'Object POST failed', body)
|
[
"def",
"post_object",
"(",
"url",
",",
"token",
",",
"container",
",",
"name",
",",
"headers",
",",
"http_conn",
"=",
"None",
",",
"response_dict",
"=",
"None",
",",
"service_token",
"=",
"None",
")",
":",
"if",
"http_conn",
":",
"(",
"parsed",
",",
"conn",
")",
"=",
"http_conn",
"else",
":",
"(",
"parsed",
",",
"conn",
")",
"=",
"http_connection",
"(",
"url",
")",
"path",
"=",
"(",
"'%s/%s/%s'",
"%",
"(",
"parsed",
".",
"path",
",",
"quote",
"(",
"container",
")",
",",
"quote",
"(",
"name",
")",
")",
")",
"headers",
"[",
"'X-Auth-Token'",
"]",
"=",
"token",
"if",
"service_token",
":",
"headers",
"[",
"'X-Service-Token'",
"]",
"=",
"service_token",
"conn",
".",
"request",
"(",
"'POST'",
",",
"path",
",",
"''",
",",
"headers",
")",
"resp",
"=",
"conn",
".",
"getresponse",
"(",
")",
"body",
"=",
"resp",
".",
"read",
"(",
")",
"http_log",
"(",
"(",
"(",
"'%s%s'",
"%",
"(",
"url",
".",
"replace",
"(",
"parsed",
".",
"path",
",",
"''",
")",
",",
"path",
")",
")",
",",
"'POST'",
")",
",",
"{",
"'headers'",
":",
"headers",
"}",
",",
"resp",
",",
"body",
")",
"store_response",
"(",
"resp",
",",
"response_dict",
")",
"if",
"(",
"(",
"resp",
".",
"status",
"<",
"200",
")",
"or",
"(",
"resp",
".",
"status",
">=",
"300",
")",
")",
":",
"raise",
"ClientException",
".",
"from_response",
"(",
"resp",
",",
"'Object POST failed'",
",",
"body",
")"
] |
update object metadata .
|
train
| false
|
48,379
|
def getNewRepository():
return ExportRepository()
|
[
"def",
"getNewRepository",
"(",
")",
":",
"return",
"ExportRepository",
"(",
")"
] |
get new repository .
|
train
| false
|
48,380
|
def with_kill_srv(f):
@wraps(f)
def wrapper(self, *args):
pidfile = args[(-1)]
try:
return f(self, *args)
finally:
_kill_srv(pidfile)
return wrapper
|
[
"def",
"with_kill_srv",
"(",
"f",
")",
":",
"@",
"wraps",
"(",
"f",
")",
"def",
"wrapper",
"(",
"self",
",",
"*",
"args",
")",
":",
"pidfile",
"=",
"args",
"[",
"(",
"-",
"1",
")",
"]",
"try",
":",
"return",
"f",
"(",
"self",
",",
"*",
"args",
")",
"finally",
":",
"_kill_srv",
"(",
"pidfile",
")",
"return",
"wrapper"
] |
helper to decorate tests which receive in the last argument tmpdir to pass to kill_srv to be used in tandem with @with_tmpdir .
|
train
| false
|
48,381
|
def _get_default_retry_params():
default = getattr(_thread_local_settings, 'default_retry_params', None)
if ((default is None) or (not default.belong_to_current_request())):
return RetryParams()
else:
return copy.copy(default)
|
[
"def",
"_get_default_retry_params",
"(",
")",
":",
"default",
"=",
"getattr",
"(",
"_thread_local_settings",
",",
"'default_retry_params'",
",",
"None",
")",
"if",
"(",
"(",
"default",
"is",
"None",
")",
"or",
"(",
"not",
"default",
".",
"belong_to_current_request",
"(",
")",
")",
")",
":",
"return",
"RetryParams",
"(",
")",
"else",
":",
"return",
"copy",
".",
"copy",
"(",
"default",
")"
] |
get default retryparams for current request and current thread .
|
train
| true
|
48,383
|
def _get_env():
env = dict(os.environ)
env['PYTHONPATH'] = os.pathsep.join(sys.path)
return env
|
[
"def",
"_get_env",
"(",
")",
":",
"env",
"=",
"dict",
"(",
"os",
".",
"environ",
")",
"env",
"[",
"'PYTHONPATH'",
"]",
"=",
"os",
".",
"pathsep",
".",
"join",
"(",
"sys",
".",
"path",
")",
"return",
"env"
] |
extracts the environment pythonpath and appends the current sys .
|
train
| true
|
48,384
|
def list_manage_opts():
return [(g, copy.deepcopy(o)) for (g, o) in _manage_opts]
|
[
"def",
"list_manage_opts",
"(",
")",
":",
"return",
"[",
"(",
"g",
",",
"copy",
".",
"deepcopy",
"(",
"o",
")",
")",
"for",
"(",
"g",
",",
"o",
")",
"in",
"_manage_opts",
"]"
] |
return a list of oslo_config options available in glance manage .
|
train
| false
|
48,385
|
def _get_dtype_from_object(dtype):
if (isinstance(dtype, type) and issubclass(dtype, np.generic)):
return dtype
elif is_categorical(dtype):
return CategoricalDtype().type
elif is_datetimetz(dtype):
return DatetimeTZDtype(dtype).type
elif isinstance(dtype, np.dtype):
try:
_validate_date_like_dtype(dtype)
except TypeError:
pass
return dtype.type
elif isinstance(dtype, string_types):
if (dtype in ['datetimetz', 'datetime64tz']):
return DatetimeTZDtype.type
elif (dtype in ['period']):
raise NotImplementedError
if ((dtype == 'datetime') or (dtype == 'timedelta')):
dtype += '64'
try:
return _get_dtype_from_object(getattr(np, dtype))
except (AttributeError, TypeError):
pass
return _get_dtype_from_object(np.dtype(dtype))
|
[
"def",
"_get_dtype_from_object",
"(",
"dtype",
")",
":",
"if",
"(",
"isinstance",
"(",
"dtype",
",",
"type",
")",
"and",
"issubclass",
"(",
"dtype",
",",
"np",
".",
"generic",
")",
")",
":",
"return",
"dtype",
"elif",
"is_categorical",
"(",
"dtype",
")",
":",
"return",
"CategoricalDtype",
"(",
")",
".",
"type",
"elif",
"is_datetimetz",
"(",
"dtype",
")",
":",
"return",
"DatetimeTZDtype",
"(",
"dtype",
")",
".",
"type",
"elif",
"isinstance",
"(",
"dtype",
",",
"np",
".",
"dtype",
")",
":",
"try",
":",
"_validate_date_like_dtype",
"(",
"dtype",
")",
"except",
"TypeError",
":",
"pass",
"return",
"dtype",
".",
"type",
"elif",
"isinstance",
"(",
"dtype",
",",
"string_types",
")",
":",
"if",
"(",
"dtype",
"in",
"[",
"'datetimetz'",
",",
"'datetime64tz'",
"]",
")",
":",
"return",
"DatetimeTZDtype",
".",
"type",
"elif",
"(",
"dtype",
"in",
"[",
"'period'",
"]",
")",
":",
"raise",
"NotImplementedError",
"if",
"(",
"(",
"dtype",
"==",
"'datetime'",
")",
"or",
"(",
"dtype",
"==",
"'timedelta'",
")",
")",
":",
"dtype",
"+=",
"'64'",
"try",
":",
"return",
"_get_dtype_from_object",
"(",
"getattr",
"(",
"np",
",",
"dtype",
")",
")",
"except",
"(",
"AttributeError",
",",
"TypeError",
")",
":",
"pass",
"return",
"_get_dtype_from_object",
"(",
"np",
".",
"dtype",
"(",
"dtype",
")",
")"
] |
get a numpy dtype .
|
train
| false
|
48,386
|
def LoadConfigsFromFile(file_path):
with open(file_path) as data:
return {d['check_id']: d for d in yaml.safe_load_all(data)}
|
[
"def",
"LoadConfigsFromFile",
"(",
"file_path",
")",
":",
"with",
"open",
"(",
"file_path",
")",
"as",
"data",
":",
"return",
"{",
"d",
"[",
"'check_id'",
"]",
":",
"d",
"for",
"d",
"in",
"yaml",
".",
"safe_load_all",
"(",
"data",
")",
"}"
] |
loads check definitions from a file .
|
train
| false
|
48,387
|
def _dot_buildout(directory):
return os.path.join(os.path.abspath(directory), '.buildout')
|
[
"def",
"_dot_buildout",
"(",
"directory",
")",
":",
"return",
"os",
".",
"path",
".",
"join",
"(",
"os",
".",
"path",
".",
"abspath",
"(",
"directory",
")",
",",
"'.buildout'",
")"
] |
get the local marker directory .
|
train
| false
|
48,388
|
def make_syncdb():
statuses = []
hue_exec = os.path.join(common.INSTALL_ROOT, 'build', 'env', 'bin', 'hue')
if os.path.exists(hue_exec):
statuses.append(runcmd([hue_exec, 'syncdb', '--noinput']))
statuses.append(runcmd([hue_exec, 'migrate', '--merge']))
return (not any(statuses))
|
[
"def",
"make_syncdb",
"(",
")",
":",
"statuses",
"=",
"[",
"]",
"hue_exec",
"=",
"os",
".",
"path",
".",
"join",
"(",
"common",
".",
"INSTALL_ROOT",
",",
"'build'",
",",
"'env'",
",",
"'bin'",
",",
"'hue'",
")",
"if",
"os",
".",
"path",
".",
"exists",
"(",
"hue_exec",
")",
":",
"statuses",
".",
"append",
"(",
"runcmd",
"(",
"[",
"hue_exec",
",",
"'syncdb'",
",",
"'--noinput'",
"]",
")",
")",
"statuses",
".",
"append",
"(",
"runcmd",
"(",
"[",
"hue_exec",
",",
"'migrate'",
",",
"'--merge'",
"]",
")",
")",
"return",
"(",
"not",
"any",
"(",
"statuses",
")",
")"
] |
make_syncdb() -> true/false .
|
train
| false
|
48,389
|
@register.tag
def get_flatpages(parser, token):
bits = token.split_contents()
syntax_message = ("%(tag_name)s expects a syntax of %(tag_name)s ['url_starts_with'] [for user] as context_name" % dict(tag_name=bits[0]))
if ((len(bits) >= 3) and (len(bits) <= 6)):
if ((len(bits) % 2) == 0):
prefix = bits[1]
else:
prefix = None
if (bits[(-2)] != 'as'):
raise template.TemplateSyntaxError(syntax_message)
context_name = bits[(-1)]
if (len(bits) >= 5):
if (bits[(-4)] != 'for'):
raise template.TemplateSyntaxError(syntax_message)
user = bits[(-3)]
else:
user = None
return FlatpageNode(context_name, starts_with=prefix, user=user)
else:
raise template.TemplateSyntaxError(syntax_message)
|
[
"@",
"register",
".",
"tag",
"def",
"get_flatpages",
"(",
"parser",
",",
"token",
")",
":",
"bits",
"=",
"token",
".",
"split_contents",
"(",
")",
"syntax_message",
"=",
"(",
"\"%(tag_name)s expects a syntax of %(tag_name)s ['url_starts_with'] [for user] as context_name\"",
"%",
"dict",
"(",
"tag_name",
"=",
"bits",
"[",
"0",
"]",
")",
")",
"if",
"(",
"(",
"len",
"(",
"bits",
")",
">=",
"3",
")",
"and",
"(",
"len",
"(",
"bits",
")",
"<=",
"6",
")",
")",
":",
"if",
"(",
"(",
"len",
"(",
"bits",
")",
"%",
"2",
")",
"==",
"0",
")",
":",
"prefix",
"=",
"bits",
"[",
"1",
"]",
"else",
":",
"prefix",
"=",
"None",
"if",
"(",
"bits",
"[",
"(",
"-",
"2",
")",
"]",
"!=",
"'as'",
")",
":",
"raise",
"template",
".",
"TemplateSyntaxError",
"(",
"syntax_message",
")",
"context_name",
"=",
"bits",
"[",
"(",
"-",
"1",
")",
"]",
"if",
"(",
"len",
"(",
"bits",
")",
">=",
"5",
")",
":",
"if",
"(",
"bits",
"[",
"(",
"-",
"4",
")",
"]",
"!=",
"'for'",
")",
":",
"raise",
"template",
".",
"TemplateSyntaxError",
"(",
"syntax_message",
")",
"user",
"=",
"bits",
"[",
"(",
"-",
"3",
")",
"]",
"else",
":",
"user",
"=",
"None",
"return",
"FlatpageNode",
"(",
"context_name",
",",
"starts_with",
"=",
"prefix",
",",
"user",
"=",
"user",
")",
"else",
":",
"raise",
"template",
".",
"TemplateSyntaxError",
"(",
"syntax_message",
")"
] |
retrieves all flatpage objects available for the current site and visible to the specific user .
|
train
| false
|
48,390
|
@receiver(SignalHandler.library_updated)
def listen_for_library_update(sender, library_key, **kwargs):
if LibrarySearchIndexer.indexing_is_enabled():
from .tasks import update_library_index
update_library_index.delay(unicode(library_key), datetime.now(UTC).isoformat())
|
[
"@",
"receiver",
"(",
"SignalHandler",
".",
"library_updated",
")",
"def",
"listen_for_library_update",
"(",
"sender",
",",
"library_key",
",",
"**",
"kwargs",
")",
":",
"if",
"LibrarySearchIndexer",
".",
"indexing_is_enabled",
"(",
")",
":",
"from",
".",
"tasks",
"import",
"update_library_index",
"update_library_index",
".",
"delay",
"(",
"unicode",
"(",
"library_key",
")",
",",
"datetime",
".",
"now",
"(",
"UTC",
")",
".",
"isoformat",
"(",
")",
")"
] |
receives signal and kicks off celery task to update search index .
|
train
| false
|
48,391
|
def all_users(number=(-1), etag=None):
return gh.all_users(number, etag)
|
[
"def",
"all_users",
"(",
"number",
"=",
"(",
"-",
"1",
")",
",",
"etag",
"=",
"None",
")",
":",
"return",
"gh",
".",
"all_users",
"(",
"number",
",",
"etag",
")"
] |
iterate over every user in the order they signed up for github .
|
train
| false
|
48,394
|
def listening_ports(attrs=None, where=None):
return _osquery_cmd(table='listening_ports', attrs=attrs, where=where)
|
[
"def",
"listening_ports",
"(",
"attrs",
"=",
"None",
",",
"where",
"=",
"None",
")",
":",
"return",
"_osquery_cmd",
"(",
"table",
"=",
"'listening_ports'",
",",
"attrs",
"=",
"attrs",
",",
"where",
"=",
"where",
")"
] |
return listening_ports information from osquery cli example: .
|
train
| false
|
48,395
|
def standardise_name(name):
try:
return numeric_to_rational(''.join(name))
except (ValueError, ZeroDivisionError):
return ''.join((ch for ch in name if (ch not in '_- '))).upper()
|
[
"def",
"standardise_name",
"(",
"name",
")",
":",
"try",
":",
"return",
"numeric_to_rational",
"(",
"''",
".",
"join",
"(",
"name",
")",
")",
"except",
"(",
"ValueError",
",",
"ZeroDivisionError",
")",
":",
"return",
"''",
".",
"join",
"(",
"(",
"ch",
"for",
"ch",
"in",
"name",
"if",
"(",
"ch",
"not",
"in",
"'_- '",
")",
")",
")",
".",
"upper",
"(",
")"
] |
standardises a property or value name .
|
train
| false
|
48,396
|
def is_ascii_encodable(s):
try:
s.encode('ascii')
except UnicodeEncodeError:
return False
except UnicodeDecodeError:
return False
except AttributeError:
return False
return True
|
[
"def",
"is_ascii_encodable",
"(",
"s",
")",
":",
"try",
":",
"s",
".",
"encode",
"(",
"'ascii'",
")",
"except",
"UnicodeEncodeError",
":",
"return",
"False",
"except",
"UnicodeDecodeError",
":",
"return",
"False",
"except",
"AttributeError",
":",
"return",
"False",
"return",
"True"
] |
check if argument encodes to ascii without error .
|
train
| false
|
48,397
|
def cloudfiles(module, container_, state, meta_, clear_meta, typ, ttl, public, private, web_index, web_error):
cf = pyrax.cloudfiles
if (cf is None):
module.fail_json(msg='Failed to instantiate client. This typically indicates an invalid region or an incorrectly capitalized region name.')
if (typ == 'container'):
container(cf, module, container_, state, meta_, clear_meta, ttl, public, private, web_index, web_error)
else:
meta(cf, module, container_, state, meta_, clear_meta)
|
[
"def",
"cloudfiles",
"(",
"module",
",",
"container_",
",",
"state",
",",
"meta_",
",",
"clear_meta",
",",
"typ",
",",
"ttl",
",",
"public",
",",
"private",
",",
"web_index",
",",
"web_error",
")",
":",
"cf",
"=",
"pyrax",
".",
"cloudfiles",
"if",
"(",
"cf",
"is",
"None",
")",
":",
"module",
".",
"fail_json",
"(",
"msg",
"=",
"'Failed to instantiate client. This typically indicates an invalid region or an incorrectly capitalized region name.'",
")",
"if",
"(",
"typ",
"==",
"'container'",
")",
":",
"container",
"(",
"cf",
",",
"module",
",",
"container_",
",",
"state",
",",
"meta_",
",",
"clear_meta",
",",
"ttl",
",",
"public",
",",
"private",
",",
"web_index",
",",
"web_error",
")",
"else",
":",
"meta",
"(",
"cf",
",",
"module",
",",
"container_",
",",
"state",
",",
"meta_",
",",
"clear_meta",
")"
] |
dispatch from here to work with metadata or file objects .
|
train
| false
|
48,400
|
def bin_list_to_int(bin_list):
return (bin_list << range(len(bin_list))).sum(0)
|
[
"def",
"bin_list_to_int",
"(",
"bin_list",
")",
":",
"return",
"(",
"bin_list",
"<<",
"range",
"(",
"len",
"(",
"bin_list",
")",
")",
")",
".",
"sum",
"(",
"0",
")"
] |
create int from binary repr from URL .
|
train
| false
|
48,401
|
def _create_row_request(table_name, row_key=None, start_key=None, end_key=None, filter_=None, limit=None):
request_kwargs = {'table_name': table_name}
if ((row_key is not None) and ((start_key is not None) or (end_key is not None))):
raise ValueError('Row key and row range cannot be set simultaneously')
range_kwargs = {}
if ((start_key is not None) or (end_key is not None)):
if (start_key is not None):
range_kwargs['start_key_closed'] = _to_bytes(start_key)
if (end_key is not None):
range_kwargs['end_key_open'] = _to_bytes(end_key)
if (filter_ is not None):
request_kwargs['filter'] = filter_.to_pb()
if (limit is not None):
request_kwargs['rows_limit'] = limit
message = data_messages_v2_pb2.ReadRowsRequest(**request_kwargs)
if (row_key is not None):
message.rows.row_keys.append(_to_bytes(row_key))
if range_kwargs:
message.rows.row_ranges.add(**range_kwargs)
return message
|
[
"def",
"_create_row_request",
"(",
"table_name",
",",
"row_key",
"=",
"None",
",",
"start_key",
"=",
"None",
",",
"end_key",
"=",
"None",
",",
"filter_",
"=",
"None",
",",
"limit",
"=",
"None",
")",
":",
"request_kwargs",
"=",
"{",
"'table_name'",
":",
"table_name",
"}",
"if",
"(",
"(",
"row_key",
"is",
"not",
"None",
")",
"and",
"(",
"(",
"start_key",
"is",
"not",
"None",
")",
"or",
"(",
"end_key",
"is",
"not",
"None",
")",
")",
")",
":",
"raise",
"ValueError",
"(",
"'Row key and row range cannot be set simultaneously'",
")",
"range_kwargs",
"=",
"{",
"}",
"if",
"(",
"(",
"start_key",
"is",
"not",
"None",
")",
"or",
"(",
"end_key",
"is",
"not",
"None",
")",
")",
":",
"if",
"(",
"start_key",
"is",
"not",
"None",
")",
":",
"range_kwargs",
"[",
"'start_key_closed'",
"]",
"=",
"_to_bytes",
"(",
"start_key",
")",
"if",
"(",
"end_key",
"is",
"not",
"None",
")",
":",
"range_kwargs",
"[",
"'end_key_open'",
"]",
"=",
"_to_bytes",
"(",
"end_key",
")",
"if",
"(",
"filter_",
"is",
"not",
"None",
")",
":",
"request_kwargs",
"[",
"'filter'",
"]",
"=",
"filter_",
".",
"to_pb",
"(",
")",
"if",
"(",
"limit",
"is",
"not",
"None",
")",
":",
"request_kwargs",
"[",
"'rows_limit'",
"]",
"=",
"limit",
"message",
"=",
"data_messages_v2_pb2",
".",
"ReadRowsRequest",
"(",
"**",
"request_kwargs",
")",
"if",
"(",
"row_key",
"is",
"not",
"None",
")",
":",
"message",
".",
"rows",
".",
"row_keys",
".",
"append",
"(",
"_to_bytes",
"(",
"row_key",
")",
")",
"if",
"range_kwargs",
":",
"message",
".",
"rows",
".",
"row_ranges",
".",
"add",
"(",
"**",
"range_kwargs",
")",
"return",
"message"
] |
creates a request to read rows in a table .
|
train
| false
|
48,402
|
def _pyopenssl_cert_or_req_san(cert_or_req):
part_separator = ':'
parts_separator = ', '
prefix = ('DNS' + part_separator)
if isinstance(cert_or_req, OpenSSL.crypto.X509):
func = OpenSSL.crypto.dump_certificate
else:
func = OpenSSL.crypto.dump_certificate_request
text = func(OpenSSL.crypto.FILETYPE_TEXT, cert_or_req).decode('utf-8')
match = re.search('X509v3 Subject Alternative Name:\\s*(.*)', text)
sans_parts = ([] if (match is None) else match.group(1).split(parts_separator))
return [part.split(part_separator)[1] for part in sans_parts if part.startswith(prefix)]
|
[
"def",
"_pyopenssl_cert_or_req_san",
"(",
"cert_or_req",
")",
":",
"part_separator",
"=",
"':'",
"parts_separator",
"=",
"', '",
"prefix",
"=",
"(",
"'DNS'",
"+",
"part_separator",
")",
"if",
"isinstance",
"(",
"cert_or_req",
",",
"OpenSSL",
".",
"crypto",
".",
"X509",
")",
":",
"func",
"=",
"OpenSSL",
".",
"crypto",
".",
"dump_certificate",
"else",
":",
"func",
"=",
"OpenSSL",
".",
"crypto",
".",
"dump_certificate_request",
"text",
"=",
"func",
"(",
"OpenSSL",
".",
"crypto",
".",
"FILETYPE_TEXT",
",",
"cert_or_req",
")",
".",
"decode",
"(",
"'utf-8'",
")",
"match",
"=",
"re",
".",
"search",
"(",
"'X509v3 Subject Alternative Name:\\\\s*(.*)'",
",",
"text",
")",
"sans_parts",
"=",
"(",
"[",
"]",
"if",
"(",
"match",
"is",
"None",
")",
"else",
"match",
".",
"group",
"(",
"1",
")",
".",
"split",
"(",
"parts_separator",
")",
")",
"return",
"[",
"part",
".",
"split",
"(",
"part_separator",
")",
"[",
"1",
"]",
"for",
"part",
"in",
"sans_parts",
"if",
"part",
".",
"startswith",
"(",
"prefix",
")",
"]"
] |
get subject alternative names from certificate or csr using pyopenssl .
|
train
| false
|
48,404
|
def test_render_to_file(Chart, datas):
file_name = ('/tmp/test_graph-%s.svg' % uuid.uuid4())
if os.path.exists(file_name):
os.remove(file_name)
chart = Chart()
chart = make_data(chart, datas)
chart.render_to_file(file_name)
with io.open(file_name, encoding='utf-8') as f:
assert ('pygal' in f.read())
os.remove(file_name)
|
[
"def",
"test_render_to_file",
"(",
"Chart",
",",
"datas",
")",
":",
"file_name",
"=",
"(",
"'/tmp/test_graph-%s.svg'",
"%",
"uuid",
".",
"uuid4",
"(",
")",
")",
"if",
"os",
".",
"path",
".",
"exists",
"(",
"file_name",
")",
":",
"os",
".",
"remove",
"(",
"file_name",
")",
"chart",
"=",
"Chart",
"(",
")",
"chart",
"=",
"make_data",
"(",
"chart",
",",
"datas",
")",
"chart",
".",
"render_to_file",
"(",
"file_name",
")",
"with",
"io",
".",
"open",
"(",
"file_name",
",",
"encoding",
"=",
"'utf-8'",
")",
"as",
"f",
":",
"assert",
"(",
"'pygal'",
"in",
"f",
".",
"read",
"(",
")",
")",
"os",
".",
"remove",
"(",
"file_name",
")"
] |
test in file rendering .
|
train
| false
|
48,405
|
def H(s):
return md5_hex(s)
|
[
"def",
"H",
"(",
"s",
")",
":",
"return",
"md5_hex",
"(",
"s",
")"
] |
the hash function h .
|
train
| false
|
48,406
|
def unpickle_backend(cls, args, kwargs):
return cls(app=current_app._get_current_object(), *args, **kwargs)
|
[
"def",
"unpickle_backend",
"(",
"cls",
",",
"args",
",",
"kwargs",
")",
":",
"return",
"cls",
"(",
"app",
"=",
"current_app",
".",
"_get_current_object",
"(",
")",
",",
"*",
"args",
",",
"**",
"kwargs",
")"
] |
return an unpickled backend .
|
train
| false
|
48,409
|
def install_openblas():
chdir(SRC_DIR)
apt_command('build-dep libopenblas-dev')
if glob.glob('*openblas*.deb'):
run_command('dpkg -i *openblas*.deb')
else:
apt_command('source libopenblas-dev')
chdir('openblas-*')
patch = open('fix_makefile_system.patch', 'w')
patch.write(OPENBLAS_0_1ALPHA_2_PATCH)
patch.close()
run_command(('patch -p1 < %s' % patch.name))
rule_file = open('Makefile.rule', 'a')
lines = ['DYNAMIC_ARCH=1', 'NUM_THREADS=64', 'NO_LAPACK=1', 'NO_AFFINITY=1']
rule_file.write('\n'.join(lines))
rule_file.close()
run_command('fakeroot debian/rules custom')
run_command('dpkg -i ../*openblas*.deb')
run_command('echo libopenblas-base hold | dpkg --set-selections')
run_command('echo libopenblas-dev hold | dpkg --set-selections')
|
[
"def",
"install_openblas",
"(",
")",
":",
"chdir",
"(",
"SRC_DIR",
")",
"apt_command",
"(",
"'build-dep libopenblas-dev'",
")",
"if",
"glob",
".",
"glob",
"(",
"'*openblas*.deb'",
")",
":",
"run_command",
"(",
"'dpkg -i *openblas*.deb'",
")",
"else",
":",
"apt_command",
"(",
"'source libopenblas-dev'",
")",
"chdir",
"(",
"'openblas-*'",
")",
"patch",
"=",
"open",
"(",
"'fix_makefile_system.patch'",
",",
"'w'",
")",
"patch",
".",
"write",
"(",
"OPENBLAS_0_1ALPHA_2_PATCH",
")",
"patch",
".",
"close",
"(",
")",
"run_command",
"(",
"(",
"'patch -p1 < %s'",
"%",
"patch",
".",
"name",
")",
")",
"rule_file",
"=",
"open",
"(",
"'Makefile.rule'",
",",
"'a'",
")",
"lines",
"=",
"[",
"'DYNAMIC_ARCH=1'",
",",
"'NUM_THREADS=64'",
",",
"'NO_LAPACK=1'",
",",
"'NO_AFFINITY=1'",
"]",
"rule_file",
".",
"write",
"(",
"'\\n'",
".",
"join",
"(",
"lines",
")",
")",
"rule_file",
".",
"close",
"(",
")",
"run_command",
"(",
"'fakeroot debian/rules custom'",
")",
"run_command",
"(",
"'dpkg -i ../*openblas*.deb'",
")",
"run_command",
"(",
"'echo libopenblas-base hold | dpkg --set-selections'",
")",
"run_command",
"(",
"'echo libopenblas-dev hold | dpkg --set-selections'",
")"
] |
docstring for install_openblas .
|
train
| false
|
48,412
|
def __IndexListForQuery(query):
(required, kind, ancestor, props) = datastore_index.CompositeIndexForQuery(query)
if (not required):
return []
index_pb = entity_pb.Index()
index_pb.set_entity_type(kind)
index_pb.set_ancestor(bool(ancestor))
for (name, direction) in datastore_index.GetRecommendedIndexProperties(props):
prop_pb = entity_pb.Index_Property()
prop_pb.set_name(name)
prop_pb.set_direction(direction)
index_pb.property_list().append(prop_pb)
return [index_pb]
|
[
"def",
"__IndexListForQuery",
"(",
"query",
")",
":",
"(",
"required",
",",
"kind",
",",
"ancestor",
",",
"props",
")",
"=",
"datastore_index",
".",
"CompositeIndexForQuery",
"(",
"query",
")",
"if",
"(",
"not",
"required",
")",
":",
"return",
"[",
"]",
"index_pb",
"=",
"entity_pb",
".",
"Index",
"(",
")",
"index_pb",
".",
"set_entity_type",
"(",
"kind",
")",
"index_pb",
".",
"set_ancestor",
"(",
"bool",
"(",
"ancestor",
")",
")",
"for",
"(",
"name",
",",
"direction",
")",
"in",
"datastore_index",
".",
"GetRecommendedIndexProperties",
"(",
"props",
")",
":",
"prop_pb",
"=",
"entity_pb",
".",
"Index_Property",
"(",
")",
"prop_pb",
".",
"set_name",
"(",
"name",
")",
"prop_pb",
".",
"set_direction",
"(",
"direction",
")",
"index_pb",
".",
"property_list",
"(",
")",
".",
"append",
"(",
"prop_pb",
")",
"return",
"[",
"index_pb",
"]"
] |
get the composite index definition used by the query .
|
train
| false
|
48,413
|
def issue_section_order(issue):
try:
return LOG_SECTION.values().index(issue_section(issue))
except:
return (-1)
|
[
"def",
"issue_section_order",
"(",
"issue",
")",
":",
"try",
":",
"return",
"LOG_SECTION",
".",
"values",
"(",
")",
".",
"index",
"(",
"issue_section",
"(",
"issue",
")",
")",
"except",
":",
"return",
"(",
"-",
"1",
")"
] |
returns the section order for the given issue .
|
train
| false
|
48,414
|
def test_routing_class_based_method_view_with_cli_routing():
@hug.object.http_methods()
class EndPoint(object, ):
@hug.object.cli
def get(self):
return 'hi there!'
def post(self):
return 'bye'
assert (hug.test.get(api, 'endpoint').data == 'hi there!')
assert (hug.test.post(api, 'endpoint').data == 'bye')
assert (hug.test.cli(EndPoint.get) == 'hi there!')
|
[
"def",
"test_routing_class_based_method_view_with_cli_routing",
"(",
")",
":",
"@",
"hug",
".",
"object",
".",
"http_methods",
"(",
")",
"class",
"EndPoint",
"(",
"object",
",",
")",
":",
"@",
"hug",
".",
"object",
".",
"cli",
"def",
"get",
"(",
"self",
")",
":",
"return",
"'hi there!'",
"def",
"post",
"(",
"self",
")",
":",
"return",
"'bye'",
"assert",
"(",
"hug",
".",
"test",
".",
"get",
"(",
"api",
",",
"'endpoint'",
")",
".",
"data",
"==",
"'hi there!'",
")",
"assert",
"(",
"hug",
".",
"test",
".",
"post",
"(",
"api",
",",
"'endpoint'",
")",
".",
"data",
"==",
"'bye'",
")",
"assert",
"(",
"hug",
".",
"test",
".",
"cli",
"(",
"EndPoint",
".",
"get",
")",
"==",
"'hi there!'",
")"
] |
test creating class based routers using method mappings exposing cli endpoints .
|
train
| false
|
48,415
|
def get_dist_ops(operator):
return (SpatiaLiteDistance(operator),)
|
[
"def",
"get_dist_ops",
"(",
"operator",
")",
":",
"return",
"(",
"SpatiaLiteDistance",
"(",
"operator",
")",
",",
")"
] |
returns operations for regular distances; spherical distances are not currently supported .
|
train
| false
|
48,416
|
def is_fakes3(s3_url):
if (s3_url is not None):
return (urlparse.urlparse(s3_url).scheme in ('fakes3', 'fakes3s'))
else:
return False
|
[
"def",
"is_fakes3",
"(",
"s3_url",
")",
":",
"if",
"(",
"s3_url",
"is",
"not",
"None",
")",
":",
"return",
"(",
"urlparse",
".",
"urlparse",
"(",
"s3_url",
")",
".",
"scheme",
"in",
"(",
"'fakes3'",
",",
"'fakes3s'",
")",
")",
"else",
":",
"return",
"False"
] |
return true if s3_url has scheme fakes3:// .
|
train
| false
|
48,417
|
@require_admin_context
def volume_type_qos_disassociate_all(context, qos_specs_id):
session = get_session()
with session.begin():
session.query(models.VolumeTypes).filter_by(qos_specs_id=qos_specs_id).update({'qos_specs_id': None, 'updated_at': timeutils.utcnow()})
|
[
"@",
"require_admin_context",
"def",
"volume_type_qos_disassociate_all",
"(",
"context",
",",
"qos_specs_id",
")",
":",
"session",
"=",
"get_session",
"(",
")",
"with",
"session",
".",
"begin",
"(",
")",
":",
"session",
".",
"query",
"(",
"models",
".",
"VolumeTypes",
")",
".",
"filter_by",
"(",
"qos_specs_id",
"=",
"qos_specs_id",
")",
".",
"update",
"(",
"{",
"'qos_specs_id'",
":",
"None",
",",
"'updated_at'",
":",
"timeutils",
".",
"utcnow",
"(",
")",
"}",
")"
] |
disassociate all volume types associated with specified qos specs .
|
train
| false
|
48,419
|
def parse_externals_xml(decoded_str, prefix=''):
prefix = os.path.normpath(prefix)
prefix = os.path.normcase(prefix)
doc = xml.dom.pulldom.parseString(_get_xml_data(decoded_str))
externals = list()
for (event, node) in doc:
if ((event == 'START_ELEMENT') and (node.nodeName == 'target')):
doc.expandNode(node)
path = os.path.normpath(node.getAttribute('path'))
if os.path.normcase(path).startswith(prefix):
path = path[(len(prefix) + 1):]
data = _get_target_property(node)
for external in parse_external_prop(data):
externals.append(joinpath(path, external))
return externals
|
[
"def",
"parse_externals_xml",
"(",
"decoded_str",
",",
"prefix",
"=",
"''",
")",
":",
"prefix",
"=",
"os",
".",
"path",
".",
"normpath",
"(",
"prefix",
")",
"prefix",
"=",
"os",
".",
"path",
".",
"normcase",
"(",
"prefix",
")",
"doc",
"=",
"xml",
".",
"dom",
".",
"pulldom",
".",
"parseString",
"(",
"_get_xml_data",
"(",
"decoded_str",
")",
")",
"externals",
"=",
"list",
"(",
")",
"for",
"(",
"event",
",",
"node",
")",
"in",
"doc",
":",
"if",
"(",
"(",
"event",
"==",
"'START_ELEMENT'",
")",
"and",
"(",
"node",
".",
"nodeName",
"==",
"'target'",
")",
")",
":",
"doc",
".",
"expandNode",
"(",
"node",
")",
"path",
"=",
"os",
".",
"path",
".",
"normpath",
"(",
"node",
".",
"getAttribute",
"(",
"'path'",
")",
")",
"if",
"os",
".",
"path",
".",
"normcase",
"(",
"path",
")",
".",
"startswith",
"(",
"prefix",
")",
":",
"path",
"=",
"path",
"[",
"(",
"len",
"(",
"prefix",
")",
"+",
"1",
")",
":",
"]",
"data",
"=",
"_get_target_property",
"(",
"node",
")",
"for",
"external",
"in",
"parse_external_prop",
"(",
"data",
")",
":",
"externals",
".",
"append",
"(",
"joinpath",
"(",
"path",
",",
"external",
")",
")",
"return",
"externals"
] |
parse a propget svn:externals xml .
|
train
| false
|
48,420
|
def testunsupportedpagebreak():
document = newdocument()
docbody = document.xpath('/w:document/w:body', namespaces=nsprefixes)[0]
try:
docbody.append(pagebreak(type='unsup'))
except ValueError:
return
assert False
|
[
"def",
"testunsupportedpagebreak",
"(",
")",
":",
"document",
"=",
"newdocument",
"(",
")",
"docbody",
"=",
"document",
".",
"xpath",
"(",
"'/w:document/w:body'",
",",
"namespaces",
"=",
"nsprefixes",
")",
"[",
"0",
"]",
"try",
":",
"docbody",
".",
"append",
"(",
"pagebreak",
"(",
"type",
"=",
"'unsup'",
")",
")",
"except",
"ValueError",
":",
"return",
"assert",
"False"
] |
ensure unsupported page break types are trapped .
|
train
| false
|
48,421
|
def random_partition(n, n_data):
all_idxs = numpy.arange(n_data)
numpy.random.shuffle(all_idxs)
idxs1 = all_idxs[:n]
idxs2 = all_idxs[n:]
return (idxs1, idxs2)
|
[
"def",
"random_partition",
"(",
"n",
",",
"n_data",
")",
":",
"all_idxs",
"=",
"numpy",
".",
"arange",
"(",
"n_data",
")",
"numpy",
".",
"random",
".",
"shuffle",
"(",
"all_idxs",
")",
"idxs1",
"=",
"all_idxs",
"[",
":",
"n",
"]",
"idxs2",
"=",
"all_idxs",
"[",
"n",
":",
"]",
"return",
"(",
"idxs1",
",",
"idxs2",
")"
] |
return n random rows of data (and also the other len-n rows) .
|
train
| true
|
48,422
|
def plot_bar_graphs(ax, prng, min_value=5, max_value=25, nb_samples=5):
x = np.arange(nb_samples)
(ya, yb) = prng.randint(min_value, max_value, size=(2, nb_samples))
width = 0.25
ax.bar(x, ya, width)
ax.bar((x + width), yb, width, color='C2')
ax.set_xticks((x + width))
ax.set_xticklabels(['a', 'b', 'c', 'd', 'e'])
return ax
|
[
"def",
"plot_bar_graphs",
"(",
"ax",
",",
"prng",
",",
"min_value",
"=",
"5",
",",
"max_value",
"=",
"25",
",",
"nb_samples",
"=",
"5",
")",
":",
"x",
"=",
"np",
".",
"arange",
"(",
"nb_samples",
")",
"(",
"ya",
",",
"yb",
")",
"=",
"prng",
".",
"randint",
"(",
"min_value",
",",
"max_value",
",",
"size",
"=",
"(",
"2",
",",
"nb_samples",
")",
")",
"width",
"=",
"0.25",
"ax",
".",
"bar",
"(",
"x",
",",
"ya",
",",
"width",
")",
"ax",
".",
"bar",
"(",
"(",
"x",
"+",
"width",
")",
",",
"yb",
",",
"width",
",",
"color",
"=",
"'C2'",
")",
"ax",
".",
"set_xticks",
"(",
"(",
"x",
"+",
"width",
")",
")",
"ax",
".",
"set_xticklabels",
"(",
"[",
"'a'",
",",
"'b'",
",",
"'c'",
",",
"'d'",
",",
"'e'",
"]",
")",
"return",
"ax"
] |
plot two bar graphs side by side .
|
train
| false
|
48,423
|
def kwarg(**kwargs):
return kwargs
|
[
"def",
"kwarg",
"(",
"**",
"kwargs",
")",
":",
"return",
"kwargs"
] |
print out the data passed into the function **kwargs .
|
train
| false
|
48,426
|
def get_google_api_auth(module, scopes=[], user_agent_product='ansible-python-api', user_agent_version='NA'):
if (not HAS_GOOGLE_API_LIB):
module.fail_json(msg='Please install google-api-python-client library')
if (not scopes):
scopes = ['https://www.googleapis.com/auth/cloud-platform']
try:
(credentials, conn_params) = get_google_credentials(module, scopes, require_valid_json=True, check_libcloud=False)
http = set_user_agent(Http(), ('%s-%s' % (user_agent_product, user_agent_version)))
http_auth = credentials.authorize(http)
return (http_auth, conn_params)
except Exception as e:
module.fail_json(msg=unexpected_error_msg(e), changed=False)
return (None, None)
|
[
"def",
"get_google_api_auth",
"(",
"module",
",",
"scopes",
"=",
"[",
"]",
",",
"user_agent_product",
"=",
"'ansible-python-api'",
",",
"user_agent_version",
"=",
"'NA'",
")",
":",
"if",
"(",
"not",
"HAS_GOOGLE_API_LIB",
")",
":",
"module",
".",
"fail_json",
"(",
"msg",
"=",
"'Please install google-api-python-client library'",
")",
"if",
"(",
"not",
"scopes",
")",
":",
"scopes",
"=",
"[",
"'https://www.googleapis.com/auth/cloud-platform'",
"]",
"try",
":",
"(",
"credentials",
",",
"conn_params",
")",
"=",
"get_google_credentials",
"(",
"module",
",",
"scopes",
",",
"require_valid_json",
"=",
"True",
",",
"check_libcloud",
"=",
"False",
")",
"http",
"=",
"set_user_agent",
"(",
"Http",
"(",
")",
",",
"(",
"'%s-%s'",
"%",
"(",
"user_agent_product",
",",
"user_agent_version",
")",
")",
")",
"http_auth",
"=",
"credentials",
".",
"authorize",
"(",
"http",
")",
"return",
"(",
"http_auth",
",",
"conn_params",
")",
"except",
"Exception",
"as",
"e",
":",
"module",
".",
"fail_json",
"(",
"msg",
"=",
"unexpected_error_msg",
"(",
"e",
")",
",",
"changed",
"=",
"False",
")",
"return",
"(",
"None",
",",
"None",
")"
] |
authentication for use with google-python-api-client .
|
train
| false
|
48,427
|
@task
def test_pypi(release='2'):
version = get_sympy_version()
release = str(release)
if (release not in {'2', '3'}):
raise ValueError(("release must be one of '2', '3', not %s" % release))
venv = '/home/vagrant/repos/test-{release}-pip-virtualenv'.format(release=release)
with use_venv(release):
make_virtualenv(venv)
with virtualenv(venv):
run('pip install sympy')
run('python -c "import sympy; assert sympy.__version__ == \'{version}\'"'.format(version=version))
|
[
"@",
"task",
"def",
"test_pypi",
"(",
"release",
"=",
"'2'",
")",
":",
"version",
"=",
"get_sympy_version",
"(",
")",
"release",
"=",
"str",
"(",
"release",
")",
"if",
"(",
"release",
"not",
"in",
"{",
"'2'",
",",
"'3'",
"}",
")",
":",
"raise",
"ValueError",
"(",
"(",
"\"release must be one of '2', '3', not %s\"",
"%",
"release",
")",
")",
"venv",
"=",
"'/home/vagrant/repos/test-{release}-pip-virtualenv'",
".",
"format",
"(",
"release",
"=",
"release",
")",
"with",
"use_venv",
"(",
"release",
")",
":",
"make_virtualenv",
"(",
"venv",
")",
"with",
"virtualenv",
"(",
"venv",
")",
":",
"run",
"(",
"'pip install sympy'",
")",
"run",
"(",
"'python -c \"import sympy; assert sympy.__version__ == \\'{version}\\'\"'",
".",
"format",
"(",
"version",
"=",
"version",
")",
")"
] |
test that the sympy can be pip installed .
|
train
| false
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.