id_within_dataset
int64 1
55.5k
| snippet
stringlengths 19
14.2k
| tokens
listlengths 6
1.63k
| nl
stringlengths 6
352
| split_within_dataset
stringclasses 1
value | is_duplicated
bool 2
classes |
|---|---|---|---|---|---|
51,857
|
def _get_old_package_license_ids(migrate_engine):
old_ids = {}
select_licenses = 'SELECT id, license_id FROM package;'
q = migrate_engine.execute(select_licenses)
for (id, license_id) in q:
old_ids[id] = license_id
return old_ids
|
[
"def",
"_get_old_package_license_ids",
"(",
"migrate_engine",
")",
":",
"old_ids",
"=",
"{",
"}",
"select_licenses",
"=",
"'SELECT id, license_id FROM package;'",
"q",
"=",
"migrate_engine",
".",
"execute",
"(",
"select_licenses",
")",
"for",
"(",
"id",
",",
"license_id",
")",
"in",
"q",
":",
"old_ids",
"[",
"id",
"]",
"=",
"license_id",
"return",
"old_ids"
] |
returns a dict of old license ids .
|
train
| false
|
51,860
|
def get_disks(vm_):
with _get_xapi_session() as xapi:
disk = {}
vm_uuid = _get_label_uuid(xapi, 'VM', vm_)
if (vm_uuid is False):
return False
for vbd in xapi.VM.get_VBDs(vm_uuid):
dev = xapi.VBD.get_device(vbd)
if (not dev):
continue
prop = xapi.VBD.get_runtime_properties(vbd)
disk[dev] = {'backend': prop['backend'], 'type': prop['device-type'], 'protocol': prop['protocol']}
return disk
|
[
"def",
"get_disks",
"(",
"vm_",
")",
":",
"with",
"_get_xapi_session",
"(",
")",
"as",
"xapi",
":",
"disk",
"=",
"{",
"}",
"vm_uuid",
"=",
"_get_label_uuid",
"(",
"xapi",
",",
"'VM'",
",",
"vm_",
")",
"if",
"(",
"vm_uuid",
"is",
"False",
")",
":",
"return",
"False",
"for",
"vbd",
"in",
"xapi",
".",
"VM",
".",
"get_VBDs",
"(",
"vm_uuid",
")",
":",
"dev",
"=",
"xapi",
".",
"VBD",
".",
"get_device",
"(",
"vbd",
")",
"if",
"(",
"not",
"dev",
")",
":",
"continue",
"prop",
"=",
"xapi",
".",
"VBD",
".",
"get_runtime_properties",
"(",
"vbd",
")",
"disk",
"[",
"dev",
"]",
"=",
"{",
"'backend'",
":",
"prop",
"[",
"'backend'",
"]",
",",
"'type'",
":",
"prop",
"[",
"'device-type'",
"]",
",",
"'protocol'",
":",
"prop",
"[",
"'protocol'",
"]",
"}",
"return",
"disk"
] |
return the disks of a named vm cli example: .
|
train
| true
|
51,861
|
def decode_der(obj_class, binstr):
der = obj_class()
der.decode(binstr)
return der
|
[
"def",
"decode_der",
"(",
"obj_class",
",",
"binstr",
")",
":",
"der",
"=",
"obj_class",
"(",
")",
"der",
".",
"decode",
"(",
"binstr",
")",
"return",
"der"
] |
instantiate a der object class .
|
train
| false
|
51,862
|
def _instrumented_test_render(self, *args, **data):
with _MAKO_LOCK:
def mako_callable_(context, *args, **kwargs):
template_rendered.send(sender=self, template=self, context=context)
return self.original_callable_[(-1)](context, *args, **kwargs)
if hasattr(self, 'original_callable_'):
self.original_callable_.append(self.callable_)
else:
self.original_callable_ = [self.callable_]
self.callable_ = mako_callable_
try:
response = runtime._render(self, self.original_callable_[(-1)], args, data)
finally:
self.callable_ = self.original_callable_.pop()
return response
|
[
"def",
"_instrumented_test_render",
"(",
"self",
",",
"*",
"args",
",",
"**",
"data",
")",
":",
"with",
"_MAKO_LOCK",
":",
"def",
"mako_callable_",
"(",
"context",
",",
"*",
"args",
",",
"**",
"kwargs",
")",
":",
"template_rendered",
".",
"send",
"(",
"sender",
"=",
"self",
",",
"template",
"=",
"self",
",",
"context",
"=",
"context",
")",
"return",
"self",
".",
"original_callable_",
"[",
"(",
"-",
"1",
")",
"]",
"(",
"context",
",",
"*",
"args",
",",
"**",
"kwargs",
")",
"if",
"hasattr",
"(",
"self",
",",
"'original_callable_'",
")",
":",
"self",
".",
"original_callable_",
".",
"append",
"(",
"self",
".",
"callable_",
")",
"else",
":",
"self",
".",
"original_callable_",
"=",
"[",
"self",
".",
"callable_",
"]",
"self",
".",
"callable_",
"=",
"mako_callable_",
"try",
":",
"response",
"=",
"runtime",
".",
"_render",
"(",
"self",
",",
"self",
".",
"original_callable_",
"[",
"(",
"-",
"1",
")",
"]",
",",
"args",
",",
"data",
")",
"finally",
":",
"self",
".",
"callable_",
"=",
"self",
".",
"original_callable_",
".",
"pop",
"(",
")",
"return",
"response"
] |
an instrumented template render method .
|
train
| false
|
51,864
|
def find_full_path(path_to_file):
for (subdir, dirs, files) in os.walk('.'):
full = os.path.relpath(os.path.join(subdir, path_to_file))
if os.path.exists(full):
return full
|
[
"def",
"find_full_path",
"(",
"path_to_file",
")",
":",
"for",
"(",
"subdir",
",",
"dirs",
",",
"files",
")",
"in",
"os",
".",
"walk",
"(",
"'.'",
")",
":",
"full",
"=",
"os",
".",
"path",
".",
"relpath",
"(",
"os",
".",
"path",
".",
"join",
"(",
"subdir",
",",
"path_to_file",
")",
")",
"if",
"os",
".",
"path",
".",
"exists",
"(",
"full",
")",
":",
"return",
"full"
] |
find the full path where we only have a relative path from somewhere in the tree .
|
train
| false
|
51,867
|
def aifile_list(request, page=None):
if (not test_user_authenticated(request)):
return login(request, next='/cobbler_web/aifile/list', expired=True)
aifiles = remote.get_autoinstall_templates(request.session['token'])
aifile_list = []
for aifile in aifiles:
aifile_list.append((aifile, 'editable'))
t = get_template('aifile_list.tmpl')
html = t.render(RequestContext(request, {'what': 'aifile', 'ai_files': aifile_list, 'version': remote.extended_version(request.session['token'])['version'], 'username': username, 'item_count': len(aifile_list[0])}))
return HttpResponse(html)
|
[
"def",
"aifile_list",
"(",
"request",
",",
"page",
"=",
"None",
")",
":",
"if",
"(",
"not",
"test_user_authenticated",
"(",
"request",
")",
")",
":",
"return",
"login",
"(",
"request",
",",
"next",
"=",
"'/cobbler_web/aifile/list'",
",",
"expired",
"=",
"True",
")",
"aifiles",
"=",
"remote",
".",
"get_autoinstall_templates",
"(",
"request",
".",
"session",
"[",
"'token'",
"]",
")",
"aifile_list",
"=",
"[",
"]",
"for",
"aifile",
"in",
"aifiles",
":",
"aifile_list",
".",
"append",
"(",
"(",
"aifile",
",",
"'editable'",
")",
")",
"t",
"=",
"get_template",
"(",
"'aifile_list.tmpl'",
")",
"html",
"=",
"t",
".",
"render",
"(",
"RequestContext",
"(",
"request",
",",
"{",
"'what'",
":",
"'aifile'",
",",
"'ai_files'",
":",
"aifile_list",
",",
"'version'",
":",
"remote",
".",
"extended_version",
"(",
"request",
".",
"session",
"[",
"'token'",
"]",
")",
"[",
"'version'",
"]",
",",
"'username'",
":",
"username",
",",
"'item_count'",
":",
"len",
"(",
"aifile_list",
"[",
"0",
"]",
")",
"}",
")",
")",
"return",
"HttpResponse",
"(",
"html",
")"
] |
list all automatic os installation templates and link to their edit pages .
|
train
| false
|
51,868
|
def upload_project(local_dir=None, remote_dir='', use_sudo=False):
runner = ((use_sudo and sudo) or run)
local_dir = (local_dir or os.getcwd())
local_dir = local_dir.rstrip(os.sep)
(local_path, local_name) = os.path.split(local_dir)
tar_file = ('%s.tar.gz' % local_name)
target_tar = os.path.join(remote_dir, tar_file)
tmp_folder = mkdtemp()
try:
tar_path = os.path.join(tmp_folder, tar_file)
local(('tar -czf %s -C %s %s' % (tar_path, local_path, local_name)))
put(tar_path, target_tar, use_sudo=use_sudo)
with cd(remote_dir):
try:
runner(('tar -xzf %s' % tar_file))
finally:
runner(('rm -f %s' % tar_file))
finally:
local(('rm -rf %s' % tmp_folder))
|
[
"def",
"upload_project",
"(",
"local_dir",
"=",
"None",
",",
"remote_dir",
"=",
"''",
",",
"use_sudo",
"=",
"False",
")",
":",
"runner",
"=",
"(",
"(",
"use_sudo",
"and",
"sudo",
")",
"or",
"run",
")",
"local_dir",
"=",
"(",
"local_dir",
"or",
"os",
".",
"getcwd",
"(",
")",
")",
"local_dir",
"=",
"local_dir",
".",
"rstrip",
"(",
"os",
".",
"sep",
")",
"(",
"local_path",
",",
"local_name",
")",
"=",
"os",
".",
"path",
".",
"split",
"(",
"local_dir",
")",
"tar_file",
"=",
"(",
"'%s.tar.gz'",
"%",
"local_name",
")",
"target_tar",
"=",
"os",
".",
"path",
".",
"join",
"(",
"remote_dir",
",",
"tar_file",
")",
"tmp_folder",
"=",
"mkdtemp",
"(",
")",
"try",
":",
"tar_path",
"=",
"os",
".",
"path",
".",
"join",
"(",
"tmp_folder",
",",
"tar_file",
")",
"local",
"(",
"(",
"'tar -czf %s -C %s %s'",
"%",
"(",
"tar_path",
",",
"local_path",
",",
"local_name",
")",
")",
")",
"put",
"(",
"tar_path",
",",
"target_tar",
",",
"use_sudo",
"=",
"use_sudo",
")",
"with",
"cd",
"(",
"remote_dir",
")",
":",
"try",
":",
"runner",
"(",
"(",
"'tar -xzf %s'",
"%",
"tar_file",
")",
")",
"finally",
":",
"runner",
"(",
"(",
"'rm -f %s'",
"%",
"tar_file",
")",
")",
"finally",
":",
"local",
"(",
"(",
"'rm -rf %s'",
"%",
"tmp_folder",
")",
")"
] |
upload the current project to a remote system via tar/gzip .
|
train
| false
|
51,869
|
def _TestAuthViewfinderUser(action, tester, user_dict, device_dict=None, user_cookie=None):
if ('email' in user_dict):
ident_dict = {'key': ('Email:%s' % user_dict['email']), 'authority': 'Viewfinder'}
else:
ident_dict = {'key': ('Phone:%s' % user_dict['phone']), 'authority': 'Viewfinder'}
response = _AuthViewfinderUser(tester, action, user_dict, ident_dict, device_dict, user_cookie)
cookie_user_dict = tester.DecodeUserCookie(tester.GetCookieFromResponse(response))
if ((action != 'link') and (not ((action == 'login') and ('password' in user_dict)))):
assert ('confirm_time' in cookie_user_dict), cookie_user_dict
else:
assert ('confirm_time' not in cookie_user_dict), cookie_user_dict
return auth_test._ValidateAuthUser(tester, action, user_dict, ident_dict, device_dict, user_cookie, response)
|
[
"def",
"_TestAuthViewfinderUser",
"(",
"action",
",",
"tester",
",",
"user_dict",
",",
"device_dict",
"=",
"None",
",",
"user_cookie",
"=",
"None",
")",
":",
"if",
"(",
"'email'",
"in",
"user_dict",
")",
":",
"ident_dict",
"=",
"{",
"'key'",
":",
"(",
"'Email:%s'",
"%",
"user_dict",
"[",
"'email'",
"]",
")",
",",
"'authority'",
":",
"'Viewfinder'",
"}",
"else",
":",
"ident_dict",
"=",
"{",
"'key'",
":",
"(",
"'Phone:%s'",
"%",
"user_dict",
"[",
"'phone'",
"]",
")",
",",
"'authority'",
":",
"'Viewfinder'",
"}",
"response",
"=",
"_AuthViewfinderUser",
"(",
"tester",
",",
"action",
",",
"user_dict",
",",
"ident_dict",
",",
"device_dict",
",",
"user_cookie",
")",
"cookie_user_dict",
"=",
"tester",
".",
"DecodeUserCookie",
"(",
"tester",
".",
"GetCookieFromResponse",
"(",
"response",
")",
")",
"if",
"(",
"(",
"action",
"!=",
"'link'",
")",
"and",
"(",
"not",
"(",
"(",
"action",
"==",
"'login'",
")",
"and",
"(",
"'password'",
"in",
"user_dict",
")",
")",
")",
")",
":",
"assert",
"(",
"'confirm_time'",
"in",
"cookie_user_dict",
")",
",",
"cookie_user_dict",
"else",
":",
"assert",
"(",
"'confirm_time'",
"not",
"in",
"cookie_user_dict",
")",
",",
"cookie_user_dict",
"return",
"auth_test",
".",
"_ValidateAuthUser",
"(",
"tester",
",",
"action",
",",
"user_dict",
",",
"ident_dict",
",",
"device_dict",
",",
"user_cookie",
",",
"response",
")"
] |
called by the servicetester in order to test login/viewfinder .
|
train
| false
|
51,870
|
def start_cycle():
dg['cyc'] = init_cycle()
dg['cache'] = {}
dg['humanize_unsupported'] = False
|
[
"def",
"start_cycle",
"(",
")",
":",
"dg",
"[",
"'cyc'",
"]",
"=",
"init_cycle",
"(",
")",
"dg",
"[",
"'cache'",
"]",
"=",
"{",
"}",
"dg",
"[",
"'humanize_unsupported'",
"]",
"=",
"False"
] |
notify from rainbow .
|
train
| false
|
51,871
|
@synchronized(NZB_LOCK)
def backup_exists(filename):
path = cfg.nzb_backup_dir.get_path()
return (path and os.path.exists(os.path.join(path, (filename + '.gz'))))
|
[
"@",
"synchronized",
"(",
"NZB_LOCK",
")",
"def",
"backup_exists",
"(",
"filename",
")",
":",
"path",
"=",
"cfg",
".",
"nzb_backup_dir",
".",
"get_path",
"(",
")",
"return",
"(",
"path",
"and",
"os",
".",
"path",
".",
"exists",
"(",
"os",
".",
"path",
".",
"join",
"(",
"path",
",",
"(",
"filename",
"+",
"'.gz'",
")",
")",
")",
")"
] |
return true if backup exists and no_dupes is set .
|
train
| false
|
51,873
|
@missing_name
def decoratedFunction():
return None
|
[
"@",
"missing_name",
"def",
"decoratedFunction",
"(",
")",
":",
"return",
"None"
] |
decoratedfunction docstring .
|
train
| false
|
51,874
|
def krogh_interpolate(xi, yi, x, der=0, axis=0):
P = KroghInterpolator(xi, yi, axis=axis)
if (der == 0):
return P(x)
elif _isscalar(der):
return P.derivative(x, der=der)
else:
return P.derivatives(x, der=(np.amax(der) + 1))[der]
|
[
"def",
"krogh_interpolate",
"(",
"xi",
",",
"yi",
",",
"x",
",",
"der",
"=",
"0",
",",
"axis",
"=",
"0",
")",
":",
"P",
"=",
"KroghInterpolator",
"(",
"xi",
",",
"yi",
",",
"axis",
"=",
"axis",
")",
"if",
"(",
"der",
"==",
"0",
")",
":",
"return",
"P",
"(",
"x",
")",
"elif",
"_isscalar",
"(",
"der",
")",
":",
"return",
"P",
".",
"derivative",
"(",
"x",
",",
"der",
"=",
"der",
")",
"else",
":",
"return",
"P",
".",
"derivatives",
"(",
"x",
",",
"der",
"=",
"(",
"np",
".",
"amax",
"(",
"der",
")",
"+",
"1",
")",
")",
"[",
"der",
"]"
] |
convenience function for polynomial interpolation .
|
train
| false
|
51,875
|
def make_violin_rugplot(vals, pdf_max, distance, color='#1f77b4'):
return graph_objs.Scatter(y=vals, x=([((- pdf_max) - distance)] * len(vals)), marker=graph_objs.Marker(color=color, symbol='line-ew-open'), mode='markers', name='', showlegend=False, hoverinfo='y')
|
[
"def",
"make_violin_rugplot",
"(",
"vals",
",",
"pdf_max",
",",
"distance",
",",
"color",
"=",
"'#1f77b4'",
")",
":",
"return",
"graph_objs",
".",
"Scatter",
"(",
"y",
"=",
"vals",
",",
"x",
"=",
"(",
"[",
"(",
"(",
"-",
"pdf_max",
")",
"-",
"distance",
")",
"]",
"*",
"len",
"(",
"vals",
")",
")",
",",
"marker",
"=",
"graph_objs",
".",
"Marker",
"(",
"color",
"=",
"color",
",",
"symbol",
"=",
"'line-ew-open'",
")",
",",
"mode",
"=",
"'markers'",
",",
"name",
"=",
"''",
",",
"showlegend",
"=",
"False",
",",
"hoverinfo",
"=",
"'y'",
")"
] |
returns a rugplot fig for a violin plot .
|
train
| false
|
51,876
|
def update_downtime(guest, instance, olddowntime, downtime_steps, elapsed):
LOG.debug('Current %(dt)s elapsed %(elapsed)d steps %(steps)s', {'dt': olddowntime, 'elapsed': elapsed, 'steps': downtime_steps}, instance=instance)
thisstep = None
for step in downtime_steps:
if (elapsed > step[0]):
thisstep = step
if (thisstep is None):
LOG.debug('No current step', instance=instance)
return olddowntime
if (thisstep[1] == olddowntime):
LOG.debug('Downtime does not need to change', instance=instance)
return olddowntime
LOG.info(_LI('Increasing downtime to %(downtime)d ms after %(waittime)d sec elapsed time'), {'downtime': thisstep[1], 'waittime': thisstep[0]}, instance=instance)
try:
guest.migrate_configure_max_downtime(thisstep[1])
except libvirt.libvirtError as e:
LOG.warning(_LW('Unable to increase max downtime to %(time)dms: %(e)s'), {'time': thisstep[1], 'e': e}, instance=instance)
return thisstep[1]
|
[
"def",
"update_downtime",
"(",
"guest",
",",
"instance",
",",
"olddowntime",
",",
"downtime_steps",
",",
"elapsed",
")",
":",
"LOG",
".",
"debug",
"(",
"'Current %(dt)s elapsed %(elapsed)d steps %(steps)s'",
",",
"{",
"'dt'",
":",
"olddowntime",
",",
"'elapsed'",
":",
"elapsed",
",",
"'steps'",
":",
"downtime_steps",
"}",
",",
"instance",
"=",
"instance",
")",
"thisstep",
"=",
"None",
"for",
"step",
"in",
"downtime_steps",
":",
"if",
"(",
"elapsed",
">",
"step",
"[",
"0",
"]",
")",
":",
"thisstep",
"=",
"step",
"if",
"(",
"thisstep",
"is",
"None",
")",
":",
"LOG",
".",
"debug",
"(",
"'No current step'",
",",
"instance",
"=",
"instance",
")",
"return",
"olddowntime",
"if",
"(",
"thisstep",
"[",
"1",
"]",
"==",
"olddowntime",
")",
":",
"LOG",
".",
"debug",
"(",
"'Downtime does not need to change'",
",",
"instance",
"=",
"instance",
")",
"return",
"olddowntime",
"LOG",
".",
"info",
"(",
"_LI",
"(",
"'Increasing downtime to %(downtime)d ms after %(waittime)d sec elapsed time'",
")",
",",
"{",
"'downtime'",
":",
"thisstep",
"[",
"1",
"]",
",",
"'waittime'",
":",
"thisstep",
"[",
"0",
"]",
"}",
",",
"instance",
"=",
"instance",
")",
"try",
":",
"guest",
".",
"migrate_configure_max_downtime",
"(",
"thisstep",
"[",
"1",
"]",
")",
"except",
"libvirt",
".",
"libvirtError",
"as",
"e",
":",
"LOG",
".",
"warning",
"(",
"_LW",
"(",
"'Unable to increase max downtime to %(time)dms: %(e)s'",
")",
",",
"{",
"'time'",
":",
"thisstep",
"[",
"1",
"]",
",",
"'e'",
":",
"e",
"}",
",",
"instance",
"=",
"instance",
")",
"return",
"thisstep",
"[",
"1",
"]"
] |
update max downtime if needed .
|
train
| false
|
51,878
|
def cm(value):
return dpi2px(value, 'cm')
|
[
"def",
"cm",
"(",
"value",
")",
":",
"return",
"dpi2px",
"(",
"value",
",",
"'cm'",
")"
] |
convert from centimeters to pixels .
|
train
| false
|
51,879
|
def get_br_int_port_name(prefix, port_id):
return ('%si-%s' % (prefix, port_id))[:constants.DEVICE_NAME_MAX_LEN]
|
[
"def",
"get_br_int_port_name",
"(",
"prefix",
",",
"port_id",
")",
":",
"return",
"(",
"'%si-%s'",
"%",
"(",
"prefix",
",",
"port_id",
")",
")",
"[",
":",
"constants",
".",
"DEVICE_NAME_MAX_LEN",
"]"
] |
return the ovs port name for the given port id .
|
train
| false
|
51,882
|
def discretize_integrate_1D(model, x_range):
from scipy.integrate import quad
x = np.arange((x_range[0] - 0.5), (x_range[1] + 0.5))
values = np.array([])
for i in range((x.size - 1)):
values = np.append(values, quad(model, x[i], x[(i + 1)])[0])
return values
|
[
"def",
"discretize_integrate_1D",
"(",
"model",
",",
"x_range",
")",
":",
"from",
"scipy",
".",
"integrate",
"import",
"quad",
"x",
"=",
"np",
".",
"arange",
"(",
"(",
"x_range",
"[",
"0",
"]",
"-",
"0.5",
")",
",",
"(",
"x_range",
"[",
"1",
"]",
"+",
"0.5",
")",
")",
"values",
"=",
"np",
".",
"array",
"(",
"[",
"]",
")",
"for",
"i",
"in",
"range",
"(",
"(",
"x",
".",
"size",
"-",
"1",
")",
")",
":",
"values",
"=",
"np",
".",
"append",
"(",
"values",
",",
"quad",
"(",
"model",
",",
"x",
"[",
"i",
"]",
",",
"x",
"[",
"(",
"i",
"+",
"1",
")",
"]",
")",
"[",
"0",
"]",
")",
"return",
"values"
] |
discretize model by integrating numerically the model over the bin .
|
train
| false
|
51,885
|
def setup_streams_fixtures(testcase):
testcase.mock_tracker = scaffold.MockTracker()
testcase.stream_file_paths = dict(stdin=tempfile.mktemp(), stdout=tempfile.mktemp(), stderr=tempfile.mktemp())
testcase.stream_files_by_name = dict(((name, FakeFileDescriptorStringIO()) for name in ['stdin', 'stdout', 'stderr']))
testcase.stream_files_by_path = dict(((testcase.stream_file_paths[name], testcase.stream_files_by_name[name]) for name in ['stdin', 'stdout', 'stderr']))
scaffold.mock('os.dup2', tracker=testcase.mock_tracker)
|
[
"def",
"setup_streams_fixtures",
"(",
"testcase",
")",
":",
"testcase",
".",
"mock_tracker",
"=",
"scaffold",
".",
"MockTracker",
"(",
")",
"testcase",
".",
"stream_file_paths",
"=",
"dict",
"(",
"stdin",
"=",
"tempfile",
".",
"mktemp",
"(",
")",
",",
"stdout",
"=",
"tempfile",
".",
"mktemp",
"(",
")",
",",
"stderr",
"=",
"tempfile",
".",
"mktemp",
"(",
")",
")",
"testcase",
".",
"stream_files_by_name",
"=",
"dict",
"(",
"(",
"(",
"name",
",",
"FakeFileDescriptorStringIO",
"(",
")",
")",
"for",
"name",
"in",
"[",
"'stdin'",
",",
"'stdout'",
",",
"'stderr'",
"]",
")",
")",
"testcase",
".",
"stream_files_by_path",
"=",
"dict",
"(",
"(",
"(",
"testcase",
".",
"stream_file_paths",
"[",
"name",
"]",
",",
"testcase",
".",
"stream_files_by_name",
"[",
"name",
"]",
")",
"for",
"name",
"in",
"[",
"'stdin'",
",",
"'stdout'",
",",
"'stderr'",
"]",
")",
")",
"scaffold",
".",
"mock",
"(",
"'os.dup2'",
",",
"tracker",
"=",
"testcase",
".",
"mock_tracker",
")"
] |
set up common test fixtures for standard streams .
|
train
| false
|
51,886
|
def update_vouch_flags_action():
def update_vouch_flags(modeladmin, request, queryset):
for profile in queryset:
vouches_received = profile.vouches_received.count()
profile.can_vouch = (vouches_received >= settings.CAN_VOUCH_THRESHOLD)
profile.is_vouched = (vouches_received > 0)
profile.save()
update_vouch_flags.short_description = 'Update vouch flags'
return update_vouch_flags
|
[
"def",
"update_vouch_flags_action",
"(",
")",
":",
"def",
"update_vouch_flags",
"(",
"modeladmin",
",",
"request",
",",
"queryset",
")",
":",
"for",
"profile",
"in",
"queryset",
":",
"vouches_received",
"=",
"profile",
".",
"vouches_received",
".",
"count",
"(",
")",
"profile",
".",
"can_vouch",
"=",
"(",
"vouches_received",
">=",
"settings",
".",
"CAN_VOUCH_THRESHOLD",
")",
"profile",
".",
"is_vouched",
"=",
"(",
"vouches_received",
">",
"0",
")",
"profile",
".",
"save",
"(",
")",
"update_vouch_flags",
".",
"short_description",
"=",
"'Update vouch flags'",
"return",
"update_vouch_flags"
] |
update can_vouch .
|
train
| false
|
51,889
|
def md5sum_file(file_name):
f = open(file_name)
digest = md5sum_str(f.read())
f.close()
return digest
|
[
"def",
"md5sum_file",
"(",
"file_name",
")",
":",
"f",
"=",
"open",
"(",
"file_name",
")",
"digest",
"=",
"md5sum_str",
"(",
"f",
".",
"read",
"(",
")",
")",
"f",
".",
"close",
"(",
")",
"return",
"digest"
] |
calculate md5sum of the file .
|
train
| false
|
51,890
|
def _zero_volume(path, volume_size):
bs = units.Mi
direct_flags = ('oflag=direct',)
sync_flags = ()
remaining_bytes = volume_size
while remaining_bytes:
zero_blocks = (remaining_bytes // bs)
seek_blocks = ((volume_size - remaining_bytes) // bs)
zero_cmd = ('dd', ('bs=%s' % bs), 'if=/dev/zero', ('of=%s' % path), ('seek=%s' % seek_blocks), ('count=%s' % zero_blocks))
zero_cmd += direct_flags
zero_cmd += sync_flags
if zero_blocks:
utils.execute(run_as_root=True, *zero_cmd)
remaining_bytes %= bs
bs //= units.Ki
direct_flags = ()
sync_flags = ('conv=fdatasync',)
|
[
"def",
"_zero_volume",
"(",
"path",
",",
"volume_size",
")",
":",
"bs",
"=",
"units",
".",
"Mi",
"direct_flags",
"=",
"(",
"'oflag=direct'",
",",
")",
"sync_flags",
"=",
"(",
")",
"remaining_bytes",
"=",
"volume_size",
"while",
"remaining_bytes",
":",
"zero_blocks",
"=",
"(",
"remaining_bytes",
"//",
"bs",
")",
"seek_blocks",
"=",
"(",
"(",
"volume_size",
"-",
"remaining_bytes",
")",
"//",
"bs",
")",
"zero_cmd",
"=",
"(",
"'dd'",
",",
"(",
"'bs=%s'",
"%",
"bs",
")",
",",
"'if=/dev/zero'",
",",
"(",
"'of=%s'",
"%",
"path",
")",
",",
"(",
"'seek=%s'",
"%",
"seek_blocks",
")",
",",
"(",
"'count=%s'",
"%",
"zero_blocks",
")",
")",
"zero_cmd",
"+=",
"direct_flags",
"zero_cmd",
"+=",
"sync_flags",
"if",
"zero_blocks",
":",
"utils",
".",
"execute",
"(",
"run_as_root",
"=",
"True",
",",
"*",
"zero_cmd",
")",
"remaining_bytes",
"%=",
"bs",
"bs",
"//=",
"units",
".",
"Ki",
"direct_flags",
"=",
"(",
")",
"sync_flags",
"=",
"(",
"'conv=fdatasync'",
",",
")"
] |
write zeros over the specified path .
|
train
| false
|
51,891
|
def attr_ne(accessing_obj, accessed_obj, *args, **kwargs):
return attr(accessing_obj, accessed_obj, *args, **{'compare': 'ne'})
|
[
"def",
"attr_ne",
"(",
"accessing_obj",
",",
"accessed_obj",
",",
"*",
"args",
",",
"**",
"kwargs",
")",
":",
"return",
"attr",
"(",
"accessing_obj",
",",
"accessed_obj",
",",
"*",
"args",
",",
"**",
"{",
"'compare'",
":",
"'ne'",
"}",
")"
] |
usage: attr_gt only true if access_objs attribute != the value given .
|
train
| false
|
51,893
|
@cache_permission
def can_see_repository_status(user, project):
return (can_commit_translation(user, project) or can_update_translation(user, project))
|
[
"@",
"cache_permission",
"def",
"can_see_repository_status",
"(",
"user",
",",
"project",
")",
":",
"return",
"(",
"can_commit_translation",
"(",
"user",
",",
"project",
")",
"or",
"can_update_translation",
"(",
"user",
",",
"project",
")",
")"
] |
checks whether user can view repository status .
|
train
| false
|
51,897
|
def tidy_cli(arguments):
TidyCommandLine().execute_cli(arguments)
|
[
"def",
"tidy_cli",
"(",
"arguments",
")",
":",
"TidyCommandLine",
"(",
")",
".",
"execute_cli",
"(",
"arguments",
")"
] |
executes tidy similarly as from the command line .
|
train
| false
|
51,898
|
def _resolve_looppart(parts, asspath, context):
asspath = asspath[:]
index = asspath.pop(0)
for part in parts:
if (part is YES):
continue
if (not hasattr(part, 'itered')):
continue
try:
itered = part.itered()
except TypeError:
continue
for stmt in itered:
try:
assigned = stmt.getitem(index, context)
except (AttributeError, IndexError):
continue
except TypeError:
continue
if (not asspath):
(yield assigned)
elif (assigned is YES):
break
else:
try:
for infered in _resolve_looppart(assigned.infer(context), asspath, context):
(yield infered)
except InferenceError:
break
|
[
"def",
"_resolve_looppart",
"(",
"parts",
",",
"asspath",
",",
"context",
")",
":",
"asspath",
"=",
"asspath",
"[",
":",
"]",
"index",
"=",
"asspath",
".",
"pop",
"(",
"0",
")",
"for",
"part",
"in",
"parts",
":",
"if",
"(",
"part",
"is",
"YES",
")",
":",
"continue",
"if",
"(",
"not",
"hasattr",
"(",
"part",
",",
"'itered'",
")",
")",
":",
"continue",
"try",
":",
"itered",
"=",
"part",
".",
"itered",
"(",
")",
"except",
"TypeError",
":",
"continue",
"for",
"stmt",
"in",
"itered",
":",
"try",
":",
"assigned",
"=",
"stmt",
".",
"getitem",
"(",
"index",
",",
"context",
")",
"except",
"(",
"AttributeError",
",",
"IndexError",
")",
":",
"continue",
"except",
"TypeError",
":",
"continue",
"if",
"(",
"not",
"asspath",
")",
":",
"(",
"yield",
"assigned",
")",
"elif",
"(",
"assigned",
"is",
"YES",
")",
":",
"break",
"else",
":",
"try",
":",
"for",
"infered",
"in",
"_resolve_looppart",
"(",
"assigned",
".",
"infer",
"(",
"context",
")",
",",
"asspath",
",",
"context",
")",
":",
"(",
"yield",
"infered",
")",
"except",
"InferenceError",
":",
"break"
] |
recursive function to resolve multiple assignments on loops .
|
train
| false
|
51,899
|
def lookup_group_plugin(group_type=None):
if (group_type is None):
return _default_group_plugin
return _group_plugins.get(group_type, (_default_organization_plugin if (group_type == 'organization') else _default_group_plugin))
|
[
"def",
"lookup_group_plugin",
"(",
"group_type",
"=",
"None",
")",
":",
"if",
"(",
"group_type",
"is",
"None",
")",
":",
"return",
"_default_group_plugin",
"return",
"_group_plugins",
".",
"get",
"(",
"group_type",
",",
"(",
"_default_organization_plugin",
"if",
"(",
"group_type",
"==",
"'organization'",
")",
"else",
"_default_group_plugin",
")",
")"
] |
returns the form plugin associated with the given group type .
|
train
| false
|
51,900
|
def setup_firewall():
from fabtools.require.shorewall import firewall, started
zones = [{'name': 'fw', 'type': 'firewall'}, {'name': 'net', 'type': 'ipv4'}, {'name': 'vz', 'type': 'ipv4'}]
interfaces = [{'zone': 'net', 'interface': 'eth0', 'options': 'proxyarp=1'}, {'zone': 'vz', 'interface': 'venet0', 'options': 'routeback,arp_filter=0'}]
masq = [{'interface': 'eth0', 'source': '192.168.1.0/24'}]
policy = [{'source': '$FW', 'dest': 'net', 'policy': 'ACCEPT'}, {'source': '$FW', 'dest': 'vz', 'policy': 'ACCEPT'}, {'source': 'vz', 'dest': 'net', 'policy': 'ACCEPT'}, {'source': 'net', 'dest': 'all', 'policy': 'DROP', 'log_level': 'info'}, {'source': 'all', 'dest': 'all', 'policy': 'REJECT', 'log_level': 'info'}]
firewall(zones=zones, interfaces=interfaces, policy=policy, masq=masq)
started()
|
[
"def",
"setup_firewall",
"(",
")",
":",
"from",
"fabtools",
".",
"require",
".",
"shorewall",
"import",
"firewall",
",",
"started",
"zones",
"=",
"[",
"{",
"'name'",
":",
"'fw'",
",",
"'type'",
":",
"'firewall'",
"}",
",",
"{",
"'name'",
":",
"'net'",
",",
"'type'",
":",
"'ipv4'",
"}",
",",
"{",
"'name'",
":",
"'vz'",
",",
"'type'",
":",
"'ipv4'",
"}",
"]",
"interfaces",
"=",
"[",
"{",
"'zone'",
":",
"'net'",
",",
"'interface'",
":",
"'eth0'",
",",
"'options'",
":",
"'proxyarp=1'",
"}",
",",
"{",
"'zone'",
":",
"'vz'",
",",
"'interface'",
":",
"'venet0'",
",",
"'options'",
":",
"'routeback,arp_filter=0'",
"}",
"]",
"masq",
"=",
"[",
"{",
"'interface'",
":",
"'eth0'",
",",
"'source'",
":",
"'192.168.1.0/24'",
"}",
"]",
"policy",
"=",
"[",
"{",
"'source'",
":",
"'$FW'",
",",
"'dest'",
":",
"'net'",
",",
"'policy'",
":",
"'ACCEPT'",
"}",
",",
"{",
"'source'",
":",
"'$FW'",
",",
"'dest'",
":",
"'vz'",
",",
"'policy'",
":",
"'ACCEPT'",
"}",
",",
"{",
"'source'",
":",
"'vz'",
",",
"'dest'",
":",
"'net'",
",",
"'policy'",
":",
"'ACCEPT'",
"}",
",",
"{",
"'source'",
":",
"'net'",
",",
"'dest'",
":",
"'all'",
",",
"'policy'",
":",
"'DROP'",
",",
"'log_level'",
":",
"'info'",
"}",
",",
"{",
"'source'",
":",
"'all'",
",",
"'dest'",
":",
"'all'",
",",
"'policy'",
":",
"'REJECT'",
",",
"'log_level'",
":",
"'info'",
"}",
"]",
"firewall",
"(",
"zones",
"=",
"zones",
",",
"interfaces",
"=",
"interfaces",
",",
"policy",
"=",
"policy",
",",
"masq",
"=",
"masq",
")",
"started",
"(",
")"
] |
shorewall config .
|
train
| false
|
51,901
|
def arg_byref(args, offset=(-1)):
return args[offset]._obj.value
|
[
"def",
"arg_byref",
"(",
"args",
",",
"offset",
"=",
"(",
"-",
"1",
")",
")",
":",
"return",
"args",
"[",
"offset",
"]",
".",
"_obj",
".",
"value"
] |
returns the pointer arguments by-reference value .
|
train
| false
|
51,902
|
def index_satisfying(iterable, condition):
for (i, x) in enumerate(iterable):
if condition(x):
return i
try:
return (i + 1)
except NameError:
raise ValueError('iterable must be non-empty')
|
[
"def",
"index_satisfying",
"(",
"iterable",
",",
"condition",
")",
":",
"for",
"(",
"i",
",",
"x",
")",
"in",
"enumerate",
"(",
"iterable",
")",
":",
"if",
"condition",
"(",
"x",
")",
":",
"return",
"i",
"try",
":",
"return",
"(",
"i",
"+",
"1",
")",
"except",
"NameError",
":",
"raise",
"ValueError",
"(",
"'iterable must be non-empty'",
")"
] |
returns the index of the first element in iterable that satisfies the given condition .
|
train
| false
|
51,903
|
def test_gnb_priors():
clf = GaussianNB(priors=np.array([0.3, 0.7])).fit(X, y)
assert_array_almost_equal(clf.predict_proba([[(-0.1), (-0.1)]]), np.array([[0.825303662161683, 0.174696337838317]]), 8)
assert_array_equal(clf.class_prior_, np.array([0.3, 0.7]))
|
[
"def",
"test_gnb_priors",
"(",
")",
":",
"clf",
"=",
"GaussianNB",
"(",
"priors",
"=",
"np",
".",
"array",
"(",
"[",
"0.3",
",",
"0.7",
"]",
")",
")",
".",
"fit",
"(",
"X",
",",
"y",
")",
"assert_array_almost_equal",
"(",
"clf",
".",
"predict_proba",
"(",
"[",
"[",
"(",
"-",
"0.1",
")",
",",
"(",
"-",
"0.1",
")",
"]",
"]",
")",
",",
"np",
".",
"array",
"(",
"[",
"[",
"0.825303662161683",
",",
"0.174696337838317",
"]",
"]",
")",
",",
"8",
")",
"assert_array_equal",
"(",
"clf",
".",
"class_prior_",
",",
"np",
".",
"array",
"(",
"[",
"0.3",
",",
"0.7",
"]",
")",
")"
] |
test whether the class prior override is properly used .
|
train
| false
|
51,904
|
def getJumpPoint(begin, end, loop, runningJumpSpace):
segment = (begin - end)
segmentLength = abs(segment)
if (segmentLength == 0.0):
return begin
segment /= segmentLength
distancePoint = DistancePoint(begin, loop, runningJumpSpace, segment)
if (distancePoint.distance == runningJumpSpace):
return distancePoint.point
effectiveDistance = distancePoint.distance
jumpPoint = distancePoint.point
segmentLeft = complex(0.7071067811865476, (-0.7071067811865476))
distancePoint = DistancePoint(begin, loop, runningJumpSpace, segmentLeft)
distancePoint.distance *= 0.5
if (distancePoint.distance > effectiveDistance):
effectiveDistance = distancePoint.distance
jumpPoint = distancePoint.point
segmentRight = complex(0.7071067811865476, 0.7071067811865476)
distancePoint = DistancePoint(begin, loop, runningJumpSpace, segmentRight)
distancePoint.distance *= 0.5
if (distancePoint.distance > effectiveDistance):
effectiveDistance = distancePoint.distance
jumpPoint = distancePoint.point
return jumpPoint
|
[
"def",
"getJumpPoint",
"(",
"begin",
",",
"end",
",",
"loop",
",",
"runningJumpSpace",
")",
":",
"segment",
"=",
"(",
"begin",
"-",
"end",
")",
"segmentLength",
"=",
"abs",
"(",
"segment",
")",
"if",
"(",
"segmentLength",
"==",
"0.0",
")",
":",
"return",
"begin",
"segment",
"/=",
"segmentLength",
"distancePoint",
"=",
"DistancePoint",
"(",
"begin",
",",
"loop",
",",
"runningJumpSpace",
",",
"segment",
")",
"if",
"(",
"distancePoint",
".",
"distance",
"==",
"runningJumpSpace",
")",
":",
"return",
"distancePoint",
".",
"point",
"effectiveDistance",
"=",
"distancePoint",
".",
"distance",
"jumpPoint",
"=",
"distancePoint",
".",
"point",
"segmentLeft",
"=",
"complex",
"(",
"0.7071067811865476",
",",
"(",
"-",
"0.7071067811865476",
")",
")",
"distancePoint",
"=",
"DistancePoint",
"(",
"begin",
",",
"loop",
",",
"runningJumpSpace",
",",
"segmentLeft",
")",
"distancePoint",
".",
"distance",
"*=",
"0.5",
"if",
"(",
"distancePoint",
".",
"distance",
">",
"effectiveDistance",
")",
":",
"effectiveDistance",
"=",
"distancePoint",
".",
"distance",
"jumpPoint",
"=",
"distancePoint",
".",
"point",
"segmentRight",
"=",
"complex",
"(",
"0.7071067811865476",
",",
"0.7071067811865476",
")",
"distancePoint",
"=",
"DistancePoint",
"(",
"begin",
",",
"loop",
",",
"runningJumpSpace",
",",
"segmentRight",
")",
"distancePoint",
".",
"distance",
"*=",
"0.5",
"if",
"(",
"distancePoint",
".",
"distance",
">",
"effectiveDistance",
")",
":",
"effectiveDistance",
"=",
"distancePoint",
".",
"distance",
"jumpPoint",
"=",
"distancePoint",
".",
"point",
"return",
"jumpPoint"
] |
get running jump point inside loop .
|
train
| false
|
51,906
|
def iter_blocks(course):
def visit(block):
' get child blocks '
(yield block)
for child in block.get_children():
for descendant in visit(child):
(yield descendant)
return visit(course)
|
[
"def",
"iter_blocks",
"(",
"course",
")",
":",
"def",
"visit",
"(",
"block",
")",
":",
"(",
"yield",
"block",
")",
"for",
"child",
"in",
"block",
".",
"get_children",
"(",
")",
":",
"for",
"descendant",
"in",
"visit",
"(",
"child",
")",
":",
"(",
"yield",
"descendant",
")",
"return",
"visit",
"(",
"course",
")"
] |
returns an iterator over all of the blocks in a course .
|
train
| false
|
51,908
|
def _calculate_num_threads(batch_size, shuffle):
if shuffle:
return min(10, int(round(math.sqrt(batch_size))))
else:
return 1
|
[
"def",
"_calculate_num_threads",
"(",
"batch_size",
",",
"shuffle",
")",
":",
"if",
"shuffle",
":",
"return",
"min",
"(",
"10",
",",
"int",
"(",
"round",
"(",
"math",
".",
"sqrt",
"(",
"batch_size",
")",
")",
")",
")",
"else",
":",
"return",
"1"
] |
calculates an appropriate number of threads for creating this database .
|
train
| false
|
51,909
|
def reverseNameFromIPv6Address(address):
fullHex = ''.join((('%02x' % (ord(c),)) for c in socket.inet_pton(socket.AF_INET6, address)))
tokens = (list(reversed(fullHex)) + ['ip6', 'arpa', ''])
return '.'.join(tokens)
|
[
"def",
"reverseNameFromIPv6Address",
"(",
"address",
")",
":",
"fullHex",
"=",
"''",
".",
"join",
"(",
"(",
"(",
"'%02x'",
"%",
"(",
"ord",
"(",
"c",
")",
",",
")",
")",
"for",
"c",
"in",
"socket",
".",
"inet_pton",
"(",
"socket",
".",
"AF_INET6",
",",
"address",
")",
")",
")",
"tokens",
"=",
"(",
"list",
"(",
"reversed",
"(",
"fullHex",
")",
")",
"+",
"[",
"'ip6'",
",",
"'arpa'",
",",
"''",
"]",
")",
"return",
"'.'",
".",
"join",
"(",
"tokens",
")"
] |
return a reverse domain name for the given ipv6 address .
|
train
| false
|
51,910
|
def energy_corrections(perturbation, n, a=10, mass=0.5):
(x, _a) = var('x _a')
Vnm = (lambda n, m, a: Integral(((X_n(n, a, x) * X_n(m, a, x)) * perturbation.subs({_a: a})), (x, 0, a)).n())
return (E_n(n, a, mass).evalf(), Vnm(n, n, a).evalf(), (((Vnm(n, (n - 1), a) ** 2) / (E_n(n, a, mass) - E_n((n - 1), a, mass))) + ((Vnm(n, (n + 1), a) ** 2) / (E_n(n, a, mass) - E_n((n + 1), a, mass)))).evalf())
|
[
"def",
"energy_corrections",
"(",
"perturbation",
",",
"n",
",",
"a",
"=",
"10",
",",
"mass",
"=",
"0.5",
")",
":",
"(",
"x",
",",
"_a",
")",
"=",
"var",
"(",
"'x _a'",
")",
"Vnm",
"=",
"(",
"lambda",
"n",
",",
"m",
",",
"a",
":",
"Integral",
"(",
"(",
"(",
"X_n",
"(",
"n",
",",
"a",
",",
"x",
")",
"*",
"X_n",
"(",
"m",
",",
"a",
",",
"x",
")",
")",
"*",
"perturbation",
".",
"subs",
"(",
"{",
"_a",
":",
"a",
"}",
")",
")",
",",
"(",
"x",
",",
"0",
",",
"a",
")",
")",
".",
"n",
"(",
")",
")",
"return",
"(",
"E_n",
"(",
"n",
",",
"a",
",",
"mass",
")",
".",
"evalf",
"(",
")",
",",
"Vnm",
"(",
"n",
",",
"n",
",",
"a",
")",
".",
"evalf",
"(",
")",
",",
"(",
"(",
"(",
"Vnm",
"(",
"n",
",",
"(",
"n",
"-",
"1",
")",
",",
"a",
")",
"**",
"2",
")",
"/",
"(",
"E_n",
"(",
"n",
",",
"a",
",",
"mass",
")",
"-",
"E_n",
"(",
"(",
"n",
"-",
"1",
")",
",",
"a",
",",
"mass",
")",
")",
")",
"+",
"(",
"(",
"Vnm",
"(",
"n",
",",
"(",
"n",
"+",
"1",
")",
",",
"a",
")",
"**",
"2",
")",
"/",
"(",
"E_n",
"(",
"n",
",",
"a",
",",
"mass",
")",
"-",
"E_n",
"(",
"(",
"n",
"+",
"1",
")",
",",
"a",
",",
"mass",
")",
")",
")",
")",
".",
"evalf",
"(",
")",
")"
] |
calculating first two order corrections due to perturbation theory and returns tuple where zero element is unperturbated energy .
|
train
| false
|
51,911
|
def get_pickleable_exception(exc):
nearest = find_nearest_pickleable_exception(exc)
if nearest:
return nearest
try:
pickle.dumps(deepcopy(exc))
except Exception:
return UnpickleableExceptionWrapper.from_exception(exc)
return exc
|
[
"def",
"get_pickleable_exception",
"(",
"exc",
")",
":",
"nearest",
"=",
"find_nearest_pickleable_exception",
"(",
"exc",
")",
"if",
"nearest",
":",
"return",
"nearest",
"try",
":",
"pickle",
".",
"dumps",
"(",
"deepcopy",
"(",
"exc",
")",
")",
"except",
"Exception",
":",
"return",
"UnpickleableExceptionWrapper",
".",
"from_exception",
"(",
"exc",
")",
"return",
"exc"
] |
make sure exception is pickleable .
|
train
| false
|
51,912
|
def add_arguments(parser):
adder = (getattr(parser, 'add_argument', None) or getattr(parser, 'add_option'))
adder('-l', '--log-level', default=logging.INFO, type=log_level, help='Set log level (DEBUG, INFO, WARNING, ERROR)')
|
[
"def",
"add_arguments",
"(",
"parser",
")",
":",
"adder",
"=",
"(",
"getattr",
"(",
"parser",
",",
"'add_argument'",
",",
"None",
")",
"or",
"getattr",
"(",
"parser",
",",
"'add_option'",
")",
")",
"adder",
"(",
"'-l'",
",",
"'--log-level'",
",",
"default",
"=",
"logging",
".",
"INFO",
",",
"type",
"=",
"log_level",
",",
"help",
"=",
"'Set log level (DEBUG, INFO, WARNING, ERROR)'",
")"
] |
add arguments to an argumentparser or optionparser for purposes of grabbing a logging level .
|
train
| true
|
51,914
|
def send_tarball(tarball):
wrapper = FileWrapper(tarball)
response = HttpResponse(wrapper, content_type='application/x-tgz')
response['Content-Disposition'] = ('attachment; filename=%s' % os.path.basename(tarball.name.encode('utf-8')))
response['Content-Length'] = os.path.getsize(tarball.name)
return response
|
[
"def",
"send_tarball",
"(",
"tarball",
")",
":",
"wrapper",
"=",
"FileWrapper",
"(",
"tarball",
")",
"response",
"=",
"HttpResponse",
"(",
"wrapper",
",",
"content_type",
"=",
"'application/x-tgz'",
")",
"response",
"[",
"'Content-Disposition'",
"]",
"=",
"(",
"'attachment; filename=%s'",
"%",
"os",
".",
"path",
".",
"basename",
"(",
"tarball",
".",
"name",
".",
"encode",
"(",
"'utf-8'",
")",
")",
")",
"response",
"[",
"'Content-Length'",
"]",
"=",
"os",
".",
"path",
".",
"getsize",
"(",
"tarball",
".",
"name",
")",
"return",
"response"
] |
renders a tarball to response .
|
train
| false
|
51,915
|
def create_exploration_summary(exploration_id, contributor_id_to_add):
exploration = get_exploration_by_id(exploration_id)
exp_summary = compute_summary_of_exploration(exploration, contributor_id_to_add)
save_exploration_summary(exp_summary)
|
[
"def",
"create_exploration_summary",
"(",
"exploration_id",
",",
"contributor_id_to_add",
")",
":",
"exploration",
"=",
"get_exploration_by_id",
"(",
"exploration_id",
")",
"exp_summary",
"=",
"compute_summary_of_exploration",
"(",
"exploration",
",",
"contributor_id_to_add",
")",
"save_exploration_summary",
"(",
"exp_summary",
")"
] |
create summary of an exploration and store in datastore .
|
train
| false
|
51,916
|
@step('I use the log record configuration')
def step_use_log_record_configuration(context):
assert context.table, 'REQUIRE: context.table'
context.table.require_columns(['property', 'value'])
for row in context.table.rows:
property_name = row['property']
value = row['value']
if (property_name == 'format'):
context.log_record_format = value
elif (property_name == 'datefmt'):
context.log_record_datefmt = value
else:
raise KeyError(('Unknown property=%s' % property_name))
|
[
"@",
"step",
"(",
"'I use the log record configuration'",
")",
"def",
"step_use_log_record_configuration",
"(",
"context",
")",
":",
"assert",
"context",
".",
"table",
",",
"'REQUIRE: context.table'",
"context",
".",
"table",
".",
"require_columns",
"(",
"[",
"'property'",
",",
"'value'",
"]",
")",
"for",
"row",
"in",
"context",
".",
"table",
".",
"rows",
":",
"property_name",
"=",
"row",
"[",
"'property'",
"]",
"value",
"=",
"row",
"[",
"'value'",
"]",
"if",
"(",
"property_name",
"==",
"'format'",
")",
":",
"context",
".",
"log_record_format",
"=",
"value",
"elif",
"(",
"property_name",
"==",
"'datefmt'",
")",
":",
"context",
".",
"log_record_datefmt",
"=",
"value",
"else",
":",
"raise",
"KeyError",
"(",
"(",
"'Unknown property=%s'",
"%",
"property_name",
")",
")"
] |
define log record configuration parameters .
|
train
| true
|
51,917
|
def is_allowed_country(country_object, context=None):
if (context and context.get('allowed_countries')):
allowed_countries = context.get('allowed_countries')
return ((country_object.name.lower() in allowed_countries) or (country_object.alpha2.lower() in allowed_countries))
return True
|
[
"def",
"is_allowed_country",
"(",
"country_object",
",",
"context",
"=",
"None",
")",
":",
"if",
"(",
"context",
"and",
"context",
".",
"get",
"(",
"'allowed_countries'",
")",
")",
":",
"allowed_countries",
"=",
"context",
".",
"get",
"(",
"'allowed_countries'",
")",
"return",
"(",
"(",
"country_object",
".",
"name",
".",
"lower",
"(",
")",
"in",
"allowed_countries",
")",
"or",
"(",
"country_object",
".",
"alpha2",
".",
"lower",
"(",
")",
"in",
"allowed_countries",
")",
")",
"return",
"True"
] |
check if country is allowed .
|
train
| false
|
51,918
|
@utils.arg('--all-tenants', dest='all_tenants', metavar='<0|1>', nargs='?', type=int, const=1, default=int(strutils.bool_from_string(os.environ.get('ALL_TENANTS', 'false'), True)), help=_('Display information from all tenants (Admin only).'))
@deprecated_network
def do_secgroup_list(cs, args):
search_opts = {'all_tenants': args.all_tenants}
columns = ['Id', 'Name', 'Description']
if args.all_tenants:
columns.append('Tenant_ID')
groups = cs.security_groups.list(search_opts=search_opts)
utils.print_list(groups, columns)
|
[
"@",
"utils",
".",
"arg",
"(",
"'--all-tenants'",
",",
"dest",
"=",
"'all_tenants'",
",",
"metavar",
"=",
"'<0|1>'",
",",
"nargs",
"=",
"'?'",
",",
"type",
"=",
"int",
",",
"const",
"=",
"1",
",",
"default",
"=",
"int",
"(",
"strutils",
".",
"bool_from_string",
"(",
"os",
".",
"environ",
".",
"get",
"(",
"'ALL_TENANTS'",
",",
"'false'",
")",
",",
"True",
")",
")",
",",
"help",
"=",
"_",
"(",
"'Display information from all tenants (Admin only).'",
")",
")",
"@",
"deprecated_network",
"def",
"do_secgroup_list",
"(",
"cs",
",",
"args",
")",
":",
"search_opts",
"=",
"{",
"'all_tenants'",
":",
"args",
".",
"all_tenants",
"}",
"columns",
"=",
"[",
"'Id'",
",",
"'Name'",
",",
"'Description'",
"]",
"if",
"args",
".",
"all_tenants",
":",
"columns",
".",
"append",
"(",
"'Tenant_ID'",
")",
"groups",
"=",
"cs",
".",
"security_groups",
".",
"list",
"(",
"search_opts",
"=",
"search_opts",
")",
"utils",
".",
"print_list",
"(",
"groups",
",",
"columns",
")"
] |
list security groups for the current tenant .
|
train
| false
|
51,919
|
def img_to_graph(img, mask=None, return_as=sparse.coo_matrix, dtype=None):
img = np.atleast_3d(img)
(n_x, n_y, n_z) = img.shape
return _to_graph(n_x, n_y, n_z, mask, img, return_as, dtype)
|
[
"def",
"img_to_graph",
"(",
"img",
",",
"mask",
"=",
"None",
",",
"return_as",
"=",
"sparse",
".",
"coo_matrix",
",",
"dtype",
"=",
"None",
")",
":",
"img",
"=",
"np",
".",
"atleast_3d",
"(",
"img",
")",
"(",
"n_x",
",",
"n_y",
",",
"n_z",
")",
"=",
"img",
".",
"shape",
"return",
"_to_graph",
"(",
"n_x",
",",
"n_y",
",",
"n_z",
",",
"mask",
",",
"img",
",",
"return_as",
",",
"dtype",
")"
] |
graph of the pixel-to-pixel gradient connections edges are weighted with the gradient values .
|
train
| false
|
51,920
|
def get_closest(word, possibilities, n=3, cutoff=0.6, fallback_to_first=True):
possibilities = list(possibilities)
try:
return get_close_matches(word, possibilities, n, cutoff)[0]
except IndexError:
if fallback_to_first:
return possibilities[0]
|
[
"def",
"get_closest",
"(",
"word",
",",
"possibilities",
",",
"n",
"=",
"3",
",",
"cutoff",
"=",
"0.6",
",",
"fallback_to_first",
"=",
"True",
")",
":",
"possibilities",
"=",
"list",
"(",
"possibilities",
")",
"try",
":",
"return",
"get_close_matches",
"(",
"word",
",",
"possibilities",
",",
"n",
",",
"cutoff",
")",
"[",
"0",
"]",
"except",
"IndexError",
":",
"if",
"fallback_to_first",
":",
"return",
"possibilities",
"[",
"0",
"]"
] |
returns closest match or just first from possibilities .
|
train
| false
|
51,921
|
def iter_markdown_lines(markdown_html):
warnings.warn(u'reviewboard.reviews.markdown_utils.iter_markdown_lines is deprecated. Please use djblets.markdown.iter_markdown_lines.', DeprecationWarning)
return djblets_markdown.iter_markdown_lines(markdown_html)
|
[
"def",
"iter_markdown_lines",
"(",
"markdown_html",
")",
":",
"warnings",
".",
"warn",
"(",
"u'reviewboard.reviews.markdown_utils.iter_markdown_lines is deprecated. Please use djblets.markdown.iter_markdown_lines.'",
",",
"DeprecationWarning",
")",
"return",
"djblets_markdown",
".",
"iter_markdown_lines",
"(",
"markdown_html",
")"
] |
iterates over lines of markdown .
|
train
| false
|
51,922
|
def user_domain_match(A, B):
A = A.lower()
B = B.lower()
if (not (liberal_is_HDN(A) and liberal_is_HDN(B))):
if (A == B):
return True
return False
initial_dot = B.startswith('.')
if (initial_dot and A.endswith(B)):
return True
if ((not initial_dot) and (A == B)):
return True
return False
|
[
"def",
"user_domain_match",
"(",
"A",
",",
"B",
")",
":",
"A",
"=",
"A",
".",
"lower",
"(",
")",
"B",
"=",
"B",
".",
"lower",
"(",
")",
"if",
"(",
"not",
"(",
"liberal_is_HDN",
"(",
"A",
")",
"and",
"liberal_is_HDN",
"(",
"B",
")",
")",
")",
":",
"if",
"(",
"A",
"==",
"B",
")",
":",
"return",
"True",
"return",
"False",
"initial_dot",
"=",
"B",
".",
"startswith",
"(",
"'.'",
")",
"if",
"(",
"initial_dot",
"and",
"A",
".",
"endswith",
"(",
"B",
")",
")",
":",
"return",
"True",
"if",
"(",
"(",
"not",
"initial_dot",
")",
"and",
"(",
"A",
"==",
"B",
")",
")",
":",
"return",
"True",
"return",
"False"
] |
for blocking/accepting domains .
|
train
| true
|
51,924
|
@mock_ec2
def test_eip_allocate_invalid_domain():
conn = boto.connect_ec2(u'the_key', u'the_secret')
with assert_raises(EC2ResponseError) as cm:
conn.allocate_address(domain=u'bogus')
cm.exception.code.should.equal(u'InvalidParameterValue')
cm.exception.status.should.equal(400)
cm.exception.request_id.should_not.be.none
|
[
"@",
"mock_ec2",
"def",
"test_eip_allocate_invalid_domain",
"(",
")",
":",
"conn",
"=",
"boto",
".",
"connect_ec2",
"(",
"u'the_key'",
",",
"u'the_secret'",
")",
"with",
"assert_raises",
"(",
"EC2ResponseError",
")",
"as",
"cm",
":",
"conn",
".",
"allocate_address",
"(",
"domain",
"=",
"u'bogus'",
")",
"cm",
".",
"exception",
".",
"code",
".",
"should",
".",
"equal",
"(",
"u'InvalidParameterValue'",
")",
"cm",
".",
"exception",
".",
"status",
".",
"should",
".",
"equal",
"(",
"400",
")",
"cm",
".",
"exception",
".",
"request_id",
".",
"should_not",
".",
"be",
".",
"none"
] |
allocate eip invalid domain .
|
train
| false
|
51,925
|
def get_conf():
if (_CORE_SITE_DICT is None):
_parse_core_site()
return _CORE_SITE_DICT
|
[
"def",
"get_conf",
"(",
")",
":",
"if",
"(",
"_CORE_SITE_DICT",
"is",
"None",
")",
":",
"_parse_core_site",
"(",
")",
"return",
"_CORE_SITE_DICT"
] |
get_conf() -> confparse object for core-site .
|
train
| false
|
51,927
|
def authentication_required(url, authenticator, abort_on):
realm = authenticator.realm()
if realm:
msg = '<b>{}</b> says:<br/>{}'.format(html.escape(url.toDisplayString()), html.escape(realm))
else:
msg = '<b>{}</b> needs authentication'.format(html.escape(url.toDisplayString()))
answer = message.ask(title='Authentication required', text=msg, mode=usertypes.PromptMode.user_pwd, abort_on=abort_on)
if (answer is not None):
authenticator.setUser(answer.user)
authenticator.setPassword(answer.password)
return answer
|
[
"def",
"authentication_required",
"(",
"url",
",",
"authenticator",
",",
"abort_on",
")",
":",
"realm",
"=",
"authenticator",
".",
"realm",
"(",
")",
"if",
"realm",
":",
"msg",
"=",
"'<b>{}</b> says:<br/>{}'",
".",
"format",
"(",
"html",
".",
"escape",
"(",
"url",
".",
"toDisplayString",
"(",
")",
")",
",",
"html",
".",
"escape",
"(",
"realm",
")",
")",
"else",
":",
"msg",
"=",
"'<b>{}</b> needs authentication'",
".",
"format",
"(",
"html",
".",
"escape",
"(",
"url",
".",
"toDisplayString",
"(",
")",
")",
")",
"answer",
"=",
"message",
".",
"ask",
"(",
"title",
"=",
"'Authentication required'",
",",
"text",
"=",
"msg",
",",
"mode",
"=",
"usertypes",
".",
"PromptMode",
".",
"user_pwd",
",",
"abort_on",
"=",
"abort_on",
")",
"if",
"(",
"answer",
"is",
"not",
"None",
")",
":",
"authenticator",
".",
"setUser",
"(",
"answer",
".",
"user",
")",
"authenticator",
".",
"setPassword",
"(",
"answer",
".",
"password",
")",
"return",
"answer"
] |
ask a prompt for an authentication question .
|
train
| false
|
51,928
|
def _cmp_recformats(f1, f2):
if ((f1[0] == 'a') and (f2[0] == 'a')):
return cmp(int(f1[1:]), int(f2[1:]))
else:
(f1, f2) = (NUMPY2FITS[f1], NUMPY2FITS[f2])
return cmp(FORMATORDER.index(f1), FORMATORDER.index(f2))
|
[
"def",
"_cmp_recformats",
"(",
"f1",
",",
"f2",
")",
":",
"if",
"(",
"(",
"f1",
"[",
"0",
"]",
"==",
"'a'",
")",
"and",
"(",
"f2",
"[",
"0",
"]",
"==",
"'a'",
")",
")",
":",
"return",
"cmp",
"(",
"int",
"(",
"f1",
"[",
"1",
":",
"]",
")",
",",
"int",
"(",
"f2",
"[",
"1",
":",
"]",
")",
")",
"else",
":",
"(",
"f1",
",",
"f2",
")",
"=",
"(",
"NUMPY2FITS",
"[",
"f1",
"]",
",",
"NUMPY2FITS",
"[",
"f2",
"]",
")",
"return",
"cmp",
"(",
"FORMATORDER",
".",
"index",
"(",
"f1",
")",
",",
"FORMATORDER",
".",
"index",
"(",
"f2",
")",
")"
] |
compares two numpy recformats using the ordering given by formatorder .
|
train
| false
|
51,929
|
def mark_sender():
try:
mid = request.args[0]
except:
raise SyntaxError
mtable = s3db.msg_message
stable = s3db.msg_sender
srecord = db((mtable.id == mid)).select(mtable.from_address, limitby=(0, 1)).first()
sender = srecord.from_address
record = db((stable.sender == sender)).select(stable.id, limitby=(0, 1)).first()
if record:
args = 'update'
else:
args = 'create'
url = URL(f='sender', args=args, vars=dict(sender=sender))
redirect(url)
|
[
"def",
"mark_sender",
"(",
")",
":",
"try",
":",
"mid",
"=",
"request",
".",
"args",
"[",
"0",
"]",
"except",
":",
"raise",
"SyntaxError",
"mtable",
"=",
"s3db",
".",
"msg_message",
"stable",
"=",
"s3db",
".",
"msg_sender",
"srecord",
"=",
"db",
"(",
"(",
"mtable",
".",
"id",
"==",
"mid",
")",
")",
".",
"select",
"(",
"mtable",
".",
"from_address",
",",
"limitby",
"=",
"(",
"0",
",",
"1",
")",
")",
".",
"first",
"(",
")",
"sender",
"=",
"srecord",
".",
"from_address",
"record",
"=",
"db",
"(",
"(",
"stable",
".",
"sender",
"==",
"sender",
")",
")",
".",
"select",
"(",
"stable",
".",
"id",
",",
"limitby",
"=",
"(",
"0",
",",
"1",
")",
")",
".",
"first",
"(",
")",
"if",
"record",
":",
"args",
"=",
"'update'",
"else",
":",
"args",
"=",
"'create'",
"url",
"=",
"URL",
"(",
"f",
"=",
"'sender'",
",",
"args",
"=",
"args",
",",
"vars",
"=",
"dict",
"(",
"sender",
"=",
"sender",
")",
")",
"redirect",
"(",
"url",
")"
] |
assign priority to the given sender .
|
train
| false
|
51,931
|
def _parallel_pairwise(X, Y, func, n_jobs, **kwds):
if (n_jobs < 0):
n_jobs = max(((cpu_count() + 1) + n_jobs), 1)
if (Y is None):
Y = X
if (n_jobs == 1):
return func(X, Y, **kwds)
fd = delayed(func)
ret = Parallel(n_jobs=n_jobs, verbose=0)((fd(X, Y[s], **kwds) for s in gen_even_slices(Y.shape[0], n_jobs)))
return np.hstack(ret)
|
[
"def",
"_parallel_pairwise",
"(",
"X",
",",
"Y",
",",
"func",
",",
"n_jobs",
",",
"**",
"kwds",
")",
":",
"if",
"(",
"n_jobs",
"<",
"0",
")",
":",
"n_jobs",
"=",
"max",
"(",
"(",
"(",
"cpu_count",
"(",
")",
"+",
"1",
")",
"+",
"n_jobs",
")",
",",
"1",
")",
"if",
"(",
"Y",
"is",
"None",
")",
":",
"Y",
"=",
"X",
"if",
"(",
"n_jobs",
"==",
"1",
")",
":",
"return",
"func",
"(",
"X",
",",
"Y",
",",
"**",
"kwds",
")",
"fd",
"=",
"delayed",
"(",
"func",
")",
"ret",
"=",
"Parallel",
"(",
"n_jobs",
"=",
"n_jobs",
",",
"verbose",
"=",
"0",
")",
"(",
"(",
"fd",
"(",
"X",
",",
"Y",
"[",
"s",
"]",
",",
"**",
"kwds",
")",
"for",
"s",
"in",
"gen_even_slices",
"(",
"Y",
".",
"shape",
"[",
"0",
"]",
",",
"n_jobs",
")",
")",
")",
"return",
"np",
".",
"hstack",
"(",
"ret",
")"
] |
break the pairwise matrix in n_jobs even slices and compute them in parallel .
|
train
| false
|
51,933
|
@handle_response_format
@treeio_login_required
@_process_mass_form
def index_sent(request, response_format='html'):
query = (Q(reply_to__isnull=True) & Q(author=request.user.profile.get_contact()))
if request.GET:
query = (query & _get_filter_query(request.GET))
objects = Object.filter_by_request(request, Message.objects.filter(query))
else:
objects = Object.filter_by_request(request, Message.objects.filter(query))
filters = FilterForm(request.user.profile, 'title', request.GET)
context = _get_default_context(request)
context.update({'filters': filters, 'messages': objects})
return render_to_response('messaging/index_sent', context, context_instance=RequestContext(request), response_format=response_format)
|
[
"@",
"handle_response_format",
"@",
"treeio_login_required",
"@",
"_process_mass_form",
"def",
"index_sent",
"(",
"request",
",",
"response_format",
"=",
"'html'",
")",
":",
"query",
"=",
"(",
"Q",
"(",
"reply_to__isnull",
"=",
"True",
")",
"&",
"Q",
"(",
"author",
"=",
"request",
".",
"user",
".",
"profile",
".",
"get_contact",
"(",
")",
")",
")",
"if",
"request",
".",
"GET",
":",
"query",
"=",
"(",
"query",
"&",
"_get_filter_query",
"(",
"request",
".",
"GET",
")",
")",
"objects",
"=",
"Object",
".",
"filter_by_request",
"(",
"request",
",",
"Message",
".",
"objects",
".",
"filter",
"(",
"query",
")",
")",
"else",
":",
"objects",
"=",
"Object",
".",
"filter_by_request",
"(",
"request",
",",
"Message",
".",
"objects",
".",
"filter",
"(",
"query",
")",
")",
"filters",
"=",
"FilterForm",
"(",
"request",
".",
"user",
".",
"profile",
",",
"'title'",
",",
"request",
".",
"GET",
")",
"context",
"=",
"_get_default_context",
"(",
"request",
")",
"context",
".",
"update",
"(",
"{",
"'filters'",
":",
"filters",
",",
"'messages'",
":",
"objects",
"}",
")",
"return",
"render_to_response",
"(",
"'messaging/index_sent'",
",",
"context",
",",
"context_instance",
"=",
"RequestContext",
"(",
"request",
")",
",",
"response_format",
"=",
"response_format",
")"
] |
sent messages index page .
|
train
| false
|
51,935
|
def p_command_stop(p):
p[0] = ('STOP',)
|
[
"def",
"p_command_stop",
"(",
"p",
")",
":",
"p",
"[",
"0",
"]",
"=",
"(",
"'STOP'",
",",
")"
] |
command : stop .
|
train
| false
|
51,936
|
def _validate_resource(resource):
try:
if (resource is not None):
resource = resourceprep(resource)
if (not resource):
raise InvalidJID(u'Resource must not be 0 bytes')
if (len(resource) > 1023):
raise InvalidJID(u'Resource must be less than 1024 bytes')
return resource
except stringprep_profiles.StringPrepError:
raise InvalidJID(u'Invalid resource')
|
[
"def",
"_validate_resource",
"(",
"resource",
")",
":",
"try",
":",
"if",
"(",
"resource",
"is",
"not",
"None",
")",
":",
"resource",
"=",
"resourceprep",
"(",
"resource",
")",
"if",
"(",
"not",
"resource",
")",
":",
"raise",
"InvalidJID",
"(",
"u'Resource must not be 0 bytes'",
")",
"if",
"(",
"len",
"(",
"resource",
")",
">",
"1023",
")",
":",
"raise",
"InvalidJID",
"(",
"u'Resource must be less than 1024 bytes'",
")",
"return",
"resource",
"except",
"stringprep_profiles",
".",
"StringPrepError",
":",
"raise",
"InvalidJID",
"(",
"u'Invalid resource'",
")"
] |
validate the resource portion of a jid .
|
train
| false
|
51,937
|
@flake8ext
def no_translate_debug_logs(logical_line, filename):
for hint in _all_hints:
if logical_line.startswith(('LOG.debug(%s(' % hint)):
(yield (0, "N319 Don't translate debug level logs"))
|
[
"@",
"flake8ext",
"def",
"no_translate_debug_logs",
"(",
"logical_line",
",",
"filename",
")",
":",
"for",
"hint",
"in",
"_all_hints",
":",
"if",
"logical_line",
".",
"startswith",
"(",
"(",
"'LOG.debug(%s('",
"%",
"hint",
")",
")",
":",
"(",
"yield",
"(",
"0",
",",
"\"N319 Don't translate debug level logs\"",
")",
")"
] |
n319 - check for log .
|
train
| false
|
51,939
|
def wait_for_debugger(pid):
with log.waitfor('Waiting for debugger') as l:
while (tracer(pid) is None):
time.sleep(0.01)
l.success()
|
[
"def",
"wait_for_debugger",
"(",
"pid",
")",
":",
"with",
"log",
".",
"waitfor",
"(",
"'Waiting for debugger'",
")",
"as",
"l",
":",
"while",
"(",
"tracer",
"(",
"pid",
")",
"is",
"None",
")",
":",
"time",
".",
"sleep",
"(",
"0.01",
")",
"l",
".",
"success",
"(",
")"
] |
wait_for_debugger -> none sleeps until the process with pid pid is being traced .
|
train
| false
|
51,940
|
def f1_score(y_real, y_pred):
return fbeta_score(y_real, y_pred, 1)
|
[
"def",
"f1_score",
"(",
"y_real",
",",
"y_pred",
")",
":",
"return",
"fbeta_score",
"(",
"y_real",
",",
"y_pred",
",",
"1",
")"
] |
compute f1 score the f1 score can be interpreted as a weighted average of the precision and recall .
|
train
| false
|
51,942
|
def install_middleware():
name = u'raven.contrib.django.middleware.SentryMiddleware'
all_names = (name, u'raven.contrib.django.middleware.SentryLogMiddleware')
with settings_lock:
middleware_attr = (u'MIDDLEWARE' if (getattr(settings, u'MIDDLEWARE', None) is not None) else u'MIDDLEWARE_CLASSES')
middleware = (getattr(settings, middleware_attr, ()) or ())
if set(all_names).isdisjoint(set(middleware)):
setattr(settings, middleware_attr, (type(middleware)((name,)) + middleware))
|
[
"def",
"install_middleware",
"(",
")",
":",
"name",
"=",
"u'raven.contrib.django.middleware.SentryMiddleware'",
"all_names",
"=",
"(",
"name",
",",
"u'raven.contrib.django.middleware.SentryLogMiddleware'",
")",
"with",
"settings_lock",
":",
"middleware_attr",
"=",
"(",
"u'MIDDLEWARE'",
"if",
"(",
"getattr",
"(",
"settings",
",",
"u'MIDDLEWARE'",
",",
"None",
")",
"is",
"not",
"None",
")",
"else",
"u'MIDDLEWARE_CLASSES'",
")",
"middleware",
"=",
"(",
"getattr",
"(",
"settings",
",",
"middleware_attr",
",",
"(",
")",
")",
"or",
"(",
")",
")",
"if",
"set",
"(",
"all_names",
")",
".",
"isdisjoint",
"(",
"set",
"(",
"middleware",
")",
")",
":",
"setattr",
"(",
"settings",
",",
"middleware_attr",
",",
"(",
"type",
"(",
"middleware",
")",
"(",
"(",
"name",
",",
")",
")",
"+",
"middleware",
")",
")"
] |
force installation of sentrymiddlware if its not explicitly present .
|
train
| false
|
51,943
|
def latest_records(resource, layout, list_id, limit, list_fields, orderby):
(datalist, numrows, ids) = resource.datalist(fields=list_fields, start=None, limit=limit, list_id=list_id, orderby=orderby, layout=layout)
if (numrows == 0):
table = resource.table
available_records = current.db((table.deleted != True))
if available_records.select(table._id, limitby=(0, 1)).first():
msg = DIV(S3CRUD.crud_string(resource.tablename, 'msg_no_match'), _class='empty')
else:
msg = DIV(S3CRUD.crud_string(resource.tablename, 'msg_list_empty'), _class='empty')
data = msg
else:
data = datalist.html()
return data
|
[
"def",
"latest_records",
"(",
"resource",
",",
"layout",
",",
"list_id",
",",
"limit",
",",
"list_fields",
",",
"orderby",
")",
":",
"(",
"datalist",
",",
"numrows",
",",
"ids",
")",
"=",
"resource",
".",
"datalist",
"(",
"fields",
"=",
"list_fields",
",",
"start",
"=",
"None",
",",
"limit",
"=",
"limit",
",",
"list_id",
"=",
"list_id",
",",
"orderby",
"=",
"orderby",
",",
"layout",
"=",
"layout",
")",
"if",
"(",
"numrows",
"==",
"0",
")",
":",
"table",
"=",
"resource",
".",
"table",
"available_records",
"=",
"current",
".",
"db",
"(",
"(",
"table",
".",
"deleted",
"!=",
"True",
")",
")",
"if",
"available_records",
".",
"select",
"(",
"table",
".",
"_id",
",",
"limitby",
"=",
"(",
"0",
",",
"1",
")",
")",
".",
"first",
"(",
")",
":",
"msg",
"=",
"DIV",
"(",
"S3CRUD",
".",
"crud_string",
"(",
"resource",
".",
"tablename",
",",
"'msg_no_match'",
")",
",",
"_class",
"=",
"'empty'",
")",
"else",
":",
"msg",
"=",
"DIV",
"(",
"S3CRUD",
".",
"crud_string",
"(",
"resource",
".",
"tablename",
",",
"'msg_list_empty'",
")",
",",
"_class",
"=",
"'empty'",
")",
"data",
"=",
"msg",
"else",
":",
"data",
"=",
"datalist",
".",
"html",
"(",
")",
"return",
"data"
] |
display a datalist of the latest records for a resource .
|
train
| false
|
51,945
|
def _blocklist_json(request):
(items, _) = get_items(groupby='id')
plugins = get_plugins()
issuerCertBlocks = BlocklistIssuerCert.objects.all()
gfxs = BlocklistGfx.objects.all()
ca = None
try:
ca = BlocklistCA.objects.all()[0]
ca = base64.b64encode(ca.data.encode('utf-8'))
except IndexError:
pass
last_update = int(round((time.time() * 1000)))
results = {'last_update': last_update, 'certificates': certificates_to_json(issuerCertBlocks), 'addons': addons_to_json(items), 'plugins': plugins_to_json(plugins), 'gfx': gfxs_to_json(gfxs), 'ca': ca}
return JsonResponse(results)
|
[
"def",
"_blocklist_json",
"(",
"request",
")",
":",
"(",
"items",
",",
"_",
")",
"=",
"get_items",
"(",
"groupby",
"=",
"'id'",
")",
"plugins",
"=",
"get_plugins",
"(",
")",
"issuerCertBlocks",
"=",
"BlocklistIssuerCert",
".",
"objects",
".",
"all",
"(",
")",
"gfxs",
"=",
"BlocklistGfx",
".",
"objects",
".",
"all",
"(",
")",
"ca",
"=",
"None",
"try",
":",
"ca",
"=",
"BlocklistCA",
".",
"objects",
".",
"all",
"(",
")",
"[",
"0",
"]",
"ca",
"=",
"base64",
".",
"b64encode",
"(",
"ca",
".",
"data",
".",
"encode",
"(",
"'utf-8'",
")",
")",
"except",
"IndexError",
":",
"pass",
"last_update",
"=",
"int",
"(",
"round",
"(",
"(",
"time",
".",
"time",
"(",
")",
"*",
"1000",
")",
")",
")",
"results",
"=",
"{",
"'last_update'",
":",
"last_update",
",",
"'certificates'",
":",
"certificates_to_json",
"(",
"issuerCertBlocks",
")",
",",
"'addons'",
":",
"addons_to_json",
"(",
"items",
")",
",",
"'plugins'",
":",
"plugins_to_json",
"(",
"plugins",
")",
",",
"'gfx'",
":",
"gfxs_to_json",
"(",
"gfxs",
")",
",",
"'ca'",
":",
"ca",
"}",
"return",
"JsonResponse",
"(",
"results",
")"
] |
export the whole blocklist in json .
|
train
| false
|
51,946
|
def _getFriends():
path = os.path.expanduser((('~/Library/Application Support/Skype/' + getUserName()) + '/main.db'))
with contextlib.closing(sqlite3.connect(path).cursor()) as db:
db.execute('SELECT skypename,fullname,displayname FROM Contacts WHERE type=1 AND is_permanent=1')
return db.fetchall()
|
[
"def",
"_getFriends",
"(",
")",
":",
"path",
"=",
"os",
".",
"path",
".",
"expanduser",
"(",
"(",
"(",
"'~/Library/Application Support/Skype/'",
"+",
"getUserName",
"(",
")",
")",
"+",
"'/main.db'",
")",
")",
"with",
"contextlib",
".",
"closing",
"(",
"sqlite3",
".",
"connect",
"(",
"path",
")",
".",
"cursor",
"(",
")",
")",
"as",
"db",
":",
"db",
".",
"execute",
"(",
"'SELECT skypename,fullname,displayname FROM Contacts WHERE type=1 AND is_permanent=1'",
")",
"return",
"db",
".",
"fetchall",
"(",
")"
] |
get friends from skype database :return: list of tuples of friends .
|
train
| false
|
51,947
|
def assert_has_text_matching(output, expression):
match = re.search(expression, output)
assert (match is not None), ("No text matching expression '%s' was found in output file." % expression)
|
[
"def",
"assert_has_text_matching",
"(",
"output",
",",
"expression",
")",
":",
"match",
"=",
"re",
".",
"search",
"(",
"expression",
",",
"output",
")",
"assert",
"(",
"match",
"is",
"not",
"None",
")",
",",
"(",
"\"No text matching expression '%s' was found in output file.\"",
"%",
"expression",
")"
] |
asserts the specified output contains text matching the regular expression specified by the argument expression .
|
train
| false
|
51,948
|
def view_entries(search_query=None):
query = Entry.select().order_by(Entry.timestamp.desc())
if search_query:
query = query.where(Entry.content.contains(search_query))
for entry in query:
timestamp = entry.timestamp.strftime('%A %B %d, %Y %I:%M%p')
print timestamp
print ('=' * len(timestamp))
print entry.content
print 'n) next entry'
print 'd) delete entry'
print 'q) return to main menu'
action = raw_input('Choice? (Ndq) ').lower().strip()
if (action == 'q'):
break
elif (action == 'd'):
entry.delete_instance()
break
|
[
"def",
"view_entries",
"(",
"search_query",
"=",
"None",
")",
":",
"query",
"=",
"Entry",
".",
"select",
"(",
")",
".",
"order_by",
"(",
"Entry",
".",
"timestamp",
".",
"desc",
"(",
")",
")",
"if",
"search_query",
":",
"query",
"=",
"query",
".",
"where",
"(",
"Entry",
".",
"content",
".",
"contains",
"(",
"search_query",
")",
")",
"for",
"entry",
"in",
"query",
":",
"timestamp",
"=",
"entry",
".",
"timestamp",
".",
"strftime",
"(",
"'%A %B %d, %Y %I:%M%p'",
")",
"print",
"timestamp",
"print",
"(",
"'='",
"*",
"len",
"(",
"timestamp",
")",
")",
"print",
"entry",
".",
"content",
"print",
"'n) next entry'",
"print",
"'d) delete entry'",
"print",
"'q) return to main menu'",
"action",
"=",
"raw_input",
"(",
"'Choice? (Ndq) '",
")",
".",
"lower",
"(",
")",
".",
"strip",
"(",
")",
"if",
"(",
"action",
"==",
"'q'",
")",
":",
"break",
"elif",
"(",
"action",
"==",
"'d'",
")",
":",
"entry",
".",
"delete_instance",
"(",
")",
"break"
] |
view previous entries .
|
train
| true
|
51,949
|
def extract_valid_args(args, func, startidx=0):
func_args = inspect.getargspec(func).args[startidx:]
return dict(((k, v) for (k, v) in list(vars(args).items()) if (k in func_args)))
|
[
"def",
"extract_valid_args",
"(",
"args",
",",
"func",
",",
"startidx",
"=",
"0",
")",
":",
"func_args",
"=",
"inspect",
".",
"getargspec",
"(",
"func",
")",
".",
"args",
"[",
"startidx",
":",
"]",
"return",
"dict",
"(",
"(",
"(",
"k",
",",
"v",
")",
"for",
"(",
"k",
",",
"v",
")",
"in",
"list",
"(",
"vars",
"(",
"args",
")",
".",
"items",
"(",
")",
")",
"if",
"(",
"k",
"in",
"func_args",
")",
")",
")"
] |
given a namespace of argparser args .
|
train
| false
|
51,950
|
def friendly_number(number, base=1000, decimals=0, suffix='', powers=['', 'k', 'M', 'G', 'T', 'P', 'E', 'Z', 'Y']):
count = 0
number = float(number)
while ((number > base) and (count < len(powers))):
number /= base
count += 1
if decimals:
fmt = ('%%.%df%%s%%s' % decimals)
else:
number = round(number)
fmt = '%d%s%s'
return (fmt % (number, powers[count], suffix))
|
[
"def",
"friendly_number",
"(",
"number",
",",
"base",
"=",
"1000",
",",
"decimals",
"=",
"0",
",",
"suffix",
"=",
"''",
",",
"powers",
"=",
"[",
"''",
",",
"'k'",
",",
"'M'",
",",
"'G'",
",",
"'T'",
",",
"'P'",
",",
"'E'",
",",
"'Z'",
",",
"'Y'",
"]",
")",
":",
"count",
"=",
"0",
"number",
"=",
"float",
"(",
"number",
")",
"while",
"(",
"(",
"number",
">",
"base",
")",
"and",
"(",
"count",
"<",
"len",
"(",
"powers",
")",
")",
")",
":",
"number",
"/=",
"base",
"count",
"+=",
"1",
"if",
"decimals",
":",
"fmt",
"=",
"(",
"'%%.%df%%s%%s'",
"%",
"decimals",
")",
"else",
":",
"number",
"=",
"round",
"(",
"number",
")",
"fmt",
"=",
"'%d%s%s'",
"return",
"(",
"fmt",
"%",
"(",
"number",
",",
"powers",
"[",
"count",
"]",
",",
"suffix",
")",
")"
] |
format a number as friendly text .
|
train
| false
|
51,953
|
def addElementToListTable(element, key, listTable):
if (key in listTable):
listTable[key].append(element)
else:
listTable[key] = [element]
|
[
"def",
"addElementToListTable",
"(",
"element",
",",
"key",
",",
"listTable",
")",
":",
"if",
"(",
"key",
"in",
"listTable",
")",
":",
"listTable",
"[",
"key",
"]",
".",
"append",
"(",
"element",
")",
"else",
":",
"listTable",
"[",
"key",
"]",
"=",
"[",
"element",
"]"
] |
add an element to the list table .
|
train
| false
|
51,954
|
def server_update(s_name, s_ip, **connection_args):
altered = False
cur_server = _server_get(s_name, **connection_args)
if (cur_server is None):
return False
alt_server = NSServer()
alt_server.set_name(s_name)
if (cur_server.get_ipaddress() != s_ip):
alt_server.set_ipaddress(s_ip)
altered = True
if (altered is False):
return False
nitro = _connect(**connection_args)
if (nitro is None):
return False
ret = True
try:
NSServer.update(nitro, alt_server)
except NSNitroError as error:
log.debug('netscaler module error - NSServer.update() failed: {0}'.format(error))
ret = False
_disconnect(nitro)
return ret
|
[
"def",
"server_update",
"(",
"s_name",
",",
"s_ip",
",",
"**",
"connection_args",
")",
":",
"altered",
"=",
"False",
"cur_server",
"=",
"_server_get",
"(",
"s_name",
",",
"**",
"connection_args",
")",
"if",
"(",
"cur_server",
"is",
"None",
")",
":",
"return",
"False",
"alt_server",
"=",
"NSServer",
"(",
")",
"alt_server",
".",
"set_name",
"(",
"s_name",
")",
"if",
"(",
"cur_server",
".",
"get_ipaddress",
"(",
")",
"!=",
"s_ip",
")",
":",
"alt_server",
".",
"set_ipaddress",
"(",
"s_ip",
")",
"altered",
"=",
"True",
"if",
"(",
"altered",
"is",
"False",
")",
":",
"return",
"False",
"nitro",
"=",
"_connect",
"(",
"**",
"connection_args",
")",
"if",
"(",
"nitro",
"is",
"None",
")",
":",
"return",
"False",
"ret",
"=",
"True",
"try",
":",
"NSServer",
".",
"update",
"(",
"nitro",
",",
"alt_server",
")",
"except",
"NSNitroError",
"as",
"error",
":",
"log",
".",
"debug",
"(",
"'netscaler module error - NSServer.update() failed: {0}'",
".",
"format",
"(",
"error",
")",
")",
"ret",
"=",
"False",
"_disconnect",
"(",
"nitro",
")",
"return",
"ret"
] |
update a servers attributes cli example: .
|
train
| true
|
51,956
|
def parse_version(*args):
v = None
if (len(args) == 1):
a = args[0]
if isinstance(a, tuple):
v = '.'.join((str(x) for x in a))
else:
v = str(a)
else:
v = '.'.join((str(a) for a in args))
if v.startswith('v'):
v = v[1:]
try:
return pkg_resources.SetuptoolsVersion(v)
except AttributeError:
return pkg_resources.parse_version(v)
|
[
"def",
"parse_version",
"(",
"*",
"args",
")",
":",
"v",
"=",
"None",
"if",
"(",
"len",
"(",
"args",
")",
"==",
"1",
")",
":",
"a",
"=",
"args",
"[",
"0",
"]",
"if",
"isinstance",
"(",
"a",
",",
"tuple",
")",
":",
"v",
"=",
"'.'",
".",
"join",
"(",
"(",
"str",
"(",
"x",
")",
"for",
"x",
"in",
"a",
")",
")",
"else",
":",
"v",
"=",
"str",
"(",
"a",
")",
"else",
":",
"v",
"=",
"'.'",
".",
"join",
"(",
"(",
"str",
"(",
"a",
")",
"for",
"a",
"in",
"args",
")",
")",
"if",
"v",
".",
"startswith",
"(",
"'v'",
")",
":",
"v",
"=",
"v",
"[",
"1",
":",
"]",
"try",
":",
"return",
"pkg_resources",
".",
"SetuptoolsVersion",
"(",
"v",
")",
"except",
"AttributeError",
":",
"return",
"pkg_resources",
".",
"parse_version",
"(",
"v",
")"
] |
returns a sortable version arguments: args -- a string .
|
train
| false
|
51,957
|
def boto_supports_kms_key_id():
return (hasattr(boto, 'Version') and (LooseVersion(boto.Version) >= LooseVersion('2.39.0')))
|
[
"def",
"boto_supports_kms_key_id",
"(",
")",
":",
"return",
"(",
"hasattr",
"(",
"boto",
",",
"'Version'",
")",
"and",
"(",
"LooseVersion",
"(",
"boto",
".",
"Version",
")",
">=",
"LooseVersion",
"(",
"'2.39.0'",
")",
")",
")"
] |
check if boto library supports kms_key_ids returns: true if version is equal to or higher then the version needed .
|
train
| false
|
51,958
|
def test_cnn_init():
cnn = CondensedNearestNeighbour(random_state=RND_SEED)
assert_equal(cnn.n_seeds_S, 1)
assert_equal(cnn.n_jobs, 1)
|
[
"def",
"test_cnn_init",
"(",
")",
":",
"cnn",
"=",
"CondensedNearestNeighbour",
"(",
"random_state",
"=",
"RND_SEED",
")",
"assert_equal",
"(",
"cnn",
".",
"n_seeds_S",
",",
"1",
")",
"assert_equal",
"(",
"cnn",
".",
"n_jobs",
",",
"1",
")"
] |
test the initialisation of the object .
|
train
| false
|
51,959
|
def for_signed_dtypes(name='dtype'):
return for_dtypes(_signed_dtypes, name=name)
|
[
"def",
"for_signed_dtypes",
"(",
"name",
"=",
"'dtype'",
")",
":",
"return",
"for_dtypes",
"(",
"_signed_dtypes",
",",
"name",
"=",
"name",
")"
] |
decorator that checks the fixture with signed dtypes .
|
train
| false
|
51,961
|
def _parse_nav_steps(arg_str):
COMMAND_ALIASES = [('LOGIN', _parse_login)]
if (not arg_str):
arg_str = ''
words = arg_str.split(' ')
for (name, callback) in COMMAND_ALIASES:
if (words[0] == name):
return callback(*words[1:])
commands = arg_str.split('|')
parsed_commands = reduce((lambda x, y: ((x + [y]) if y else x)), map(_parse_command, commands), [])
runhandler = '_command_handler'
return {'runhandler': runhandler, 'args': {'commands': parsed_commands}}
|
[
"def",
"_parse_nav_steps",
"(",
"arg_str",
")",
":",
"COMMAND_ALIASES",
"=",
"[",
"(",
"'LOGIN'",
",",
"_parse_login",
")",
"]",
"if",
"(",
"not",
"arg_str",
")",
":",
"arg_str",
"=",
"''",
"words",
"=",
"arg_str",
".",
"split",
"(",
"' '",
")",
"for",
"(",
"name",
",",
"callback",
")",
"in",
"COMMAND_ALIASES",
":",
"if",
"(",
"words",
"[",
"0",
"]",
"==",
"name",
")",
":",
"return",
"callback",
"(",
"*",
"words",
"[",
"1",
":",
"]",
")",
"commands",
"=",
"arg_str",
".",
"split",
"(",
"'|'",
")",
"parsed_commands",
"=",
"reduce",
"(",
"(",
"lambda",
"x",
",",
"y",
":",
"(",
"(",
"x",
"+",
"[",
"y",
"]",
")",
"if",
"y",
"else",
"x",
")",
")",
",",
"map",
"(",
"_parse_command",
",",
"commands",
")",
",",
"[",
"]",
")",
"runhandler",
"=",
"'_command_handler'",
"return",
"{",
"'runhandler'",
":",
"runhandler",
",",
"'args'",
":",
"{",
"'commands'",
":",
"parsed_commands",
"}",
"}"
] |
heres how to specify the navigation steps: 1 .
|
train
| false
|
51,962
|
def test_all_logarithmic(Chart):
chart = Chart(logarithmic=True)
chart.add('1', [1, 30, 8, 199, (-23)])
chart.add('2', [87, 42, 0.9, 189, 81])
assert chart.render()
|
[
"def",
"test_all_logarithmic",
"(",
"Chart",
")",
":",
"chart",
"=",
"Chart",
"(",
"logarithmic",
"=",
"True",
")",
"chart",
".",
"add",
"(",
"'1'",
",",
"[",
"1",
",",
"30",
",",
"8",
",",
"199",
",",
"(",
"-",
"23",
")",
"]",
")",
"chart",
".",
"add",
"(",
"'2'",
",",
"[",
"87",
",",
"42",
",",
"0.9",
",",
"189",
",",
"81",
"]",
")",
"assert",
"chart",
".",
"render",
"(",
")"
] |
test logarithmic view rendering .
|
train
| false
|
51,965
|
def pipeline_factory(loader, global_conf, **local_conf):
pipeline = local_conf[cfg.CONF.auth_strategy]
pipeline = pipeline.split()
filters = [loader.get_filter(n) for n in pipeline[:(-1)]]
app = loader.get_app(pipeline[(-1)])
filters.reverse()
for filter in filters:
app = filter(app)
return app
|
[
"def",
"pipeline_factory",
"(",
"loader",
",",
"global_conf",
",",
"**",
"local_conf",
")",
":",
"pipeline",
"=",
"local_conf",
"[",
"cfg",
".",
"CONF",
".",
"auth_strategy",
"]",
"pipeline",
"=",
"pipeline",
".",
"split",
"(",
")",
"filters",
"=",
"[",
"loader",
".",
"get_filter",
"(",
"n",
")",
"for",
"n",
"in",
"pipeline",
"[",
":",
"(",
"-",
"1",
")",
"]",
"]",
"app",
"=",
"loader",
".",
"get_app",
"(",
"pipeline",
"[",
"(",
"-",
"1",
")",
"]",
")",
"filters",
".",
"reverse",
"(",
")",
"for",
"filter",
"in",
"filters",
":",
"app",
"=",
"filter",
"(",
"app",
")",
"return",
"app"
] |
create a paste pipeline based on the auth_strategy config option .
|
train
| false
|
51,966
|
def crt(a_values, modulo_values):
m = 1
x = 0
for modulo in modulo_values:
m *= modulo
for (m_i, a_i) in zip(modulo_values, a_values):
M_i = (m // m_i)
inv = inverse(M_i, m_i)
x = ((x + ((a_i * M_i) * inv)) % m)
return x
|
[
"def",
"crt",
"(",
"a_values",
",",
"modulo_values",
")",
":",
"m",
"=",
"1",
"x",
"=",
"0",
"for",
"modulo",
"in",
"modulo_values",
":",
"m",
"*=",
"modulo",
"for",
"(",
"m_i",
",",
"a_i",
")",
"in",
"zip",
"(",
"modulo_values",
",",
"a_values",
")",
":",
"M_i",
"=",
"(",
"m",
"//",
"m_i",
")",
"inv",
"=",
"inverse",
"(",
"M_i",
",",
"m_i",
")",
"x",
"=",
"(",
"(",
"x",
"+",
"(",
"(",
"a_i",
"*",
"M_i",
")",
"*",
"inv",
")",
")",
"%",
"m",
")",
"return",
"x"
] |
chinese remainder theorem .
|
train
| false
|
51,967
|
def test_filefind():
f = tempfile.NamedTemporaryFile()
alt_dirs = paths.get_ipython_dir()
t = path.filefind(f.name, alt_dirs)
|
[
"def",
"test_filefind",
"(",
")",
":",
"f",
"=",
"tempfile",
".",
"NamedTemporaryFile",
"(",
")",
"alt_dirs",
"=",
"paths",
".",
"get_ipython_dir",
"(",
")",
"t",
"=",
"path",
".",
"filefind",
"(",
"f",
".",
"name",
",",
"alt_dirs",
")"
] |
various tests for filefind .
|
train
| false
|
51,968
|
def is_called_from_pytest():
return getattr(matplotlib, u'_called_from_pytest', False)
|
[
"def",
"is_called_from_pytest",
"(",
")",
":",
"return",
"getattr",
"(",
"matplotlib",
",",
"u'_called_from_pytest'",
",",
"False",
")"
] |
returns whether the call was done from pytest .
|
train
| false
|
51,971
|
def bootstrap_statistic(data, stats_fn, num_samples):
return [stats_fn(bootstrap_sample(data)) for _ in range(num_samples)]
|
[
"def",
"bootstrap_statistic",
"(",
"data",
",",
"stats_fn",
",",
"num_samples",
")",
":",
"return",
"[",
"stats_fn",
"(",
"bootstrap_sample",
"(",
"data",
")",
")",
"for",
"_",
"in",
"range",
"(",
"num_samples",
")",
"]"
] |
evaluates stats_fn on num_samples bootstrap samples from data .
|
train
| false
|
51,972
|
def resolveMUITimeZone(spec):
pattern = re.compile('@(?P<dllname>.*),-(?P<index>\\d+)(?:;(?P<comment>.*))?')
matcher = pattern.match(spec)
assert matcher, 'Could not parse MUI spec'
try:
handle = DLLCache[matcher.groupdict()['dllname']]
result = win32api.LoadString(handle, int(matcher.groupdict()['index']))
except win32api.error as e:
result = None
return result
|
[
"def",
"resolveMUITimeZone",
"(",
"spec",
")",
":",
"pattern",
"=",
"re",
".",
"compile",
"(",
"'@(?P<dllname>.*),-(?P<index>\\\\d+)(?:;(?P<comment>.*))?'",
")",
"matcher",
"=",
"pattern",
".",
"match",
"(",
"spec",
")",
"assert",
"matcher",
",",
"'Could not parse MUI spec'",
"try",
":",
"handle",
"=",
"DLLCache",
"[",
"matcher",
".",
"groupdict",
"(",
")",
"[",
"'dllname'",
"]",
"]",
"result",
"=",
"win32api",
".",
"LoadString",
"(",
"handle",
",",
"int",
"(",
"matcher",
".",
"groupdict",
"(",
")",
"[",
"'index'",
"]",
")",
")",
"except",
"win32api",
".",
"error",
"as",
"e",
":",
"result",
"=",
"None",
"return",
"result"
] |
resolve a multilingual user interface resource for the time zone name .
|
train
| false
|
51,974
|
def keyspaces():
sys = _sys_mgr()
return sys.list_keyspaces()
|
[
"def",
"keyspaces",
"(",
")",
":",
"sys",
"=",
"_sys_mgr",
"(",
")",
"return",
"sys",
".",
"list_keyspaces",
"(",
")"
] |
return existing keyspaces cli example: .
|
train
| false
|
51,975
|
def strip_boolean_result(method, exc_type=None, exc_str=None, fail_ret=None):
@wraps(method)
def wrapped(*args, **kwargs):
ret = method(*args, **kwargs)
if ret[0]:
if (len(ret) == 2):
return ret[1]
else:
return ret[1:]
else:
if exc_type:
raise exc_type((exc_str or 'call failed'))
return fail_ret
return wrapped
|
[
"def",
"strip_boolean_result",
"(",
"method",
",",
"exc_type",
"=",
"None",
",",
"exc_str",
"=",
"None",
",",
"fail_ret",
"=",
"None",
")",
":",
"@",
"wraps",
"(",
"method",
")",
"def",
"wrapped",
"(",
"*",
"args",
",",
"**",
"kwargs",
")",
":",
"ret",
"=",
"method",
"(",
"*",
"args",
",",
"**",
"kwargs",
")",
"if",
"ret",
"[",
"0",
"]",
":",
"if",
"(",
"len",
"(",
"ret",
")",
"==",
"2",
")",
":",
"return",
"ret",
"[",
"1",
"]",
"else",
":",
"return",
"ret",
"[",
"1",
":",
"]",
"else",
":",
"if",
"exc_type",
":",
"raise",
"exc_type",
"(",
"(",
"exc_str",
"or",
"'call failed'",
")",
")",
"return",
"fail_ret",
"return",
"wrapped"
] |
translate methods return value for stripping off success flag .
|
train
| true
|
51,976
|
def _list_from_layouttuple(tk, ltuple):
ltuple = tk.splitlist(ltuple)
res = []
indx = 0
while (indx < len(ltuple)):
name = ltuple[indx]
opts = {}
res.append((name, opts))
indx += 1
while (indx < len(ltuple)):
(opt, val) = ltuple[indx:(indx + 2)]
if (not opt.startswith('-')):
break
opt = opt[1:]
indx += 2
if (opt == 'children'):
val = _list_from_layouttuple(tk, val)
opts[opt] = val
return res
|
[
"def",
"_list_from_layouttuple",
"(",
"tk",
",",
"ltuple",
")",
":",
"ltuple",
"=",
"tk",
".",
"splitlist",
"(",
"ltuple",
")",
"res",
"=",
"[",
"]",
"indx",
"=",
"0",
"while",
"(",
"indx",
"<",
"len",
"(",
"ltuple",
")",
")",
":",
"name",
"=",
"ltuple",
"[",
"indx",
"]",
"opts",
"=",
"{",
"}",
"res",
".",
"append",
"(",
"(",
"name",
",",
"opts",
")",
")",
"indx",
"+=",
"1",
"while",
"(",
"indx",
"<",
"len",
"(",
"ltuple",
")",
")",
":",
"(",
"opt",
",",
"val",
")",
"=",
"ltuple",
"[",
"indx",
":",
"(",
"indx",
"+",
"2",
")",
"]",
"if",
"(",
"not",
"opt",
".",
"startswith",
"(",
"'-'",
")",
")",
":",
"break",
"opt",
"=",
"opt",
"[",
"1",
":",
"]",
"indx",
"+=",
"2",
"if",
"(",
"opt",
"==",
"'children'",
")",
":",
"val",
"=",
"_list_from_layouttuple",
"(",
"tk",
",",
"val",
")",
"opts",
"[",
"opt",
"]",
"=",
"val",
"return",
"res"
] |
construct a list from the tuple returned by ttk::layout .
|
train
| false
|
51,979
|
def freezedicts(obj):
if isinstance(obj, (list, tuple)):
return type(obj)([freezedicts(sub) for sub in obj])
if isinstance(obj, dict):
return frozenset(six.iteritems(obj))
return obj
|
[
"def",
"freezedicts",
"(",
"obj",
")",
":",
"if",
"isinstance",
"(",
"obj",
",",
"(",
"list",
",",
"tuple",
")",
")",
":",
"return",
"type",
"(",
"obj",
")",
"(",
"[",
"freezedicts",
"(",
"sub",
")",
"for",
"sub",
"in",
"obj",
"]",
")",
"if",
"isinstance",
"(",
"obj",
",",
"dict",
")",
":",
"return",
"frozenset",
"(",
"six",
".",
"iteritems",
"(",
"obj",
")",
")",
"return",
"obj"
] |
recursively iterate over obj .
|
train
| false
|
51,980
|
def val_load(db):
dbname = (db + u'.db')
if (not os.access(dbname, os.R_OK)):
sys.exit((u'Cannot read file: %s' % dbname))
else:
db_in = shelve.open(db)
from nltk.sem import Valuation
val = Valuation(db_in)
return val
|
[
"def",
"val_load",
"(",
"db",
")",
":",
"dbname",
"=",
"(",
"db",
"+",
"u'.db'",
")",
"if",
"(",
"not",
"os",
".",
"access",
"(",
"dbname",
",",
"os",
".",
"R_OK",
")",
")",
":",
"sys",
".",
"exit",
"(",
"(",
"u'Cannot read file: %s'",
"%",
"dbname",
")",
")",
"else",
":",
"db_in",
"=",
"shelve",
".",
"open",
"(",
"db",
")",
"from",
"nltk",
".",
"sem",
"import",
"Valuation",
"val",
"=",
"Valuation",
"(",
"db_in",
")",
"return",
"val"
] |
load a valuation from a persistent database .
|
train
| false
|
51,981
|
def task_absent(name):
ret = {'name': name, 'changes': {}, 'result': True, 'comment': ''}
task = __salt__['kapacitor.get_task'](name)
if task:
if __opts__['test']:
ret['result'] = None
ret['comment'] = 'Task would have been deleted'
else:
result = __salt__['kapacitor.delete_task'](name)
ret['result'] = result['success']
if (not ret['result']):
ret['comment'] = 'Could not disable task'
if result.get('stderr'):
ret['comment'] += ('\n' + result['stderr'])
return ret
ret['comment'] = 'Task was deleted'
ret['changes'][name] = 'deleted'
else:
ret['comment'] = 'Task does not exist'
return ret
|
[
"def",
"task_absent",
"(",
"name",
")",
":",
"ret",
"=",
"{",
"'name'",
":",
"name",
",",
"'changes'",
":",
"{",
"}",
",",
"'result'",
":",
"True",
",",
"'comment'",
":",
"''",
"}",
"task",
"=",
"__salt__",
"[",
"'kapacitor.get_task'",
"]",
"(",
"name",
")",
"if",
"task",
":",
"if",
"__opts__",
"[",
"'test'",
"]",
":",
"ret",
"[",
"'result'",
"]",
"=",
"None",
"ret",
"[",
"'comment'",
"]",
"=",
"'Task would have been deleted'",
"else",
":",
"result",
"=",
"__salt__",
"[",
"'kapacitor.delete_task'",
"]",
"(",
"name",
")",
"ret",
"[",
"'result'",
"]",
"=",
"result",
"[",
"'success'",
"]",
"if",
"(",
"not",
"ret",
"[",
"'result'",
"]",
")",
":",
"ret",
"[",
"'comment'",
"]",
"=",
"'Could not disable task'",
"if",
"result",
".",
"get",
"(",
"'stderr'",
")",
":",
"ret",
"[",
"'comment'",
"]",
"+=",
"(",
"'\\n'",
"+",
"result",
"[",
"'stderr'",
"]",
")",
"return",
"ret",
"ret",
"[",
"'comment'",
"]",
"=",
"'Task was deleted'",
"ret",
"[",
"'changes'",
"]",
"[",
"name",
"]",
"=",
"'deleted'",
"else",
":",
"ret",
"[",
"'comment'",
"]",
"=",
"'Task does not exist'",
"return",
"ret"
] |
ensure that a task is absent from kapacitor .
|
train
| true
|
51,982
|
def _B(slot):
return (slot * 2)
|
[
"def",
"_B",
"(",
"slot",
")",
":",
"return",
"(",
"slot",
"*",
"2",
")"
] |
convert slot to byte boundary .
|
train
| false
|
51,983
|
def cms_post_age(row):
if hasattr(row, 'cms_post'):
row = row.cms_post
try:
date = row.date
except:
return messages['NONE']
now = request.utcnow
age = (now - date)
if (age < timedelta(days=2)):
return 1
elif (age < timedelta(days=7)):
return 2
else:
return 3
|
[
"def",
"cms_post_age",
"(",
"row",
")",
":",
"if",
"hasattr",
"(",
"row",
",",
"'cms_post'",
")",
":",
"row",
"=",
"row",
".",
"cms_post",
"try",
":",
"date",
"=",
"row",
".",
"date",
"except",
":",
"return",
"messages",
"[",
"'NONE'",
"]",
"now",
"=",
"request",
".",
"utcnow",
"age",
"=",
"(",
"now",
"-",
"date",
")",
"if",
"(",
"age",
"<",
"timedelta",
"(",
"days",
"=",
"2",
")",
")",
":",
"return",
"1",
"elif",
"(",
"age",
"<",
"timedelta",
"(",
"days",
"=",
"7",
")",
")",
":",
"return",
"2",
"else",
":",
"return",
"3"
] |
the age of the post - used for colour-coding markers of alerts & incidents .
|
train
| false
|
51,984
|
def run_2to3(files, fixer_names=None, options=None, explicit=None):
if (not files):
return
from lib2to3.refactor import RefactoringTool, get_fixers_from_package
class DistutilsRefactoringTool(RefactoringTool, ):
def log_error(self, msg, *args, **kw):
log.error(msg, *args)
def log_message(self, msg, *args):
log.info(msg, *args)
def log_debug(self, msg, *args):
log.debug(msg, *args)
if (fixer_names is None):
fixer_names = get_fixers_from_package('lib2to3.fixes')
r = DistutilsRefactoringTool(fixer_names, options=options)
r.refactor(files, write=True)
|
[
"def",
"run_2to3",
"(",
"files",
",",
"fixer_names",
"=",
"None",
",",
"options",
"=",
"None",
",",
"explicit",
"=",
"None",
")",
":",
"if",
"(",
"not",
"files",
")",
":",
"return",
"from",
"lib2to3",
".",
"refactor",
"import",
"RefactoringTool",
",",
"get_fixers_from_package",
"class",
"DistutilsRefactoringTool",
"(",
"RefactoringTool",
",",
")",
":",
"def",
"log_error",
"(",
"self",
",",
"msg",
",",
"*",
"args",
",",
"**",
"kw",
")",
":",
"log",
".",
"error",
"(",
"msg",
",",
"*",
"args",
")",
"def",
"log_message",
"(",
"self",
",",
"msg",
",",
"*",
"args",
")",
":",
"log",
".",
"info",
"(",
"msg",
",",
"*",
"args",
")",
"def",
"log_debug",
"(",
"self",
",",
"msg",
",",
"*",
"args",
")",
":",
"log",
".",
"debug",
"(",
"msg",
",",
"*",
"args",
")",
"if",
"(",
"fixer_names",
"is",
"None",
")",
":",
"fixer_names",
"=",
"get_fixers_from_package",
"(",
"'lib2to3.fixes'",
")",
"r",
"=",
"DistutilsRefactoringTool",
"(",
"fixer_names",
",",
"options",
"=",
"options",
")",
"r",
".",
"refactor",
"(",
"files",
",",
"write",
"=",
"True",
")"
] |
invoke 2to3 on a list of python files .
|
train
| false
|
51,987
|
def get_visible_columns(data, table_meta, df):
columns = []
doc = (data[0] or frappe.new_doc(df.options))
def add_column(col_df):
return (is_visible(col_df, doc) and column_has_value(data, col_df.get(u'fieldname')))
if df.get(u'visible_columns'):
for col_df in df.get(u'visible_columns'):
docfield = table_meta.get_field(col_df.get(u'fieldname'))
if (not docfield):
continue
newdf = docfield.as_dict().copy()
newdf.update(col_df)
if add_column(newdf):
columns.append(newdf)
else:
for col_df in table_meta.fields:
if add_column(col_df):
columns.append(col_df)
return columns
|
[
"def",
"get_visible_columns",
"(",
"data",
",",
"table_meta",
",",
"df",
")",
":",
"columns",
"=",
"[",
"]",
"doc",
"=",
"(",
"data",
"[",
"0",
"]",
"or",
"frappe",
".",
"new_doc",
"(",
"df",
".",
"options",
")",
")",
"def",
"add_column",
"(",
"col_df",
")",
":",
"return",
"(",
"is_visible",
"(",
"col_df",
",",
"doc",
")",
"and",
"column_has_value",
"(",
"data",
",",
"col_df",
".",
"get",
"(",
"u'fieldname'",
")",
")",
")",
"if",
"df",
".",
"get",
"(",
"u'visible_columns'",
")",
":",
"for",
"col_df",
"in",
"df",
".",
"get",
"(",
"u'visible_columns'",
")",
":",
"docfield",
"=",
"table_meta",
".",
"get_field",
"(",
"col_df",
".",
"get",
"(",
"u'fieldname'",
")",
")",
"if",
"(",
"not",
"docfield",
")",
":",
"continue",
"newdf",
"=",
"docfield",
".",
"as_dict",
"(",
")",
".",
"copy",
"(",
")",
"newdf",
".",
"update",
"(",
"col_df",
")",
"if",
"add_column",
"(",
"newdf",
")",
":",
"columns",
".",
"append",
"(",
"newdf",
")",
"else",
":",
"for",
"col_df",
"in",
"table_meta",
".",
"fields",
":",
"if",
"add_column",
"(",
"col_df",
")",
":",
"columns",
".",
"append",
"(",
"col_df",
")",
"return",
"columns"
] |
returns list of visible columns based on print_hide and if all columns have value .
|
train
| false
|
51,989
|
def is_invertible(polynomial, threshold=1.0):
eigvals = np.linalg.eigvals(companion_matrix(polynomial))
return np.all((np.abs(eigvals) < threshold))
|
[
"def",
"is_invertible",
"(",
"polynomial",
",",
"threshold",
"=",
"1.0",
")",
":",
"eigvals",
"=",
"np",
".",
"linalg",
".",
"eigvals",
"(",
"companion_matrix",
"(",
"polynomial",
")",
")",
"return",
"np",
".",
"all",
"(",
"(",
"np",
".",
"abs",
"(",
"eigvals",
")",
"<",
"threshold",
")",
")"
] |
determine if a polynomial is invertible .
|
train
| false
|
51,990
|
def test_installer(args, plugin, config, temp_dir):
backup = _create_backup(config, temp_dir)
names_match = (plugin.get_all_names() == plugin.get_all_names_answer())
if names_match:
logger.info('get_all_names test succeeded')
else:
logger.error('get_all_names test failed for config %s', config)
domains = list(plugin.get_testable_domain_names())
success = test_deploy_cert(plugin, temp_dir, domains)
if (success and args.enhance):
success = test_enhancements(plugin, domains)
good_rollback = test_rollback(plugin, config, backup)
return (names_match and success and good_rollback)
|
[
"def",
"test_installer",
"(",
"args",
",",
"plugin",
",",
"config",
",",
"temp_dir",
")",
":",
"backup",
"=",
"_create_backup",
"(",
"config",
",",
"temp_dir",
")",
"names_match",
"=",
"(",
"plugin",
".",
"get_all_names",
"(",
")",
"==",
"plugin",
".",
"get_all_names_answer",
"(",
")",
")",
"if",
"names_match",
":",
"logger",
".",
"info",
"(",
"'get_all_names test succeeded'",
")",
"else",
":",
"logger",
".",
"error",
"(",
"'get_all_names test failed for config %s'",
",",
"config",
")",
"domains",
"=",
"list",
"(",
"plugin",
".",
"get_testable_domain_names",
"(",
")",
")",
"success",
"=",
"test_deploy_cert",
"(",
"plugin",
",",
"temp_dir",
",",
"domains",
")",
"if",
"(",
"success",
"and",
"args",
".",
"enhance",
")",
":",
"success",
"=",
"test_enhancements",
"(",
"plugin",
",",
"domains",
")",
"good_rollback",
"=",
"test_rollback",
"(",
"plugin",
",",
"config",
",",
"backup",
")",
"return",
"(",
"names_match",
"and",
"success",
"and",
"good_rollback",
")"
] |
tests plugin as an installer .
|
train
| false
|
51,991
|
def ver_str(version):
return '.'.join(map(str, version))
|
[
"def",
"ver_str",
"(",
"version",
")",
":",
"return",
"'.'",
".",
"join",
"(",
"map",
"(",
"str",
",",
"version",
")",
")"
] |
version tuple as string .
|
train
| false
|
51,992
|
def plot_mnist_digit(image):
fig = plt.figure()
ax = fig.add_subplot(1, 1, 1)
ax.matshow(image, cmap=matplotlib.cm.binary)
plt.xticks(np.array([]))
plt.yticks(np.array([]))
plt.show()
|
[
"def",
"plot_mnist_digit",
"(",
"image",
")",
":",
"fig",
"=",
"plt",
".",
"figure",
"(",
")",
"ax",
"=",
"fig",
".",
"add_subplot",
"(",
"1",
",",
"1",
",",
"1",
")",
"ax",
".",
"matshow",
"(",
"image",
",",
"cmap",
"=",
"matplotlib",
".",
"cm",
".",
"binary",
")",
"plt",
".",
"xticks",
"(",
"np",
".",
"array",
"(",
"[",
"]",
")",
")",
"plt",
".",
"yticks",
"(",
"np",
".",
"array",
"(",
"[",
"]",
")",
")",
"plt",
".",
"show",
"(",
")"
] |
plot a single mnist image .
|
train
| false
|
51,994
|
def _serialize_buffer(buffer, array_serialization=None):
if (array_serialization == 'binary'):
return buffer.ravel().tostring()
elif (array_serialization == 'base64'):
return {'storage_type': 'base64', 'buffer': base64.b64encode(buffer).decode('ascii')}
raise ValueError("The array serialization method should be 'binary' or 'base64'.")
|
[
"def",
"_serialize_buffer",
"(",
"buffer",
",",
"array_serialization",
"=",
"None",
")",
":",
"if",
"(",
"array_serialization",
"==",
"'binary'",
")",
":",
"return",
"buffer",
".",
"ravel",
"(",
")",
".",
"tostring",
"(",
")",
"elif",
"(",
"array_serialization",
"==",
"'base64'",
")",
":",
"return",
"{",
"'storage_type'",
":",
"'base64'",
",",
"'buffer'",
":",
"base64",
".",
"b64encode",
"(",
"buffer",
")",
".",
"decode",
"(",
"'ascii'",
")",
"}",
"raise",
"ValueError",
"(",
"\"The array serialization method should be 'binary' or 'base64'.\"",
")"
] |
serialize a numpy array .
|
train
| true
|
51,996
|
def skip_on_broken_permissions(test_method):
@wraps(test_method)
def wrapper(case, *args, **kwargs):
test_file = FilePath(case.mktemp())
test_file.touch()
test_file.chmod(0)
permissions = test_file.getPermissions()
test_file.chmod(511)
if (permissions != Permissions(0)):
raise SkipTest("Can't run test on filesystem with broken permissions.")
return test_method(case, *args, **kwargs)
return wrapper
|
[
"def",
"skip_on_broken_permissions",
"(",
"test_method",
")",
":",
"@",
"wraps",
"(",
"test_method",
")",
"def",
"wrapper",
"(",
"case",
",",
"*",
"args",
",",
"**",
"kwargs",
")",
":",
"test_file",
"=",
"FilePath",
"(",
"case",
".",
"mktemp",
"(",
")",
")",
"test_file",
".",
"touch",
"(",
")",
"test_file",
".",
"chmod",
"(",
"0",
")",
"permissions",
"=",
"test_file",
".",
"getPermissions",
"(",
")",
"test_file",
".",
"chmod",
"(",
"511",
")",
"if",
"(",
"permissions",
"!=",
"Permissions",
"(",
"0",
")",
")",
":",
"raise",
"SkipTest",
"(",
"\"Can't run test on filesystem with broken permissions.\"",
")",
"return",
"test_method",
"(",
"case",
",",
"*",
"args",
",",
"**",
"kwargs",
")",
"return",
"wrapper"
] |
skips the wrapped test when the temporary directory is on a filesystem with broken permissions .
|
train
| false
|
51,997
|
@login_required
def resetrepo(request):
if (request.method != 'POST'):
return HttpResponseNotAllowed(['POST'])
view_helpers.SvnRepository(request.user.username).reset()
view_helpers.unset_mission_completed(request.user.get_profile(), 'svn_checkout')
view_helpers.unset_mission_completed(request.user.get_profile(), 'svn_diff')
view_helpers.unset_mission_completed(request.user.get_profile(), 'svn_commit')
if ('stay_on_this_page' in request.GET):
return HttpResponseRedirect(reverse('svn_main_page'))
else:
return HttpResponseRedirect(reverse('svn_checkout'))
|
[
"@",
"login_required",
"def",
"resetrepo",
"(",
"request",
")",
":",
"if",
"(",
"request",
".",
"method",
"!=",
"'POST'",
")",
":",
"return",
"HttpResponseNotAllowed",
"(",
"[",
"'POST'",
"]",
")",
"view_helpers",
".",
"SvnRepository",
"(",
"request",
".",
"user",
".",
"username",
")",
".",
"reset",
"(",
")",
"view_helpers",
".",
"unset_mission_completed",
"(",
"request",
".",
"user",
".",
"get_profile",
"(",
")",
",",
"'svn_checkout'",
")",
"view_helpers",
".",
"unset_mission_completed",
"(",
"request",
".",
"user",
".",
"get_profile",
"(",
")",
",",
"'svn_diff'",
")",
"view_helpers",
".",
"unset_mission_completed",
"(",
"request",
".",
"user",
".",
"get_profile",
"(",
")",
",",
"'svn_commit'",
")",
"if",
"(",
"'stay_on_this_page'",
"in",
"request",
".",
"GET",
")",
":",
"return",
"HttpResponseRedirect",
"(",
"reverse",
"(",
"'svn_main_page'",
")",
")",
"else",
":",
"return",
"HttpResponseRedirect",
"(",
"reverse",
"(",
"'svn_checkout'",
")",
")"
] |
reset a users mission repository and mark steps as uncompleted .
|
train
| false
|
52,000
|
def loadApplication(filename, kind, passphrase=None):
if (kind == 'python'):
application = sob.loadValueFromFile(filename, 'application', passphrase)
else:
application = sob.load(filename, kind, passphrase)
return application
|
[
"def",
"loadApplication",
"(",
"filename",
",",
"kind",
",",
"passphrase",
"=",
"None",
")",
":",
"if",
"(",
"kind",
"==",
"'python'",
")",
":",
"application",
"=",
"sob",
".",
"loadValueFromFile",
"(",
"filename",
",",
"'application'",
",",
"passphrase",
")",
"else",
":",
"application",
"=",
"sob",
".",
"load",
"(",
"filename",
",",
"kind",
",",
"passphrase",
")",
"return",
"application"
] |
load application from a given file .
|
train
| false
|
52,003
|
def add_views_to_dataset_resources(context, dataset_dict, view_types=[], create_datastore_views=False):
created_views = []
for resource_dict in dataset_dict.get('resources', []):
new_views = add_views_to_resource(context, resource_dict, dataset_dict, view_types, create_datastore_views)
created_views.extend(new_views)
return created_views
|
[
"def",
"add_views_to_dataset_resources",
"(",
"context",
",",
"dataset_dict",
",",
"view_types",
"=",
"[",
"]",
",",
"create_datastore_views",
"=",
"False",
")",
":",
"created_views",
"=",
"[",
"]",
"for",
"resource_dict",
"in",
"dataset_dict",
".",
"get",
"(",
"'resources'",
",",
"[",
"]",
")",
":",
"new_views",
"=",
"add_views_to_resource",
"(",
"context",
",",
"resource_dict",
",",
"dataset_dict",
",",
"view_types",
",",
"create_datastore_views",
")",
"created_views",
".",
"extend",
"(",
"new_views",
")",
"return",
"created_views"
] |
creates the provided views on all resources of the provided dataset views to create are provided as a list of view_types .
|
train
| false
|
52,004
|
def remove_value(module):
consul_api = get_consul_api(module)
key = module.params.get('key')
value = module.params.get('value')
(index, existing) = consul_api.kv.get(key, recurse=module.params.get('recurse'))
changed = (existing != None)
if (changed and (not module.check_mode)):
consul_api.kv.delete(key, module.params.get('recurse'))
module.exit_json(changed=changed, index=index, key=key, data=existing)
|
[
"def",
"remove_value",
"(",
"module",
")",
":",
"consul_api",
"=",
"get_consul_api",
"(",
"module",
")",
"key",
"=",
"module",
".",
"params",
".",
"get",
"(",
"'key'",
")",
"value",
"=",
"module",
".",
"params",
".",
"get",
"(",
"'value'",
")",
"(",
"index",
",",
"existing",
")",
"=",
"consul_api",
".",
"kv",
".",
"get",
"(",
"key",
",",
"recurse",
"=",
"module",
".",
"params",
".",
"get",
"(",
"'recurse'",
")",
")",
"changed",
"=",
"(",
"existing",
"!=",
"None",
")",
"if",
"(",
"changed",
"and",
"(",
"not",
"module",
".",
"check_mode",
")",
")",
":",
"consul_api",
".",
"kv",
".",
"delete",
"(",
"key",
",",
"module",
".",
"params",
".",
"get",
"(",
"'recurse'",
")",
")",
"module",
".",
"exit_json",
"(",
"changed",
"=",
"changed",
",",
"index",
"=",
"index",
",",
"key",
"=",
"key",
",",
"data",
"=",
"existing",
")"
] |
remove the value associated with the given key .
|
train
| false
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.