id_within_dataset int64 1 55.5k | snippet stringlengths 19 14.2k | tokens listlengths 6 1.63k | nl stringlengths 6 352 | split_within_dataset stringclasses 1 value | is_duplicated bool 2 classes |
|---|---|---|---|---|---|
47,618 | def _api_get_config(name, output, kwargs):
(_, data) = config.get_dconfig(kwargs.get('section'), kwargs.get('keyword'))
return report(output, keyword='config', data=data)
| [
"def",
"_api_get_config",
"(",
"name",
",",
"output",
",",
"kwargs",
")",
":",
"(",
"_",
",",
"data",
")",
"=",
"config",
".",
"get_dconfig",
"(",
"kwargs",
".",
"get",
"(",
"'section'",
")",
",",
"kwargs",
".",
"get",
"(",
"'keyword'",
")",
")",
"return",
"report",
"(",
"output",
",",
"keyword",
"=",
"'config'",
",",
"data",
"=",
"data",
")"
] | api: accepts output . | train | false |
47,619 | def start_stub(name):
service = SERVICES.get(name, None)
if service:
fake_server = service['class'](port_num=service['port'])
setattr(world, name, fake_server)
| [
"def",
"start_stub",
"(",
"name",
")",
":",
"service",
"=",
"SERVICES",
".",
"get",
"(",
"name",
",",
"None",
")",
"if",
"service",
":",
"fake_server",
"=",
"service",
"[",
"'class'",
"]",
"(",
"port_num",
"=",
"service",
"[",
"'port'",
"]",
")",
"setattr",
"(",
"world",
",",
"name",
",",
"fake_server",
")"
] | start the required stub service running on a local port . | train | false |
47,620 | def get_model_or_none(model, *args, **kwargs):
try:
return model.objects.get(*args, **kwargs)
except model.DoesNotExist:
return None
| [
"def",
"get_model_or_none",
"(",
"model",
",",
"*",
"args",
",",
"**",
"kwargs",
")",
":",
"try",
":",
"return",
"model",
".",
"objects",
".",
"get",
"(",
"*",
"args",
",",
"**",
"kwargs",
")",
"except",
"model",
".",
"DoesNotExist",
":",
"return",
"None"
] | get model object or return none . | train | false |
47,623 | def check_free_space_in_dir(path, size):
from ..utils.console import human_file_size
space = get_free_space_in_dir(path)
if (space < size):
raise IOError(u"Not enough free space in '{0}' to download a {1} file".format(path, human_file_size(size)))
| [
"def",
"check_free_space_in_dir",
"(",
"path",
",",
"size",
")",
":",
"from",
".",
".",
"utils",
".",
"console",
"import",
"human_file_size",
"space",
"=",
"get_free_space_in_dir",
"(",
"path",
")",
"if",
"(",
"space",
"<",
"size",
")",
":",
"raise",
"IOError",
"(",
"u\"Not enough free space in '{0}' to download a {1} file\"",
".",
"format",
"(",
"path",
",",
"human_file_size",
"(",
"size",
")",
")",
")"
] | determines if a given directory has enough space to hold a file of a given size . | train | false |
47,624 | def find_listen_pids_namespace(namespace):
ip = ip_lib.IPWrapper(namespace=namespace)
pids = set()
cmd = ['netstat', '-nlp']
output = ip.netns.execute(cmd, run_as_root=True)
for line in output.splitlines():
m = NETSTAT_PIDS_REGEX.match(line)
if m:
pids.add(m.group('pid'))
return pids
| [
"def",
"find_listen_pids_namespace",
"(",
"namespace",
")",
":",
"ip",
"=",
"ip_lib",
".",
"IPWrapper",
"(",
"namespace",
"=",
"namespace",
")",
"pids",
"=",
"set",
"(",
")",
"cmd",
"=",
"[",
"'netstat'",
",",
"'-nlp'",
"]",
"output",
"=",
"ip",
".",
"netns",
".",
"execute",
"(",
"cmd",
",",
"run_as_root",
"=",
"True",
")",
"for",
"line",
"in",
"output",
".",
"splitlines",
"(",
")",
":",
"m",
"=",
"NETSTAT_PIDS_REGEX",
".",
"match",
"(",
"line",
")",
"if",
"m",
":",
"pids",
".",
"add",
"(",
"m",
".",
"group",
"(",
"'pid'",
")",
")",
"return",
"pids"
] | retrieve a list of pids of listening processes within the given netns . | train | false |
47,625 | def zset_score_pairs(response, **options):
if ((not response) or (not options['withscores'])):
return response
score_cast_func = options.get('score_cast_func', float)
it = iter(response)
return list(izip(it, imap(score_cast_func, it)))
| [
"def",
"zset_score_pairs",
"(",
"response",
",",
"**",
"options",
")",
":",
"if",
"(",
"(",
"not",
"response",
")",
"or",
"(",
"not",
"options",
"[",
"'withscores'",
"]",
")",
")",
":",
"return",
"response",
"score_cast_func",
"=",
"options",
".",
"get",
"(",
"'score_cast_func'",
",",
"float",
")",
"it",
"=",
"iter",
"(",
"response",
")",
"return",
"list",
"(",
"izip",
"(",
"it",
",",
"imap",
"(",
"score_cast_func",
",",
"it",
")",
")",
")"
] | if withscores is specified in the options . | train | true |
47,628 | def split_delimited_symbol(symbol):
if (symbol in _delimited_symbol_default_triggers):
return ('', '')
symbol = symbol.upper()
split_list = re.split(pattern=_delimited_symbol_delimiters_regex, string=symbol, maxsplit=1)
company_symbol = split_list[0]
if (len(split_list) > 1):
share_class_symbol = split_list[1]
else:
share_class_symbol = ''
return (company_symbol, share_class_symbol)
| [
"def",
"split_delimited_symbol",
"(",
"symbol",
")",
":",
"if",
"(",
"symbol",
"in",
"_delimited_symbol_default_triggers",
")",
":",
"return",
"(",
"''",
",",
"''",
")",
"symbol",
"=",
"symbol",
".",
"upper",
"(",
")",
"split_list",
"=",
"re",
".",
"split",
"(",
"pattern",
"=",
"_delimited_symbol_delimiters_regex",
",",
"string",
"=",
"symbol",
",",
"maxsplit",
"=",
"1",
")",
"company_symbol",
"=",
"split_list",
"[",
"0",
"]",
"if",
"(",
"len",
"(",
"split_list",
")",
">",
"1",
")",
":",
"share_class_symbol",
"=",
"split_list",
"[",
"1",
"]",
"else",
":",
"share_class_symbol",
"=",
"''",
"return",
"(",
"company_symbol",
",",
"share_class_symbol",
")"
] | takes in a symbol that may be delimited and splits it in to a company symbol and share class symbol . | train | true |
47,631 | def groupmean_d(x, d):
x = np.asarray(x)
nvars = (x.ndim + 1)
sli = (([slice(None)] + ([None] * (nvars - 2))) + [slice(None)])
return (((x[..., None] * d[sli]).sum(0) * 1.0) / d.sum(0))
| [
"def",
"groupmean_d",
"(",
"x",
",",
"d",
")",
":",
"x",
"=",
"np",
".",
"asarray",
"(",
"x",
")",
"nvars",
"=",
"(",
"x",
".",
"ndim",
"+",
"1",
")",
"sli",
"=",
"(",
"(",
"[",
"slice",
"(",
"None",
")",
"]",
"+",
"(",
"[",
"None",
"]",
"*",
"(",
"nvars",
"-",
"2",
")",
")",
")",
"+",
"[",
"slice",
"(",
"None",
")",
"]",
")",
"return",
"(",
"(",
"(",
"x",
"[",
"...",
",",
"None",
"]",
"*",
"d",
"[",
"sli",
"]",
")",
".",
"sum",
"(",
"0",
")",
"*",
"1.0",
")",
"/",
"d",
".",
"sum",
"(",
"0",
")",
")"
] | groupmeans using dummy variables parameters x : array_like . | train | false |
47,632 | def _get_vnic_manager(host_reference):
return host_reference.configManager.virtualNicManager
| [
"def",
"_get_vnic_manager",
"(",
"host_reference",
")",
":",
"return",
"host_reference",
".",
"configManager",
".",
"virtualNicManager"
] | helper function that returns a list of virtual nicmanagers and their information . | train | false |
47,633 | def _makeHeaderIPv6(sig=V2_SIGNATURE, verCom='!', famProto='!', addrLength='\x00$', addrs=((('\x00' * 15) + '\x01') * 2), ports='\x1f\x90"\xb8'):
return (((((sig + verCom) + famProto) + addrLength) + addrs) + ports)
| [
"def",
"_makeHeaderIPv6",
"(",
"sig",
"=",
"V2_SIGNATURE",
",",
"verCom",
"=",
"'!'",
",",
"famProto",
"=",
"'!'",
",",
"addrLength",
"=",
"'\\x00$'",
",",
"addrs",
"=",
"(",
"(",
"(",
"'\\x00'",
"*",
"15",
")",
"+",
"'\\x01'",
")",
"*",
"2",
")",
",",
"ports",
"=",
"'\\x1f\\x90\"\\xb8'",
")",
":",
"return",
"(",
"(",
"(",
"(",
"(",
"sig",
"+",
"verCom",
")",
"+",
"famProto",
")",
"+",
"addrLength",
")",
"+",
"addrs",
")",
"+",
"ports",
")"
] | construct a version 2 ipv6 header with custom bytes . | train | false |
47,635 | def convert_rgb_to_hsv(rgb):
(red, green, blue) = [(_ / BYTE_MAX) for _ in rgb]
(hue, saturation, brightness) = colorsys.rgb_to_hsv(red, green, blue)
return [int((hue * SHORT_MAX)), int((saturation * SHORT_MAX)), int((brightness * SHORT_MAX))]
| [
"def",
"convert_rgb_to_hsv",
"(",
"rgb",
")",
":",
"(",
"red",
",",
"green",
",",
"blue",
")",
"=",
"[",
"(",
"_",
"/",
"BYTE_MAX",
")",
"for",
"_",
"in",
"rgb",
"]",
"(",
"hue",
",",
"saturation",
",",
"brightness",
")",
"=",
"colorsys",
".",
"rgb_to_hsv",
"(",
"red",
",",
"green",
",",
"blue",
")",
"return",
"[",
"int",
"(",
"(",
"hue",
"*",
"SHORT_MAX",
")",
")",
",",
"int",
"(",
"(",
"saturation",
"*",
"SHORT_MAX",
")",
")",
",",
"int",
"(",
"(",
"brightness",
"*",
"SHORT_MAX",
")",
")",
"]"
] | convert home assistant rgb values to hsv values . | train | false |
47,636 | def table_description():
if (connection.vendor == u'sqlite'):
fields = connection.introspection.get_table_description(connection.cursor(), u'course_overviews_courseoverview')
return [f.name for f in fields]
else:
cursor = connection.cursor()
cursor.execute(u"\n SELECT column_name\n FROM information_schema.columns\n WHERE table_name = 'course_overviews_courseoverview' AND table_schema = DATABASE()")
rows = cursor.fetchall()
return [r[0] for r in rows]
| [
"def",
"table_description",
"(",
")",
":",
"if",
"(",
"connection",
".",
"vendor",
"==",
"u'sqlite'",
")",
":",
"fields",
"=",
"connection",
".",
"introspection",
".",
"get_table_description",
"(",
"connection",
".",
"cursor",
"(",
")",
",",
"u'course_overviews_courseoverview'",
")",
"return",
"[",
"f",
".",
"name",
"for",
"f",
"in",
"fields",
"]",
"else",
":",
"cursor",
"=",
"connection",
".",
"cursor",
"(",
")",
"cursor",
".",
"execute",
"(",
"u\"\\n SELECT column_name\\n FROM information_schema.columns\\n WHERE table_name = 'course_overviews_courseoverview' AND table_schema = DATABASE()\"",
")",
"rows",
"=",
"cursor",
".",
"fetchall",
"(",
")",
"return",
"[",
"r",
"[",
"0",
"]",
"for",
"r",
"in",
"rows",
"]"
] | handle mysql/pg vs sqlite . | train | false |
47,638 | def negotiate_locale(preferred, available, sep='_', aliases=LOCALE_ALIASES):
available = [a.lower() for a in available if a]
for locale in preferred:
ll = locale.lower()
if (ll in available):
return locale
if aliases:
alias = aliases.get(ll)
if alias:
alias = alias.replace('_', sep)
if (alias.lower() in available):
return alias
parts = locale.split(sep)
if ((len(parts) > 1) and (parts[0].lower() in available)):
return parts[0]
return None
| [
"def",
"negotiate_locale",
"(",
"preferred",
",",
"available",
",",
"sep",
"=",
"'_'",
",",
"aliases",
"=",
"LOCALE_ALIASES",
")",
":",
"available",
"=",
"[",
"a",
".",
"lower",
"(",
")",
"for",
"a",
"in",
"available",
"if",
"a",
"]",
"for",
"locale",
"in",
"preferred",
":",
"ll",
"=",
"locale",
".",
"lower",
"(",
")",
"if",
"(",
"ll",
"in",
"available",
")",
":",
"return",
"locale",
"if",
"aliases",
":",
"alias",
"=",
"aliases",
".",
"get",
"(",
"ll",
")",
"if",
"alias",
":",
"alias",
"=",
"alias",
".",
"replace",
"(",
"'_'",
",",
"sep",
")",
"if",
"(",
"alias",
".",
"lower",
"(",
")",
"in",
"available",
")",
":",
"return",
"alias",
"parts",
"=",
"locale",
".",
"split",
"(",
"sep",
")",
"if",
"(",
"(",
"len",
"(",
"parts",
")",
">",
"1",
")",
"and",
"(",
"parts",
"[",
"0",
"]",
".",
"lower",
"(",
")",
"in",
"available",
")",
")",
":",
"return",
"parts",
"[",
"0",
"]",
"return",
"None"
] | find the best match between available and requested locale strings . | train | false |
47,639 | def _isnan(num):
num = str(num).lower()
if (not num):
return 0
sign = 0
if (num[0] == '+'):
num = num[1:]
elif (num[0] == '-'):
num = num[1:]
sign = 1
if num.startswith('nan'):
if ((len(num) > 3) and (not num[3:].isdigit())):
return 0
return (1, sign, num[3:].lstrip('0'))
if num.startswith('snan'):
if ((len(num) > 4) and (not num[4:].isdigit())):
return 0
return (2, sign, num[4:].lstrip('0'))
return 0
| [
"def",
"_isnan",
"(",
"num",
")",
":",
"num",
"=",
"str",
"(",
"num",
")",
".",
"lower",
"(",
")",
"if",
"(",
"not",
"num",
")",
":",
"return",
"0",
"sign",
"=",
"0",
"if",
"(",
"num",
"[",
"0",
"]",
"==",
"'+'",
")",
":",
"num",
"=",
"num",
"[",
"1",
":",
"]",
"elif",
"(",
"num",
"[",
"0",
"]",
"==",
"'-'",
")",
":",
"num",
"=",
"num",
"[",
"1",
":",
"]",
"sign",
"=",
"1",
"if",
"num",
".",
"startswith",
"(",
"'nan'",
")",
":",
"if",
"(",
"(",
"len",
"(",
"num",
")",
">",
"3",
")",
"and",
"(",
"not",
"num",
"[",
"3",
":",
"]",
".",
"isdigit",
"(",
")",
")",
")",
":",
"return",
"0",
"return",
"(",
"1",
",",
"sign",
",",
"num",
"[",
"3",
":",
"]",
".",
"lstrip",
"(",
"'0'",
")",
")",
"if",
"num",
".",
"startswith",
"(",
"'snan'",
")",
":",
"if",
"(",
"(",
"len",
"(",
"num",
")",
">",
"4",
")",
"and",
"(",
"not",
"num",
"[",
"4",
":",
"]",
".",
"isdigit",
"(",
")",
")",
")",
":",
"return",
"0",
"return",
"(",
"2",
",",
"sign",
",",
"num",
"[",
"4",
":",
"]",
".",
"lstrip",
"(",
"'0'",
")",
")",
"return",
"0"
] | determines whether a string or float is nan => nan => snan 0 => not a nan . | train | false |
47,641 | def get_names_from_cert(csr, typ=OpenSSL.crypto.FILETYPE_PEM):
return _get_names_from_cert_or_req(csr, OpenSSL.crypto.load_certificate, typ)
| [
"def",
"get_names_from_cert",
"(",
"csr",
",",
"typ",
"=",
"OpenSSL",
".",
"crypto",
".",
"FILETYPE_PEM",
")",
":",
"return",
"_get_names_from_cert_or_req",
"(",
"csr",
",",
"OpenSSL",
".",
"crypto",
".",
"load_certificate",
",",
"typ",
")"
] | get a list of domains from a cert . | train | false |
47,642 | def parse_mime_headers(doc_file):
headers = []
while True:
line = doc_file.readline()
done = (line in ('\r\n', '\n', ''))
if six.PY3:
try:
line = line.decode('utf-8')
except UnicodeDecodeError:
line = line.decode('latin1')
headers.append(line)
if done:
break
if six.PY3:
header_string = ''.join(headers)
else:
header_string = ''.join(headers)
headers = email.parser.Parser().parsestr(header_string)
return HeaderKeyDict(headers)
| [
"def",
"parse_mime_headers",
"(",
"doc_file",
")",
":",
"headers",
"=",
"[",
"]",
"while",
"True",
":",
"line",
"=",
"doc_file",
".",
"readline",
"(",
")",
"done",
"=",
"(",
"line",
"in",
"(",
"'\\r\\n'",
",",
"'\\n'",
",",
"''",
")",
")",
"if",
"six",
".",
"PY3",
":",
"try",
":",
"line",
"=",
"line",
".",
"decode",
"(",
"'utf-8'",
")",
"except",
"UnicodeDecodeError",
":",
"line",
"=",
"line",
".",
"decode",
"(",
"'latin1'",
")",
"headers",
".",
"append",
"(",
"line",
")",
"if",
"done",
":",
"break",
"if",
"six",
".",
"PY3",
":",
"header_string",
"=",
"''",
".",
"join",
"(",
"headers",
")",
"else",
":",
"header_string",
"=",
"''",
".",
"join",
"(",
"headers",
")",
"headers",
"=",
"email",
".",
"parser",
".",
"Parser",
"(",
")",
".",
"parsestr",
"(",
"header_string",
")",
"return",
"HeaderKeyDict",
"(",
"headers",
")"
] | takes a file-like object containing a mime document and returns a headerkeydict containing the headers . | train | false |
47,643 | def get_forms_for_user(user):
editable_forms = UserPagePermissionsProxy(user).editable_pages()
editable_forms = editable_forms.filter(content_type__in=get_form_types())
for fn in hooks.get_hooks(u'filter_form_submissions_for_user'):
editable_forms = fn(user, editable_forms)
return editable_forms
| [
"def",
"get_forms_for_user",
"(",
"user",
")",
":",
"editable_forms",
"=",
"UserPagePermissionsProxy",
"(",
"user",
")",
".",
"editable_pages",
"(",
")",
"editable_forms",
"=",
"editable_forms",
".",
"filter",
"(",
"content_type__in",
"=",
"get_form_types",
"(",
")",
")",
"for",
"fn",
"in",
"hooks",
".",
"get_hooks",
"(",
"u'filter_form_submissions_for_user'",
")",
":",
"editable_forms",
"=",
"fn",
"(",
"user",
",",
"editable_forms",
")",
"return",
"editable_forms"
] | return a queryset of form pages that this user is allowed to access the submissions for . | train | false |
47,644 | def scale_to_unit_interval(ndar, eps=1e-08):
ndar = ndar.copy()
ndar -= ndar.min()
ndar *= (1.0 / (ndar.max() + eps))
return ndar
| [
"def",
"scale_to_unit_interval",
"(",
"ndar",
",",
"eps",
"=",
"1e-08",
")",
":",
"ndar",
"=",
"ndar",
".",
"copy",
"(",
")",
"ndar",
"-=",
"ndar",
".",
"min",
"(",
")",
"ndar",
"*=",
"(",
"1.0",
"/",
"(",
"ndar",
".",
"max",
"(",
")",
"+",
"eps",
")",
")",
"return",
"ndar"
] | scales all values in the ndarray ndar to be between 0 and 1 . | train | false |
47,645 | def renew(config, unused_plugins):
try:
renewal.handle_renewal_request(config)
finally:
hooks.run_saved_post_hooks()
| [
"def",
"renew",
"(",
"config",
",",
"unused_plugins",
")",
":",
"try",
":",
"renewal",
".",
"handle_renewal_request",
"(",
"config",
")",
"finally",
":",
"hooks",
".",
"run_saved_post_hooks",
"(",
")"
] | renew previously-obtained certificates . | train | false |
47,647 | def volume_present(name, volume_size, sparse=False, create_parent=False, properties=None, cloned_from=None):
ret = {'name': name, 'changes': {}, 'result': True, 'comment': ''}
if (not properties):
properties = {}
log.debug('zfs.volume_present::{0}::config::volume_size = {1}'.format(name, volume_size))
log.debug('zfs.volume_present::{0}::config::sparse = {1}'.format(name, sparse))
log.debug('zfs.volume_present::{0}::config::create_parent = {1}'.format(name, create_parent))
log.debug('zfs.volume_present::{0}::config::cloned_from = {1}'.format(name, cloned_from))
log.debug('zfs.volume_present::{0}::config::properties = {1}'.format(name, properties))
for prop in properties.keys():
if isinstance(properties[prop], bool):
properties[prop] = ('on' if properties[prop] else 'off')
if (('@' in name) or ('#' in name)):
ret['result'] = False
ret['comment'] = 'invalid filesystem or volume name: {0}'.format(name)
if cloned_from:
cloned_parent = cloned_from[:cloned_from.index('@')]
if ('@' not in cloned_from):
ret['result'] = False
ret['comment'] = '{0} is not a snapshot'.format(cloned_from)
elif (cloned_from not in __salt__['zfs.list'](cloned_from, **{'type': 'snapshot'})):
ret['result'] = False
ret['comment'] = 'snapshot {0} does not exist'.format(cloned_from)
elif (cloned_parent not in __salt__['zfs.list'](cloned_parent, **{'type': 'volume'})):
ret['result'] = False
ret['comment'] = 'snapshot {0} is not from a volume'.format(cloned_from)
if ret['result']:
if (name in __salt__['zfs.list'](name, **{'type': 'volume'})):
properties['volsize'] = volume_size
result = __salt__['zfs.get'](name, **{'properties': ','.join(properties.keys()), 'fields': 'value', 'depth': 1})
for prop in properties.keys():
if (properties[prop] != result[name][prop]['value']):
if (name not in ret['changes']):
ret['changes'][name] = {}
ret['changes'][name][prop] = properties[prop]
if (len(ret['changes']) > 0):
if (not __opts__['test']):
result = __salt__['zfs.set'](name, **ret['changes'][name])
if (name not in result):
ret['result'] = False
else:
for prop in result[name].keys():
if (result[name][prop] != 'set'):
ret['result'] = False
if ret['result']:
ret['comment'] = 'volume {0} was updated'.format(name)
else:
ret['changes'] = {}
ret['comment'] = 'volume {0} failed to be updated'.format(name)
else:
ret['comment'] = 'volume {0} is up to date'.format(name)
else:
result = {name: 'created'}
if (not __opts__['test']):
if (not cloned_from):
result = __salt__['zfs.create'](name, **{'volume_size': volume_size, 'sparse': sparse, 'create_parent': create_parent, 'properties': properties})
else:
result = __salt__['zfs.clone'](cloned_from, name, **{'create_parent': create_parent, 'properties': properties})
ret['result'] = (name in result)
if ret['result']:
ret['result'] = ((result[name] == 'created') or result[name].startswith('cloned'))
if ret['result']:
ret['changes'][name] = (properties if (len(properties) > 0) else result[name])
ret['comment'] = 'volume {0} was created'.format(name)
else:
ret['comment'] = 'failed to create volume {0}'.format(name)
if (name in result):
ret['comment'] = result[name]
return ret
| [
"def",
"volume_present",
"(",
"name",
",",
"volume_size",
",",
"sparse",
"=",
"False",
",",
"create_parent",
"=",
"False",
",",
"properties",
"=",
"None",
",",
"cloned_from",
"=",
"None",
")",
":",
"ret",
"=",
"{",
"'name'",
":",
"name",
",",
"'changes'",
":",
"{",
"}",
",",
"'result'",
":",
"True",
",",
"'comment'",
":",
"''",
"}",
"if",
"(",
"not",
"properties",
")",
":",
"properties",
"=",
"{",
"}",
"log",
".",
"debug",
"(",
"'zfs.volume_present::{0}::config::volume_size = {1}'",
".",
"format",
"(",
"name",
",",
"volume_size",
")",
")",
"log",
".",
"debug",
"(",
"'zfs.volume_present::{0}::config::sparse = {1}'",
".",
"format",
"(",
"name",
",",
"sparse",
")",
")",
"log",
".",
"debug",
"(",
"'zfs.volume_present::{0}::config::create_parent = {1}'",
".",
"format",
"(",
"name",
",",
"create_parent",
")",
")",
"log",
".",
"debug",
"(",
"'zfs.volume_present::{0}::config::cloned_from = {1}'",
".",
"format",
"(",
"name",
",",
"cloned_from",
")",
")",
"log",
".",
"debug",
"(",
"'zfs.volume_present::{0}::config::properties = {1}'",
".",
"format",
"(",
"name",
",",
"properties",
")",
")",
"for",
"prop",
"in",
"properties",
".",
"keys",
"(",
")",
":",
"if",
"isinstance",
"(",
"properties",
"[",
"prop",
"]",
",",
"bool",
")",
":",
"properties",
"[",
"prop",
"]",
"=",
"(",
"'on'",
"if",
"properties",
"[",
"prop",
"]",
"else",
"'off'",
")",
"if",
"(",
"(",
"'@'",
"in",
"name",
")",
"or",
"(",
"'#'",
"in",
"name",
")",
")",
":",
"ret",
"[",
"'result'",
"]",
"=",
"False",
"ret",
"[",
"'comment'",
"]",
"=",
"'invalid filesystem or volume name: {0}'",
".",
"format",
"(",
"name",
")",
"if",
"cloned_from",
":",
"cloned_parent",
"=",
"cloned_from",
"[",
":",
"cloned_from",
".",
"index",
"(",
"'@'",
")",
"]",
"if",
"(",
"'@'",
"not",
"in",
"cloned_from",
")",
":",
"ret",
"[",
"'result'",
"]",
"=",
"False",
"ret",
"[",
"'comment'",
"]",
"=",
"'{0} is not a snapshot'",
".",
"format",
"(",
"cloned_from",
")",
"elif",
"(",
"cloned_from",
"not",
"in",
"__salt__",
"[",
"'zfs.list'",
"]",
"(",
"cloned_from",
",",
"**",
"{",
"'type'",
":",
"'snapshot'",
"}",
")",
")",
":",
"ret",
"[",
"'result'",
"]",
"=",
"False",
"ret",
"[",
"'comment'",
"]",
"=",
"'snapshot {0} does not exist'",
".",
"format",
"(",
"cloned_from",
")",
"elif",
"(",
"cloned_parent",
"not",
"in",
"__salt__",
"[",
"'zfs.list'",
"]",
"(",
"cloned_parent",
",",
"**",
"{",
"'type'",
":",
"'volume'",
"}",
")",
")",
":",
"ret",
"[",
"'result'",
"]",
"=",
"False",
"ret",
"[",
"'comment'",
"]",
"=",
"'snapshot {0} is not from a volume'",
".",
"format",
"(",
"cloned_from",
")",
"if",
"ret",
"[",
"'result'",
"]",
":",
"if",
"(",
"name",
"in",
"__salt__",
"[",
"'zfs.list'",
"]",
"(",
"name",
",",
"**",
"{",
"'type'",
":",
"'volume'",
"}",
")",
")",
":",
"properties",
"[",
"'volsize'",
"]",
"=",
"volume_size",
"result",
"=",
"__salt__",
"[",
"'zfs.get'",
"]",
"(",
"name",
",",
"**",
"{",
"'properties'",
":",
"','",
".",
"join",
"(",
"properties",
".",
"keys",
"(",
")",
")",
",",
"'fields'",
":",
"'value'",
",",
"'depth'",
":",
"1",
"}",
")",
"for",
"prop",
"in",
"properties",
".",
"keys",
"(",
")",
":",
"if",
"(",
"properties",
"[",
"prop",
"]",
"!=",
"result",
"[",
"name",
"]",
"[",
"prop",
"]",
"[",
"'value'",
"]",
")",
":",
"if",
"(",
"name",
"not",
"in",
"ret",
"[",
"'changes'",
"]",
")",
":",
"ret",
"[",
"'changes'",
"]",
"[",
"name",
"]",
"=",
"{",
"}",
"ret",
"[",
"'changes'",
"]",
"[",
"name",
"]",
"[",
"prop",
"]",
"=",
"properties",
"[",
"prop",
"]",
"if",
"(",
"len",
"(",
"ret",
"[",
"'changes'",
"]",
")",
">",
"0",
")",
":",
"if",
"(",
"not",
"__opts__",
"[",
"'test'",
"]",
")",
":",
"result",
"=",
"__salt__",
"[",
"'zfs.set'",
"]",
"(",
"name",
",",
"**",
"ret",
"[",
"'changes'",
"]",
"[",
"name",
"]",
")",
"if",
"(",
"name",
"not",
"in",
"result",
")",
":",
"ret",
"[",
"'result'",
"]",
"=",
"False",
"else",
":",
"for",
"prop",
"in",
"result",
"[",
"name",
"]",
".",
"keys",
"(",
")",
":",
"if",
"(",
"result",
"[",
"name",
"]",
"[",
"prop",
"]",
"!=",
"'set'",
")",
":",
"ret",
"[",
"'result'",
"]",
"=",
"False",
"if",
"ret",
"[",
"'result'",
"]",
":",
"ret",
"[",
"'comment'",
"]",
"=",
"'volume {0} was updated'",
".",
"format",
"(",
"name",
")",
"else",
":",
"ret",
"[",
"'changes'",
"]",
"=",
"{",
"}",
"ret",
"[",
"'comment'",
"]",
"=",
"'volume {0} failed to be updated'",
".",
"format",
"(",
"name",
")",
"else",
":",
"ret",
"[",
"'comment'",
"]",
"=",
"'volume {0} is up to date'",
".",
"format",
"(",
"name",
")",
"else",
":",
"result",
"=",
"{",
"name",
":",
"'created'",
"}",
"if",
"(",
"not",
"__opts__",
"[",
"'test'",
"]",
")",
":",
"if",
"(",
"not",
"cloned_from",
")",
":",
"result",
"=",
"__salt__",
"[",
"'zfs.create'",
"]",
"(",
"name",
",",
"**",
"{",
"'volume_size'",
":",
"volume_size",
",",
"'sparse'",
":",
"sparse",
",",
"'create_parent'",
":",
"create_parent",
",",
"'properties'",
":",
"properties",
"}",
")",
"else",
":",
"result",
"=",
"__salt__",
"[",
"'zfs.clone'",
"]",
"(",
"cloned_from",
",",
"name",
",",
"**",
"{",
"'create_parent'",
":",
"create_parent",
",",
"'properties'",
":",
"properties",
"}",
")",
"ret",
"[",
"'result'",
"]",
"=",
"(",
"name",
"in",
"result",
")",
"if",
"ret",
"[",
"'result'",
"]",
":",
"ret",
"[",
"'result'",
"]",
"=",
"(",
"(",
"result",
"[",
"name",
"]",
"==",
"'created'",
")",
"or",
"result",
"[",
"name",
"]",
".",
"startswith",
"(",
"'cloned'",
")",
")",
"if",
"ret",
"[",
"'result'",
"]",
":",
"ret",
"[",
"'changes'",
"]",
"[",
"name",
"]",
"=",
"(",
"properties",
"if",
"(",
"len",
"(",
"properties",
")",
">",
"0",
")",
"else",
"result",
"[",
"name",
"]",
")",
"ret",
"[",
"'comment'",
"]",
"=",
"'volume {0} was created'",
".",
"format",
"(",
"name",
")",
"else",
":",
"ret",
"[",
"'comment'",
"]",
"=",
"'failed to create volume {0}'",
".",
"format",
"(",
"name",
")",
"if",
"(",
"name",
"in",
"result",
")",
":",
"ret",
"[",
"'comment'",
"]",
"=",
"result",
"[",
"name",
"]",
"return",
"ret"
] | check that a block volume exists . | train | false |
47,648 | def get_carrier_id():
carrier = get_carrier()
if (carrier is None):
return carrier
for carr in CARRIERS:
if (carr.slug == carrier):
return carr.id
return mkt.constants.carriers.UNKNOWN_CARRIER.id
| [
"def",
"get_carrier_id",
"(",
")",
":",
"carrier",
"=",
"get_carrier",
"(",
")",
"if",
"(",
"carrier",
"is",
"None",
")",
":",
"return",
"carrier",
"for",
"carr",
"in",
"CARRIERS",
":",
"if",
"(",
"carr",
".",
"slug",
"==",
"carrier",
")",
":",
"return",
"carr",
".",
"id",
"return",
"mkt",
".",
"constants",
".",
"carriers",
".",
"UNKNOWN_CARRIER",
".",
"id"
] | returns the carrier id for the request lifecycle . | train | false |
47,650 | def arcball_map_to_sphere(point, center, radius):
v0 = ((point[0] - center[0]) / radius)
v1 = ((center[1] - point[1]) / radius)
n = ((v0 * v0) + (v1 * v1))
if (n > 1.0):
n = math.sqrt(n)
return numpy.array([(v0 / n), (v1 / n), 0.0])
else:
return numpy.array([v0, v1, math.sqrt((1.0 - n))])
| [
"def",
"arcball_map_to_sphere",
"(",
"point",
",",
"center",
",",
"radius",
")",
":",
"v0",
"=",
"(",
"(",
"point",
"[",
"0",
"]",
"-",
"center",
"[",
"0",
"]",
")",
"/",
"radius",
")",
"v1",
"=",
"(",
"(",
"center",
"[",
"1",
"]",
"-",
"point",
"[",
"1",
"]",
")",
"/",
"radius",
")",
"n",
"=",
"(",
"(",
"v0",
"*",
"v0",
")",
"+",
"(",
"v1",
"*",
"v1",
")",
")",
"if",
"(",
"n",
">",
"1.0",
")",
":",
"n",
"=",
"math",
".",
"sqrt",
"(",
"n",
")",
"return",
"numpy",
".",
"array",
"(",
"[",
"(",
"v0",
"/",
"n",
")",
",",
"(",
"v1",
"/",
"n",
")",
",",
"0.0",
"]",
")",
"else",
":",
"return",
"numpy",
".",
"array",
"(",
"[",
"v0",
",",
"v1",
",",
"math",
".",
"sqrt",
"(",
"(",
"1.0",
"-",
"n",
")",
")",
"]",
")"
] | return unit sphere coordinates from window coordinates . | train | true |
47,651 | def arr_dtype_number(arr, num):
return np.dtype((arr.dtype.str[:2] + str(num)))
| [
"def",
"arr_dtype_number",
"(",
"arr",
",",
"num",
")",
":",
"return",
"np",
".",
"dtype",
"(",
"(",
"arr",
".",
"dtype",
".",
"str",
"[",
":",
"2",
"]",
"+",
"str",
"(",
"num",
")",
")",
")"
] | return dtype for given number of items per element . | train | false |
47,652 | def atomic_group_remove_labels(id, labels):
label_objs = models.Label.smart_get_bulk(labels)
models.AtomicGroup.smart_get(id).label_set.remove(*label_objs)
| [
"def",
"atomic_group_remove_labels",
"(",
"id",
",",
"labels",
")",
":",
"label_objs",
"=",
"models",
".",
"Label",
".",
"smart_get_bulk",
"(",
"labels",
")",
"models",
".",
"AtomicGroup",
".",
"smart_get",
"(",
"id",
")",
".",
"label_set",
".",
"remove",
"(",
"*",
"label_objs",
")"
] | remove labels from atomic group . | train | false |
47,653 | def HandleCommandLine(cls, serviceClassString=None, argv=None, customInstallOptions='', customOptionHandler=None):
err = 0
if (argv is None):
argv = sys.argv
if (len(argv) <= 1):
usage()
serviceName = cls._svc_name_
serviceDisplayName = cls._svc_display_name_
if (serviceClassString is None):
serviceClassString = GetServiceClassString(cls)
import getopt
try:
(opts, args) = getopt.getopt(argv[1:], customInstallOptions, ['password=', 'username=', 'startup=', 'perfmonini=', 'perfmondll=', 'interactive', 'wait='])
except getopt.error as details:
print details
usage()
userName = None
password = None
perfMonIni = perfMonDll = None
startup = None
interactive = None
waitSecs = 0
for (opt, val) in opts:
if (opt == '--username'):
userName = val
elif (opt == '--password'):
password = val
elif (opt == '--perfmonini'):
perfMonIni = val
elif (opt == '--perfmondll'):
perfMonDll = val
elif (opt == '--interactive'):
interactive = 1
elif (opt == '--startup'):
map = {'manual': win32service.SERVICE_DEMAND_START, 'auto': win32service.SERVICE_AUTO_START, 'disabled': win32service.SERVICE_DISABLED}
try:
startup = map[val.lower()]
except KeyError:
print ("'%s' is not a valid startup option" % val)
elif (opt == '--wait'):
try:
waitSecs = int(val)
except ValueError:
print '--wait must specify an integer number of seconds.'
usage()
arg = args[0]
knownArg = 0
if (arg == 'start'):
knownArg = 1
print ('Starting service %s' % serviceName)
try:
StartService(serviceName, args[1:])
if waitSecs:
WaitForServiceStatus(serviceName, win32service.SERVICE_RUNNING, waitSecs)
except win32service.error as exc:
print ('Error starting service: %s' % exc.strerror)
elif (arg == 'restart'):
knownArg = 1
print ('Restarting service %s' % serviceName)
RestartService(serviceName, args[1:])
if waitSecs:
WaitForServiceStatus(serviceName, win32service.SERVICE_RUNNING, waitSecs)
elif (arg == 'debug'):
knownArg = 1
if (not hasattr(sys, 'frozen')):
svcArgs = ' '.join(args[1:])
try:
exeName = LocateSpecificServiceExe(serviceName)
except win32api.error as exc:
if (exc[0] == winerror.ERROR_FILE_NOT_FOUND):
print 'The service does not appear to be installed.'
print 'Please install the service before debugging it.'
sys.exit(1)
raise
try:
os.system(('%s -debug %s %s' % (exeName, serviceName, svcArgs)))
except KeyboardInterrupt:
pass
else:
DebugService(cls, args)
if ((not knownArg) and (len(args) != 1)):
usage()
if (arg == 'install'):
knownArg = 1
try:
serviceDeps = cls._svc_deps_
except AttributeError:
serviceDeps = None
try:
exeName = cls._exe_name_
except AttributeError:
exeName = None
try:
exeArgs = cls._exe_args_
except AttributeError:
exeArgs = None
try:
description = cls._svc_description_
except AttributeError:
description = None
print ('Installing service %s' % (serviceName,))
try:
InstallService(serviceClassString, serviceName, serviceDisplayName, serviceDeps=serviceDeps, startType=startup, bRunInteractive=interactive, userName=userName, password=password, exeName=exeName, perfMonIni=perfMonIni, perfMonDll=perfMonDll, exeArgs=exeArgs, description=description)
if customOptionHandler:
customOptionHandler(*(opts,))
print 'Service installed'
except win32service.error as exc:
if (exc.winerror == winerror.ERROR_SERVICE_EXISTS):
arg = 'update'
else:
print ('Error installing service: %s (%d)' % (exc.strerror, exc.winerror))
err = exc.winerror
except ValueError as msg:
print ('Error installing service: %s' % str(msg))
err = (-1)
try:
RemoveService(serviceName)
except win32api.error:
print 'Warning - could not remove the partially installed service.'
if (arg == 'update'):
knownArg = 1
try:
serviceDeps = cls._svc_deps_
except AttributeError:
serviceDeps = None
try:
exeName = cls._exe_name_
except AttributeError:
exeName = None
try:
exeArgs = cls._exe_args_
except AttributeError:
exeArgs = None
try:
description = cls._svc_description_
except AttributeError:
description = None
print 'Changing service configuration'
try:
ChangeServiceConfig(serviceClassString, serviceName, serviceDeps=serviceDeps, startType=startup, bRunInteractive=interactive, userName=userName, password=password, exeName=exeName, displayName=serviceDisplayName, perfMonIni=perfMonIni, perfMonDll=perfMonDll, exeArgs=exeArgs, description=description)
if customOptionHandler:
customOptionHandler(*(opts,))
print 'Service updated'
except win32service.error as exc:
print ('Error changing service configuration: %s (%d)' % (exc.strerror, exc.winerror))
err = exc.winerror
elif (arg == 'remove'):
knownArg = 1
print ('Removing service %s' % serviceName)
try:
RemoveService(serviceName)
print 'Service removed'
except win32service.error as exc:
print ('Error removing service: %s (%d)' % (exc.strerror, exc.winerror))
err = exc.winerror
elif (arg == 'stop'):
knownArg = 1
print ('Stopping service %s' % serviceName)
try:
if waitSecs:
StopServiceWithDeps(serviceName, waitSecs=waitSecs)
else:
StopService(serviceName)
except win32service.error as exc:
print ('Error stopping service: %s (%d)' % (exc.strerror, exc.winerror))
err = exc.winerror
if (not knownArg):
err = (-1)
print ("Unknown command - '%s'" % arg)
usage()
return err
| [
"def",
"HandleCommandLine",
"(",
"cls",
",",
"serviceClassString",
"=",
"None",
",",
"argv",
"=",
"None",
",",
"customInstallOptions",
"=",
"''",
",",
"customOptionHandler",
"=",
"None",
")",
":",
"err",
"=",
"0",
"if",
"(",
"argv",
"is",
"None",
")",
":",
"argv",
"=",
"sys",
".",
"argv",
"if",
"(",
"len",
"(",
"argv",
")",
"<=",
"1",
")",
":",
"usage",
"(",
")",
"serviceName",
"=",
"cls",
".",
"_svc_name_",
"serviceDisplayName",
"=",
"cls",
".",
"_svc_display_name_",
"if",
"(",
"serviceClassString",
"is",
"None",
")",
":",
"serviceClassString",
"=",
"GetServiceClassString",
"(",
"cls",
")",
"import",
"getopt",
"try",
":",
"(",
"opts",
",",
"args",
")",
"=",
"getopt",
".",
"getopt",
"(",
"argv",
"[",
"1",
":",
"]",
",",
"customInstallOptions",
",",
"[",
"'password='",
",",
"'username='",
",",
"'startup='",
",",
"'perfmonini='",
",",
"'perfmondll='",
",",
"'interactive'",
",",
"'wait='",
"]",
")",
"except",
"getopt",
".",
"error",
"as",
"details",
":",
"print",
"details",
"usage",
"(",
")",
"userName",
"=",
"None",
"password",
"=",
"None",
"perfMonIni",
"=",
"perfMonDll",
"=",
"None",
"startup",
"=",
"None",
"interactive",
"=",
"None",
"waitSecs",
"=",
"0",
"for",
"(",
"opt",
",",
"val",
")",
"in",
"opts",
":",
"if",
"(",
"opt",
"==",
"'--username'",
")",
":",
"userName",
"=",
"val",
"elif",
"(",
"opt",
"==",
"'--password'",
")",
":",
"password",
"=",
"val",
"elif",
"(",
"opt",
"==",
"'--perfmonini'",
")",
":",
"perfMonIni",
"=",
"val",
"elif",
"(",
"opt",
"==",
"'--perfmondll'",
")",
":",
"perfMonDll",
"=",
"val",
"elif",
"(",
"opt",
"==",
"'--interactive'",
")",
":",
"interactive",
"=",
"1",
"elif",
"(",
"opt",
"==",
"'--startup'",
")",
":",
"map",
"=",
"{",
"'manual'",
":",
"win32service",
".",
"SERVICE_DEMAND_START",
",",
"'auto'",
":",
"win32service",
".",
"SERVICE_AUTO_START",
",",
"'disabled'",
":",
"win32service",
".",
"SERVICE_DISABLED",
"}",
"try",
":",
"startup",
"=",
"map",
"[",
"val",
".",
"lower",
"(",
")",
"]",
"except",
"KeyError",
":",
"print",
"(",
"\"'%s' is not a valid startup option\"",
"%",
"val",
")",
"elif",
"(",
"opt",
"==",
"'--wait'",
")",
":",
"try",
":",
"waitSecs",
"=",
"int",
"(",
"val",
")",
"except",
"ValueError",
":",
"print",
"'--wait must specify an integer number of seconds.'",
"usage",
"(",
")",
"arg",
"=",
"args",
"[",
"0",
"]",
"knownArg",
"=",
"0",
"if",
"(",
"arg",
"==",
"'start'",
")",
":",
"knownArg",
"=",
"1",
"print",
"(",
"'Starting service %s'",
"%",
"serviceName",
")",
"try",
":",
"StartService",
"(",
"serviceName",
",",
"args",
"[",
"1",
":",
"]",
")",
"if",
"waitSecs",
":",
"WaitForServiceStatus",
"(",
"serviceName",
",",
"win32service",
".",
"SERVICE_RUNNING",
",",
"waitSecs",
")",
"except",
"win32service",
".",
"error",
"as",
"exc",
":",
"print",
"(",
"'Error starting service: %s'",
"%",
"exc",
".",
"strerror",
")",
"elif",
"(",
"arg",
"==",
"'restart'",
")",
":",
"knownArg",
"=",
"1",
"print",
"(",
"'Restarting service %s'",
"%",
"serviceName",
")",
"RestartService",
"(",
"serviceName",
",",
"args",
"[",
"1",
":",
"]",
")",
"if",
"waitSecs",
":",
"WaitForServiceStatus",
"(",
"serviceName",
",",
"win32service",
".",
"SERVICE_RUNNING",
",",
"waitSecs",
")",
"elif",
"(",
"arg",
"==",
"'debug'",
")",
":",
"knownArg",
"=",
"1",
"if",
"(",
"not",
"hasattr",
"(",
"sys",
",",
"'frozen'",
")",
")",
":",
"svcArgs",
"=",
"' '",
".",
"join",
"(",
"args",
"[",
"1",
":",
"]",
")",
"try",
":",
"exeName",
"=",
"LocateSpecificServiceExe",
"(",
"serviceName",
")",
"except",
"win32api",
".",
"error",
"as",
"exc",
":",
"if",
"(",
"exc",
"[",
"0",
"]",
"==",
"winerror",
".",
"ERROR_FILE_NOT_FOUND",
")",
":",
"print",
"'The service does not appear to be installed.'",
"print",
"'Please install the service before debugging it.'",
"sys",
".",
"exit",
"(",
"1",
")",
"raise",
"try",
":",
"os",
".",
"system",
"(",
"(",
"'%s -debug %s %s'",
"%",
"(",
"exeName",
",",
"serviceName",
",",
"svcArgs",
")",
")",
")",
"except",
"KeyboardInterrupt",
":",
"pass",
"else",
":",
"DebugService",
"(",
"cls",
",",
"args",
")",
"if",
"(",
"(",
"not",
"knownArg",
")",
"and",
"(",
"len",
"(",
"args",
")",
"!=",
"1",
")",
")",
":",
"usage",
"(",
")",
"if",
"(",
"arg",
"==",
"'install'",
")",
":",
"knownArg",
"=",
"1",
"try",
":",
"serviceDeps",
"=",
"cls",
".",
"_svc_deps_",
"except",
"AttributeError",
":",
"serviceDeps",
"=",
"None",
"try",
":",
"exeName",
"=",
"cls",
".",
"_exe_name_",
"except",
"AttributeError",
":",
"exeName",
"=",
"None",
"try",
":",
"exeArgs",
"=",
"cls",
".",
"_exe_args_",
"except",
"AttributeError",
":",
"exeArgs",
"=",
"None",
"try",
":",
"description",
"=",
"cls",
".",
"_svc_description_",
"except",
"AttributeError",
":",
"description",
"=",
"None",
"print",
"(",
"'Installing service %s'",
"%",
"(",
"serviceName",
",",
")",
")",
"try",
":",
"InstallService",
"(",
"serviceClassString",
",",
"serviceName",
",",
"serviceDisplayName",
",",
"serviceDeps",
"=",
"serviceDeps",
",",
"startType",
"=",
"startup",
",",
"bRunInteractive",
"=",
"interactive",
",",
"userName",
"=",
"userName",
",",
"password",
"=",
"password",
",",
"exeName",
"=",
"exeName",
",",
"perfMonIni",
"=",
"perfMonIni",
",",
"perfMonDll",
"=",
"perfMonDll",
",",
"exeArgs",
"=",
"exeArgs",
",",
"description",
"=",
"description",
")",
"if",
"customOptionHandler",
":",
"customOptionHandler",
"(",
"*",
"(",
"opts",
",",
")",
")",
"print",
"'Service installed'",
"except",
"win32service",
".",
"error",
"as",
"exc",
":",
"if",
"(",
"exc",
".",
"winerror",
"==",
"winerror",
".",
"ERROR_SERVICE_EXISTS",
")",
":",
"arg",
"=",
"'update'",
"else",
":",
"print",
"(",
"'Error installing service: %s (%d)'",
"%",
"(",
"exc",
".",
"strerror",
",",
"exc",
".",
"winerror",
")",
")",
"err",
"=",
"exc",
".",
"winerror",
"except",
"ValueError",
"as",
"msg",
":",
"print",
"(",
"'Error installing service: %s'",
"%",
"str",
"(",
"msg",
")",
")",
"err",
"=",
"(",
"-",
"1",
")",
"try",
":",
"RemoveService",
"(",
"serviceName",
")",
"except",
"win32api",
".",
"error",
":",
"print",
"'Warning - could not remove the partially installed service.'",
"if",
"(",
"arg",
"==",
"'update'",
")",
":",
"knownArg",
"=",
"1",
"try",
":",
"serviceDeps",
"=",
"cls",
".",
"_svc_deps_",
"except",
"AttributeError",
":",
"serviceDeps",
"=",
"None",
"try",
":",
"exeName",
"=",
"cls",
".",
"_exe_name_",
"except",
"AttributeError",
":",
"exeName",
"=",
"None",
"try",
":",
"exeArgs",
"=",
"cls",
".",
"_exe_args_",
"except",
"AttributeError",
":",
"exeArgs",
"=",
"None",
"try",
":",
"description",
"=",
"cls",
".",
"_svc_description_",
"except",
"AttributeError",
":",
"description",
"=",
"None",
"print",
"'Changing service configuration'",
"try",
":",
"ChangeServiceConfig",
"(",
"serviceClassString",
",",
"serviceName",
",",
"serviceDeps",
"=",
"serviceDeps",
",",
"startType",
"=",
"startup",
",",
"bRunInteractive",
"=",
"interactive",
",",
"userName",
"=",
"userName",
",",
"password",
"=",
"password",
",",
"exeName",
"=",
"exeName",
",",
"displayName",
"=",
"serviceDisplayName",
",",
"perfMonIni",
"=",
"perfMonIni",
",",
"perfMonDll",
"=",
"perfMonDll",
",",
"exeArgs",
"=",
"exeArgs",
",",
"description",
"=",
"description",
")",
"if",
"customOptionHandler",
":",
"customOptionHandler",
"(",
"*",
"(",
"opts",
",",
")",
")",
"print",
"'Service updated'",
"except",
"win32service",
".",
"error",
"as",
"exc",
":",
"print",
"(",
"'Error changing service configuration: %s (%d)'",
"%",
"(",
"exc",
".",
"strerror",
",",
"exc",
".",
"winerror",
")",
")",
"err",
"=",
"exc",
".",
"winerror",
"elif",
"(",
"arg",
"==",
"'remove'",
")",
":",
"knownArg",
"=",
"1",
"print",
"(",
"'Removing service %s'",
"%",
"serviceName",
")",
"try",
":",
"RemoveService",
"(",
"serviceName",
")",
"print",
"'Service removed'",
"except",
"win32service",
".",
"error",
"as",
"exc",
":",
"print",
"(",
"'Error removing service: %s (%d)'",
"%",
"(",
"exc",
".",
"strerror",
",",
"exc",
".",
"winerror",
")",
")",
"err",
"=",
"exc",
".",
"winerror",
"elif",
"(",
"arg",
"==",
"'stop'",
")",
":",
"knownArg",
"=",
"1",
"print",
"(",
"'Stopping service %s'",
"%",
"serviceName",
")",
"try",
":",
"if",
"waitSecs",
":",
"StopServiceWithDeps",
"(",
"serviceName",
",",
"waitSecs",
"=",
"waitSecs",
")",
"else",
":",
"StopService",
"(",
"serviceName",
")",
"except",
"win32service",
".",
"error",
"as",
"exc",
":",
"print",
"(",
"'Error stopping service: %s (%d)'",
"%",
"(",
"exc",
".",
"strerror",
",",
"exc",
".",
"winerror",
")",
")",
"err",
"=",
"exc",
".",
"winerror",
"if",
"(",
"not",
"knownArg",
")",
":",
"err",
"=",
"(",
"-",
"1",
")",
"print",
"(",
"\"Unknown command - '%s'\"",
"%",
"arg",
")",
"usage",
"(",
")",
"return",
"err"
] | handle command line for a windows service prescribed name that will be called by py2exe . | train | false |
47,655 | def _kde_support(data, bw, gridsize, cut, clip):
support_min = max((data.min() - (bw * cut)), clip[0])
support_max = min((data.max() + (bw * cut)), clip[1])
return np.linspace(support_min, support_max, gridsize)
| [
"def",
"_kde_support",
"(",
"data",
",",
"bw",
",",
"gridsize",
",",
"cut",
",",
"clip",
")",
":",
"support_min",
"=",
"max",
"(",
"(",
"data",
".",
"min",
"(",
")",
"-",
"(",
"bw",
"*",
"cut",
")",
")",
",",
"clip",
"[",
"0",
"]",
")",
"support_max",
"=",
"min",
"(",
"(",
"data",
".",
"max",
"(",
")",
"+",
"(",
"bw",
"*",
"cut",
")",
")",
",",
"clip",
"[",
"1",
"]",
")",
"return",
"np",
".",
"linspace",
"(",
"support_min",
",",
"support_max",
",",
"gridsize",
")"
] | establish support for a kernel density estimate . | train | false |
47,656 | def write_config(config, path, variables=None):
f = file(path, 'w')
try:
f.write('<?xml version="1.0"?>\n<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>\n<configuration>\n')
keys = ((variables and (variables,)) or (config.keys(),))[0]
for name in keys:
value = config[name]
f.write(' <property>\n')
f.write((' <name>%s</name>\n' % name))
f.write((' <value>%s</value>\n' % value))
f.write(' </property>\n')
f.write('</configuration>\n')
finally:
f.close()
| [
"def",
"write_config",
"(",
"config",
",",
"path",
",",
"variables",
"=",
"None",
")",
":",
"f",
"=",
"file",
"(",
"path",
",",
"'w'",
")",
"try",
":",
"f",
".",
"write",
"(",
"'<?xml version=\"1.0\"?>\\n<?xml-stylesheet type=\"text/xsl\" href=\"configuration.xsl\"?>\\n<configuration>\\n'",
")",
"keys",
"=",
"(",
"(",
"variables",
"and",
"(",
"variables",
",",
")",
")",
"or",
"(",
"config",
".",
"keys",
"(",
")",
",",
")",
")",
"[",
"0",
"]",
"for",
"name",
"in",
"keys",
":",
"value",
"=",
"config",
"[",
"name",
"]",
"f",
".",
"write",
"(",
"' <property>\\n'",
")",
"f",
".",
"write",
"(",
"(",
"' <name>%s</name>\\n'",
"%",
"name",
")",
")",
"f",
".",
"write",
"(",
"(",
"' <value>%s</value>\\n'",
"%",
"value",
")",
")",
"f",
".",
"write",
"(",
"' </property>\\n'",
")",
"f",
".",
"write",
"(",
"'</configuration>\\n'",
")",
"finally",
":",
"f",
".",
"close",
"(",
")"
] | minimal utility to write hadoop-style configuration from a configuration map . | train | false |
47,657 | def InitSecretsForTest():
GetSharedSecretsManager(can_prompt=False)
| [
"def",
"InitSecretsForTest",
"(",
")",
":",
"GetSharedSecretsManager",
"(",
"can_prompt",
"=",
"False",
")"
] | init secrets for test . | train | false |
47,658 | def URLFreqRespTime(urlstatsdict):
resptime = []
binner = _ExponentialBinner(10, 2)
maxbins = 0
for (url, urlstats) in urlstatsdict.iteritems():
urlresptime = sorted(urlstats.GetResponseTimeList())
urlbin = binner.Bin(urlresptime)
maxbins = max(maxbins, len(urlbin))
resptime.append((url, urlresptime, urlbin))
resptime.sort(key=(lambda triple: len(triple[1])), reverse=True)
intervals = binner.Intervals(maxbins)
return (resptime, intervals)
| [
"def",
"URLFreqRespTime",
"(",
"urlstatsdict",
")",
":",
"resptime",
"=",
"[",
"]",
"binner",
"=",
"_ExponentialBinner",
"(",
"10",
",",
"2",
")",
"maxbins",
"=",
"0",
"for",
"(",
"url",
",",
"urlstats",
")",
"in",
"urlstatsdict",
".",
"iteritems",
"(",
")",
":",
"urlresptime",
"=",
"sorted",
"(",
"urlstats",
".",
"GetResponseTimeList",
"(",
")",
")",
"urlbin",
"=",
"binner",
".",
"Bin",
"(",
"urlresptime",
")",
"maxbins",
"=",
"max",
"(",
"maxbins",
",",
"len",
"(",
"urlbin",
")",
")",
"resptime",
".",
"append",
"(",
"(",
"url",
",",
"urlresptime",
",",
"urlbin",
")",
")",
"resptime",
".",
"sort",
"(",
"key",
"=",
"(",
"lambda",
"triple",
":",
"len",
"(",
"triple",
"[",
"1",
"]",
")",
")",
",",
"reverse",
"=",
"True",
")",
"intervals",
"=",
"binner",
".",
"Intervals",
"(",
"maxbins",
")",
"return",
"(",
"resptime",
",",
"intervals",
")"
] | computes request counts in different response time ranges for histograms . | train | false |
47,659 | def exec_payload(shell_obj, payload_name, args=(), use_api=False):
payload_inst = get_payload_instance(payload_name, shell_obj)
if use_api:
result = payload_inst.run_api(*args)
else:
result = payload_inst.run(*args)
return result
| [
"def",
"exec_payload",
"(",
"shell_obj",
",",
"payload_name",
",",
"args",
"=",
"(",
")",
",",
"use_api",
"=",
"False",
")",
":",
"payload_inst",
"=",
"get_payload_instance",
"(",
"payload_name",
",",
"shell_obj",
")",
"if",
"use_api",
":",
"result",
"=",
"payload_inst",
".",
"run_api",
"(",
"*",
"args",
")",
"else",
":",
"result",
"=",
"payload_inst",
".",
"run",
"(",
"*",
"args",
")",
"return",
"result"
] | now i execute the payload . | train | false |
47,661 | def remove_indents(txt):
txt = re.sub('(?miu)^\\s+', '', txt)
return txt
| [
"def",
"remove_indents",
"(",
"txt",
")",
":",
"txt",
"=",
"re",
".",
"sub",
"(",
"'(?miu)^\\\\s+'",
",",
"''",
",",
"txt",
")",
"return",
"txt"
] | remove whitespace at the beginning of each line . | train | false |
47,662 | def rand_uuid():
return uuidutils.generate_uuid()
| [
"def",
"rand_uuid",
"(",
")",
":",
"return",
"uuidutils",
".",
"generate_uuid",
"(",
")"
] | generate a random uuid string :return: a random uuid :rtype: string . | train | false |
47,663 | def duck_type_collection(specimen, default=None):
if hasattr(specimen, '__emulates__'):
if ((specimen.__emulates__ is not None) and issubclass(specimen.__emulates__, set)):
return set
else:
return specimen.__emulates__
isa = ((isinstance(specimen, type) and issubclass) or isinstance)
if isa(specimen, list):
return list
elif isa(specimen, set):
return set
elif isa(specimen, dict):
return dict
if hasattr(specimen, 'append'):
return list
elif hasattr(specimen, 'add'):
return set
elif hasattr(specimen, 'set'):
return dict
else:
return default
| [
"def",
"duck_type_collection",
"(",
"specimen",
",",
"default",
"=",
"None",
")",
":",
"if",
"hasattr",
"(",
"specimen",
",",
"'__emulates__'",
")",
":",
"if",
"(",
"(",
"specimen",
".",
"__emulates__",
"is",
"not",
"None",
")",
"and",
"issubclass",
"(",
"specimen",
".",
"__emulates__",
",",
"set",
")",
")",
":",
"return",
"set",
"else",
":",
"return",
"specimen",
".",
"__emulates__",
"isa",
"=",
"(",
"(",
"isinstance",
"(",
"specimen",
",",
"type",
")",
"and",
"issubclass",
")",
"or",
"isinstance",
")",
"if",
"isa",
"(",
"specimen",
",",
"list",
")",
":",
"return",
"list",
"elif",
"isa",
"(",
"specimen",
",",
"set",
")",
":",
"return",
"set",
"elif",
"isa",
"(",
"specimen",
",",
"dict",
")",
":",
"return",
"dict",
"if",
"hasattr",
"(",
"specimen",
",",
"'append'",
")",
":",
"return",
"list",
"elif",
"hasattr",
"(",
"specimen",
",",
"'add'",
")",
":",
"return",
"set",
"elif",
"hasattr",
"(",
"specimen",
",",
"'set'",
")",
":",
"return",
"dict",
"else",
":",
"return",
"default"
] | given an instance or class . | train | false |
47,664 | def BuildToken(request, execution_time):
token = access_control.ACLToken(username=request.user, reason=request.REQ.get('reason', ''), process='GRRAdminUI', expiry=(rdfvalue.RDFDatetime.Now() + execution_time))
for field in ['REMOTE_ADDR', 'HTTP_X_FORWARDED_FOR']:
remote_addr = request.META.get(field, '')
if remote_addr:
token.source_ips.append(remote_addr)
return token
| [
"def",
"BuildToken",
"(",
"request",
",",
"execution_time",
")",
":",
"token",
"=",
"access_control",
".",
"ACLToken",
"(",
"username",
"=",
"request",
".",
"user",
",",
"reason",
"=",
"request",
".",
"REQ",
".",
"get",
"(",
"'reason'",
",",
"''",
")",
",",
"process",
"=",
"'GRRAdminUI'",
",",
"expiry",
"=",
"(",
"rdfvalue",
".",
"RDFDatetime",
".",
"Now",
"(",
")",
"+",
"execution_time",
")",
")",
"for",
"field",
"in",
"[",
"'REMOTE_ADDR'",
",",
"'HTTP_X_FORWARDED_FOR'",
"]",
":",
"remote_addr",
"=",
"request",
".",
"META",
".",
"get",
"(",
"field",
",",
"''",
")",
"if",
"remote_addr",
":",
"token",
".",
"source_ips",
".",
"append",
"(",
"remote_addr",
")",
"return",
"token"
] | build an acltoken from the request . | train | false |
47,665 | def update_order_line_from_product(pricing_context, order_line, product, quantity=1, supplier=None):
if order_line.pk:
raise Exception('set_from_product may not be used on saved lines')
if (not product):
raise Exception('set_from_product may not be used without product')
order_line.supplier = supplier
order_line.type = OrderLineType.PRODUCT
order_line.product = product
order_line.quantity = quantity
order_line.sku = product.sku
order_line.text = (product.safe_translation_getter('name') or product.sku)
order_line.accounting_identifier = product.accounting_identifier
order_line.require_verification = bool(getattr(product, 'require_verification', False))
order_line.verified = False
if pricing_context:
price_info = product.get_price_info(pricing_context, quantity=quantity)
order_line.base_unit_price = price_info.base_unit_price
order_line.discount_amount = price_info.discount_amount
assert (order_line.price == price_info.price)
else:
order_line.base_unit_price_value = 0
order_line.discount_amount_value = 0
| [
"def",
"update_order_line_from_product",
"(",
"pricing_context",
",",
"order_line",
",",
"product",
",",
"quantity",
"=",
"1",
",",
"supplier",
"=",
"None",
")",
":",
"if",
"order_line",
".",
"pk",
":",
"raise",
"Exception",
"(",
"'set_from_product may not be used on saved lines'",
")",
"if",
"(",
"not",
"product",
")",
":",
"raise",
"Exception",
"(",
"'set_from_product may not be used without product'",
")",
"order_line",
".",
"supplier",
"=",
"supplier",
"order_line",
".",
"type",
"=",
"OrderLineType",
".",
"PRODUCT",
"order_line",
".",
"product",
"=",
"product",
"order_line",
".",
"quantity",
"=",
"quantity",
"order_line",
".",
"sku",
"=",
"product",
".",
"sku",
"order_line",
".",
"text",
"=",
"(",
"product",
".",
"safe_translation_getter",
"(",
"'name'",
")",
"or",
"product",
".",
"sku",
")",
"order_line",
".",
"accounting_identifier",
"=",
"product",
".",
"accounting_identifier",
"order_line",
".",
"require_verification",
"=",
"bool",
"(",
"getattr",
"(",
"product",
",",
"'require_verification'",
",",
"False",
")",
")",
"order_line",
".",
"verified",
"=",
"False",
"if",
"pricing_context",
":",
"price_info",
"=",
"product",
".",
"get_price_info",
"(",
"pricing_context",
",",
"quantity",
"=",
"quantity",
")",
"order_line",
".",
"base_unit_price",
"=",
"price_info",
".",
"base_unit_price",
"order_line",
".",
"discount_amount",
"=",
"price_info",
".",
"discount_amount",
"assert",
"(",
"order_line",
".",
"price",
"==",
"price_info",
".",
"price",
")",
"else",
":",
"order_line",
".",
"base_unit_price_value",
"=",
"0",
"order_line",
".",
"discount_amount_value",
"=",
"0"
] | update orderline data from a product . | train | false |
47,668 | def _get_numba_ufunc(expr):
if isinstance(expr, Broadcast):
leaves = expr._scalars
expr = expr._scalar_expr
else:
leaves = expr._leaves()
(s, scope) = funcstr(leaves, expr)
scope = dict(((k, (numba.jit(nopython=True)(v) if callable(v) else v)) for (k, v) in scope.items()))
func = eval(s, scope)
sig = compute_signature(expr)
with lock:
ufunc = numba.vectorize([sig], nopython=True)(func)
return ufunc
| [
"def",
"_get_numba_ufunc",
"(",
"expr",
")",
":",
"if",
"isinstance",
"(",
"expr",
",",
"Broadcast",
")",
":",
"leaves",
"=",
"expr",
".",
"_scalars",
"expr",
"=",
"expr",
".",
"_scalar_expr",
"else",
":",
"leaves",
"=",
"expr",
".",
"_leaves",
"(",
")",
"(",
"s",
",",
"scope",
")",
"=",
"funcstr",
"(",
"leaves",
",",
"expr",
")",
"scope",
"=",
"dict",
"(",
"(",
"(",
"k",
",",
"(",
"numba",
".",
"jit",
"(",
"nopython",
"=",
"True",
")",
"(",
"v",
")",
"if",
"callable",
"(",
"v",
")",
"else",
"v",
")",
")",
"for",
"(",
"k",
",",
"v",
")",
"in",
"scope",
".",
"items",
"(",
")",
")",
")",
"func",
"=",
"eval",
"(",
"s",
",",
"scope",
")",
"sig",
"=",
"compute_signature",
"(",
"expr",
")",
"with",
"lock",
":",
"ufunc",
"=",
"numba",
".",
"vectorize",
"(",
"[",
"sig",
"]",
",",
"nopython",
"=",
"True",
")",
"(",
"func",
")",
"return",
"ufunc"
] | construct a numba ufunc from a blaze expression parameters expr : blaze . | train | false |
47,669 | def write_double_matrix(fid, kind, mat):
FIFFT_MATRIX = (1 << 30)
FIFFT_MATRIX_DOUBLE = (FIFF.FIFFT_DOUBLE | FIFFT_MATRIX)
data_size = ((8 * mat.size) + (4 * (mat.ndim + 1)))
fid.write(np.array(kind, dtype='>i4').tostring())
fid.write(np.array(FIFFT_MATRIX_DOUBLE, dtype='>i4').tostring())
fid.write(np.array(data_size, dtype='>i4').tostring())
fid.write(np.array(FIFF.FIFFV_NEXT_SEQ, dtype='>i4').tostring())
fid.write(np.array(mat, dtype='>f8').tostring())
dims = np.empty((mat.ndim + 1), dtype=np.int32)
dims[:mat.ndim] = mat.shape[::(-1)]
dims[(-1)] = mat.ndim
fid.write(np.array(dims, dtype='>i4').tostring())
check_fiff_length(fid)
| [
"def",
"write_double_matrix",
"(",
"fid",
",",
"kind",
",",
"mat",
")",
":",
"FIFFT_MATRIX",
"=",
"(",
"1",
"<<",
"30",
")",
"FIFFT_MATRIX_DOUBLE",
"=",
"(",
"FIFF",
".",
"FIFFT_DOUBLE",
"|",
"FIFFT_MATRIX",
")",
"data_size",
"=",
"(",
"(",
"8",
"*",
"mat",
".",
"size",
")",
"+",
"(",
"4",
"*",
"(",
"mat",
".",
"ndim",
"+",
"1",
")",
")",
")",
"fid",
".",
"write",
"(",
"np",
".",
"array",
"(",
"kind",
",",
"dtype",
"=",
"'>i4'",
")",
".",
"tostring",
"(",
")",
")",
"fid",
".",
"write",
"(",
"np",
".",
"array",
"(",
"FIFFT_MATRIX_DOUBLE",
",",
"dtype",
"=",
"'>i4'",
")",
".",
"tostring",
"(",
")",
")",
"fid",
".",
"write",
"(",
"np",
".",
"array",
"(",
"data_size",
",",
"dtype",
"=",
"'>i4'",
")",
".",
"tostring",
"(",
")",
")",
"fid",
".",
"write",
"(",
"np",
".",
"array",
"(",
"FIFF",
".",
"FIFFV_NEXT_SEQ",
",",
"dtype",
"=",
"'>i4'",
")",
".",
"tostring",
"(",
")",
")",
"fid",
".",
"write",
"(",
"np",
".",
"array",
"(",
"mat",
",",
"dtype",
"=",
"'>f8'",
")",
".",
"tostring",
"(",
")",
")",
"dims",
"=",
"np",
".",
"empty",
"(",
"(",
"mat",
".",
"ndim",
"+",
"1",
")",
",",
"dtype",
"=",
"np",
".",
"int32",
")",
"dims",
"[",
":",
"mat",
".",
"ndim",
"]",
"=",
"mat",
".",
"shape",
"[",
":",
":",
"(",
"-",
"1",
")",
"]",
"dims",
"[",
"(",
"-",
"1",
")",
"]",
"=",
"mat",
".",
"ndim",
"fid",
".",
"write",
"(",
"np",
".",
"array",
"(",
"dims",
",",
"dtype",
"=",
"'>i4'",
")",
".",
"tostring",
"(",
")",
")",
"check_fiff_length",
"(",
"fid",
")"
] | write a double-precision floating-point matrix tag . | train | false |
47,670 | def fix_index(builder, idx, size):
is_negative = builder.icmp_signed('<', idx, ir.Constant(size.type, 0))
wrapped_index = builder.add(idx, size)
return builder.select(is_negative, wrapped_index, idx)
| [
"def",
"fix_index",
"(",
"builder",
",",
"idx",
",",
"size",
")",
":",
"is_negative",
"=",
"builder",
".",
"icmp_signed",
"(",
"'<'",
",",
"idx",
",",
"ir",
".",
"Constant",
"(",
"size",
".",
"type",
",",
"0",
")",
")",
"wrapped_index",
"=",
"builder",
".",
"add",
"(",
"idx",
",",
"size",
")",
"return",
"builder",
".",
"select",
"(",
"is_negative",
",",
"wrapped_index",
",",
"idx",
")"
] | fix negative index by adding *size* to it . | train | false |
47,671 | def route_table_exists(route_table_id=None, name=None, route_table_name=None, tags=None, region=None, key=None, keyid=None, profile=None):
if name:
log.warning('boto_vpc.route_table_exists: name parameter is deprecated use route_table_name instead.')
route_table_name = name
return resource_exists('route_table', name=route_table_name, resource_id=route_table_id, tags=tags, region=region, key=key, keyid=keyid, profile=profile)
| [
"def",
"route_table_exists",
"(",
"route_table_id",
"=",
"None",
",",
"name",
"=",
"None",
",",
"route_table_name",
"=",
"None",
",",
"tags",
"=",
"None",
",",
"region",
"=",
"None",
",",
"key",
"=",
"None",
",",
"keyid",
"=",
"None",
",",
"profile",
"=",
"None",
")",
":",
"if",
"name",
":",
"log",
".",
"warning",
"(",
"'boto_vpc.route_table_exists: name parameter is deprecated use route_table_name instead.'",
")",
"route_table_name",
"=",
"name",
"return",
"resource_exists",
"(",
"'route_table'",
",",
"name",
"=",
"route_table_name",
",",
"resource_id",
"=",
"route_table_id",
",",
"tags",
"=",
"tags",
",",
"region",
"=",
"region",
",",
"key",
"=",
"key",
",",
"keyid",
"=",
"keyid",
",",
"profile",
"=",
"profile",
")"
] | checks if a route table exists . | train | true |
47,672 | def HTML(html):
return markupsafe.Markup(html)
| [
"def",
"HTML",
"(",
"html",
")",
":",
"return",
"markupsafe",
".",
"Markup",
"(",
"html",
")"
] | mark a string as already html . | train | false |
47,673 | def rm_r(path):
if os.path.isdir(path):
shutil.rmtree(path)
elif os.path.isfile(path):
unlink(path)
| [
"def",
"rm_r",
"(",
"path",
")",
":",
"if",
"os",
".",
"path",
".",
"isdir",
"(",
"path",
")",
":",
"shutil",
".",
"rmtree",
"(",
"path",
")",
"elif",
"os",
".",
"path",
".",
"isfile",
"(",
"path",
")",
":",
"unlink",
"(",
"path",
")"
] | like rm -r command . | train | false |
47,674 | def volume_type_extra_specs_update_or_create(context, volume_type_id, extra_specs):
IMPL.volume_type_extra_specs_update_or_create(context, volume_type_id, extra_specs)
| [
"def",
"volume_type_extra_specs_update_or_create",
"(",
"context",
",",
"volume_type_id",
",",
"extra_specs",
")",
":",
"IMPL",
".",
"volume_type_extra_specs_update_or_create",
"(",
"context",
",",
"volume_type_id",
",",
"extra_specs",
")"
] | create or update volume type extra specs . | train | false |
47,677 | def _xfs_prune_output(out, uuid):
data = {}
cnt = []
cutpoint = False
for line in [l.strip() for l in out.split('\n') if l]:
if line.startswith('-'):
if cutpoint:
break
else:
cutpoint = True
continue
if cutpoint:
cnt.append(line)
for kset in [e for e in cnt[1:] if (':' in e)]:
(key, val) = [t.strip() for t in kset.split(':', 1)]
data[key.lower().replace(' ', '_')] = val
return (((data.get('uuid') == uuid) and data) or {})
| [
"def",
"_xfs_prune_output",
"(",
"out",
",",
"uuid",
")",
":",
"data",
"=",
"{",
"}",
"cnt",
"=",
"[",
"]",
"cutpoint",
"=",
"False",
"for",
"line",
"in",
"[",
"l",
".",
"strip",
"(",
")",
"for",
"l",
"in",
"out",
".",
"split",
"(",
"'\\n'",
")",
"if",
"l",
"]",
":",
"if",
"line",
".",
"startswith",
"(",
"'-'",
")",
":",
"if",
"cutpoint",
":",
"break",
"else",
":",
"cutpoint",
"=",
"True",
"continue",
"if",
"cutpoint",
":",
"cnt",
".",
"append",
"(",
"line",
")",
"for",
"kset",
"in",
"[",
"e",
"for",
"e",
"in",
"cnt",
"[",
"1",
":",
"]",
"if",
"(",
"':'",
"in",
"e",
")",
"]",
":",
"(",
"key",
",",
"val",
")",
"=",
"[",
"t",
".",
"strip",
"(",
")",
"for",
"t",
"in",
"kset",
".",
"split",
"(",
"':'",
",",
"1",
")",
"]",
"data",
"[",
"key",
".",
"lower",
"(",
")",
".",
"replace",
"(",
"' '",
",",
"'_'",
")",
"]",
"=",
"val",
"return",
"(",
"(",
"(",
"data",
".",
"get",
"(",
"'uuid'",
")",
"==",
"uuid",
")",
"and",
"data",
")",
"or",
"{",
"}",
")"
] | parse prune output . | train | true |
47,680 | def channel_session_user_from_http(func):
@http_session_user
@channel_session
@functools.wraps(func)
def inner(message, *args, **kwargs):
if (message.http_session is not None):
transfer_user(message.http_session, message.channel_session)
return func(message, *args, **kwargs)
return inner
| [
"def",
"channel_session_user_from_http",
"(",
"func",
")",
":",
"@",
"http_session_user",
"@",
"channel_session",
"@",
"functools",
".",
"wraps",
"(",
"func",
")",
"def",
"inner",
"(",
"message",
",",
"*",
"args",
",",
"**",
"kwargs",
")",
":",
"if",
"(",
"message",
".",
"http_session",
"is",
"not",
"None",
")",
":",
"transfer_user",
"(",
"message",
".",
"http_session",
",",
"message",
".",
"channel_session",
")",
"return",
"func",
"(",
"message",
",",
"*",
"args",
",",
"**",
"kwargs",
")",
"return",
"inner"
] | decorator that automatically transfers the user from http sessions to channel-based sessions . | train | false |
47,683 | def install_thread_excepthook():
import sys
run_old = Thread.run
def run(*args, **kwargs):
try:
run_old(*args, **kwargs)
except (KeyboardInterrupt, SystemExit):
raise
except:
sys.excepthook(*sys.exc_info())
Thread.run = run
| [
"def",
"install_thread_excepthook",
"(",
")",
":",
"import",
"sys",
"run_old",
"=",
"Thread",
".",
"run",
"def",
"run",
"(",
"*",
"args",
",",
"**",
"kwargs",
")",
":",
"try",
":",
"run_old",
"(",
"*",
"args",
",",
"**",
"kwargs",
")",
"except",
"(",
"KeyboardInterrupt",
",",
"SystemExit",
")",
":",
"raise",
"except",
":",
"sys",
".",
"excepthook",
"(",
"*",
"sys",
".",
"exc_info",
"(",
")",
")",
"Thread",
".",
"run",
"=",
"run"
] | workaround for sys . | train | false |
47,684 | def get_dataset_root():
return _dataset_root
| [
"def",
"get_dataset_root",
"(",
")",
":",
"return",
"_dataset_root"
] | gets the path to the root directory to download and cache datasets . | train | false |
47,685 | def _dict_from_tcltuple(ttuple, cut_minus=True):
opt_start = (1 if cut_minus else 0)
retdict = {}
it = iter(ttuple)
for (opt, val) in zip(it, it):
retdict[str(opt)[opt_start:]] = val
return tclobjs_to_py(retdict)
| [
"def",
"_dict_from_tcltuple",
"(",
"ttuple",
",",
"cut_minus",
"=",
"True",
")",
":",
"opt_start",
"=",
"(",
"1",
"if",
"cut_minus",
"else",
"0",
")",
"retdict",
"=",
"{",
"}",
"it",
"=",
"iter",
"(",
"ttuple",
")",
"for",
"(",
"opt",
",",
"val",
")",
"in",
"zip",
"(",
"it",
",",
"it",
")",
":",
"retdict",
"[",
"str",
"(",
"opt",
")",
"[",
"opt_start",
":",
"]",
"]",
"=",
"val",
"return",
"tclobjs_to_py",
"(",
"retdict",
")"
] | break tuple in pairs . | train | false |
47,687 | def colordiff(a, b, highlight='text_highlight'):
if config['ui']['color']:
return _colordiff(a, b, highlight)
else:
return (six.text_type(a), six.text_type(b))
| [
"def",
"colordiff",
"(",
"a",
",",
"b",
",",
"highlight",
"=",
"'text_highlight'",
")",
":",
"if",
"config",
"[",
"'ui'",
"]",
"[",
"'color'",
"]",
":",
"return",
"_colordiff",
"(",
"a",
",",
"b",
",",
"highlight",
")",
"else",
":",
"return",
"(",
"six",
".",
"text_type",
"(",
"a",
")",
",",
"six",
".",
"text_type",
"(",
"b",
")",
")"
] | colorize differences between two values if color is enabled . | train | false |
47,688 | def _select_vs(v, p):
if (v >= 120.0):
return (60, 120, inf)
elif (v >= 60.0):
return (40, 60, 120)
elif (v >= 40.0):
return (30, 40, 60)
elif (v >= 30.0):
return (24, 30, 40)
elif (v >= 24.0):
return (20, 24, 30)
elif (v >= 19.5):
return (19, 20, 24)
if (p >= 0.9):
if (v < 2.5):
return (1, 2, 3)
elif (v < 3.5):
return (2, 3, 4)
vi = int(round(v))
return ((vi - 1), vi, (vi + 1))
| [
"def",
"_select_vs",
"(",
"v",
",",
"p",
")",
":",
"if",
"(",
"v",
">=",
"120.0",
")",
":",
"return",
"(",
"60",
",",
"120",
",",
"inf",
")",
"elif",
"(",
"v",
">=",
"60.0",
")",
":",
"return",
"(",
"40",
",",
"60",
",",
"120",
")",
"elif",
"(",
"v",
">=",
"40.0",
")",
":",
"return",
"(",
"30",
",",
"40",
",",
"60",
")",
"elif",
"(",
"v",
">=",
"30.0",
")",
":",
"return",
"(",
"24",
",",
"30",
",",
"40",
")",
"elif",
"(",
"v",
">=",
"24.0",
")",
":",
"return",
"(",
"20",
",",
"24",
",",
"30",
")",
"elif",
"(",
"v",
">=",
"19.5",
")",
":",
"return",
"(",
"19",
",",
"20",
",",
"24",
")",
"if",
"(",
"p",
">=",
"0.9",
")",
":",
"if",
"(",
"v",
"<",
"2.5",
")",
":",
"return",
"(",
"1",
",",
"2",
",",
"3",
")",
"elif",
"(",
"v",
"<",
"3.5",
")",
":",
"return",
"(",
"2",
",",
"3",
",",
"4",
")",
"vi",
"=",
"int",
"(",
"round",
"(",
"v",
")",
")",
"return",
"(",
"(",
"vi",
"-",
"1",
")",
",",
"vi",
",",
"(",
"vi",
"+",
"1",
")",
")"
] | returns the points to use for interpolating v . | train | true |
47,693 | def get_account_name(account_type=None, root_type=None, is_group=None, account_currency=None, company=None):
return frappe.db.get_value(u'Account', {u'account_type': (account_type or u''), u'root_type': (root_type or u''), u'is_group': (is_group or 0), u'account_currency': (account_currency or frappe.defaults.get_defaults().currency), u'company': (company or frappe.defaults.get_defaults().company)}, u'name')
| [
"def",
"get_account_name",
"(",
"account_type",
"=",
"None",
",",
"root_type",
"=",
"None",
",",
"is_group",
"=",
"None",
",",
"account_currency",
"=",
"None",
",",
"company",
"=",
"None",
")",
":",
"return",
"frappe",
".",
"db",
".",
"get_value",
"(",
"u'Account'",
",",
"{",
"u'account_type'",
":",
"(",
"account_type",
"or",
"u''",
")",
",",
"u'root_type'",
":",
"(",
"root_type",
"or",
"u''",
")",
",",
"u'is_group'",
":",
"(",
"is_group",
"or",
"0",
")",
",",
"u'account_currency'",
":",
"(",
"account_currency",
"or",
"frappe",
".",
"defaults",
".",
"get_defaults",
"(",
")",
".",
"currency",
")",
",",
"u'company'",
":",
"(",
"company",
"or",
"frappe",
".",
"defaults",
".",
"get_defaults",
"(",
")",
".",
"company",
")",
"}",
",",
"u'name'",
")"
] | return account based on matching conditions . | train | false |
47,694 | @login_required
def choose_transcripts(request):
response = {'status': 'Error', 'subs': ''}
try:
(data, videos, item) = _validate_transcripts_data(request)
except TranscriptsRequestValidationException as e:
return error_response(response, e.message)
html5_id = data.get('html5_id')
html5_id_to_remove = [x for x in videos['html5'] if (x != html5_id)]
if html5_id_to_remove:
remove_subs_from_store(html5_id_to_remove, item)
if (item.sub != html5_id):
item.sub = html5_id
item.save_with_metadata(request.user)
response = {'status': 'Success', 'subs': item.sub}
return JsonResponse(response)
| [
"@",
"login_required",
"def",
"choose_transcripts",
"(",
"request",
")",
":",
"response",
"=",
"{",
"'status'",
":",
"'Error'",
",",
"'subs'",
":",
"''",
"}",
"try",
":",
"(",
"data",
",",
"videos",
",",
"item",
")",
"=",
"_validate_transcripts_data",
"(",
"request",
")",
"except",
"TranscriptsRequestValidationException",
"as",
"e",
":",
"return",
"error_response",
"(",
"response",
",",
"e",
".",
"message",
")",
"html5_id",
"=",
"data",
".",
"get",
"(",
"'html5_id'",
")",
"html5_id_to_remove",
"=",
"[",
"x",
"for",
"x",
"in",
"videos",
"[",
"'html5'",
"]",
"if",
"(",
"x",
"!=",
"html5_id",
")",
"]",
"if",
"html5_id_to_remove",
":",
"remove_subs_from_store",
"(",
"html5_id_to_remove",
",",
"item",
")",
"if",
"(",
"item",
".",
"sub",
"!=",
"html5_id",
")",
":",
"item",
".",
"sub",
"=",
"html5_id",
"item",
".",
"save_with_metadata",
"(",
"request",
".",
"user",
")",
"response",
"=",
"{",
"'status'",
":",
"'Success'",
",",
"'subs'",
":",
"item",
".",
"sub",
"}",
"return",
"JsonResponse",
"(",
"response",
")"
] | replaces html5 subtitles . | train | false |
47,697 | def test_no_data_with_empty_series(Chart):
chart = Chart()
chart.add('Serie1', [])
chart.add('Serie2', [])
q = chart.render_pyquery()
assert (q('.text-overlay text').text() == 'No data')
| [
"def",
"test_no_data_with_empty_series",
"(",
"Chart",
")",
":",
"chart",
"=",
"Chart",
"(",
")",
"chart",
".",
"add",
"(",
"'Serie1'",
",",
"[",
"]",
")",
"chart",
".",
"add",
"(",
"'Serie2'",
",",
"[",
"]",
")",
"q",
"=",
"chart",
".",
"render_pyquery",
"(",
")",
"assert",
"(",
"q",
"(",
"'.text-overlay text'",
")",
".",
"text",
"(",
")",
"==",
"'No data'",
")"
] | test no data for 2 empty series . | train | false |
47,699 | def evaluate_template(text, install_environment):
return Template(text).safe_substitute(get_env_var_values(install_environment))
| [
"def",
"evaluate_template",
"(",
"text",
",",
"install_environment",
")",
":",
"return",
"Template",
"(",
"text",
")",
".",
"safe_substitute",
"(",
"get_env_var_values",
"(",
"install_environment",
")",
")"
] | substitute variables defined in xml blocks from dependencies file . | train | false |
47,700 | def to_hex(num):
if (num < 0):
return ('-0x%x' % (- num))
else:
return ('0x%x' % num)
| [
"def",
"to_hex",
"(",
"num",
")",
":",
"if",
"(",
"num",
"<",
"0",
")",
":",
"return",
"(",
"'-0x%x'",
"%",
"(",
"-",
"num",
")",
")",
"else",
":",
"return",
"(",
"'0x%x'",
"%",
"num",
")"
] | convert c to a hex color . | train | false |
47,701 | def read_mri_cfg(subject, subjects_dir=None):
subjects_dir = get_subjects_dir(subjects_dir, raise_error=True)
fname = os.path.join(subjects_dir, subject, 'MRI scaling parameters.cfg')
if (not os.path.exists(fname)):
raise IOError(('%r does not seem to be a scaled mri subject: %r does not exist.' % (subject, fname)))
logger.info(('Reading MRI cfg file %s' % fname))
config = configparser.RawConfigParser()
config.read(fname)
n_params = config.getint('MRI Scaling', 'n_params')
if (n_params == 1):
scale = config.getfloat('MRI Scaling', 'scale')
elif (n_params == 3):
scale_str = config.get('MRI Scaling', 'scale')
scale = np.array([float(s) for s in scale_str.split()])
else:
raise ValueError(('Invalid n_params value in MRI cfg: %i' % n_params))
out = {'subject_from': config.get('MRI Scaling', 'subject_from'), 'n_params': n_params, 'scale': scale}
return out
| [
"def",
"read_mri_cfg",
"(",
"subject",
",",
"subjects_dir",
"=",
"None",
")",
":",
"subjects_dir",
"=",
"get_subjects_dir",
"(",
"subjects_dir",
",",
"raise_error",
"=",
"True",
")",
"fname",
"=",
"os",
".",
"path",
".",
"join",
"(",
"subjects_dir",
",",
"subject",
",",
"'MRI scaling parameters.cfg'",
")",
"if",
"(",
"not",
"os",
".",
"path",
".",
"exists",
"(",
"fname",
")",
")",
":",
"raise",
"IOError",
"(",
"(",
"'%r does not seem to be a scaled mri subject: %r does not exist.'",
"%",
"(",
"subject",
",",
"fname",
")",
")",
")",
"logger",
".",
"info",
"(",
"(",
"'Reading MRI cfg file %s'",
"%",
"fname",
")",
")",
"config",
"=",
"configparser",
".",
"RawConfigParser",
"(",
")",
"config",
".",
"read",
"(",
"fname",
")",
"n_params",
"=",
"config",
".",
"getint",
"(",
"'MRI Scaling'",
",",
"'n_params'",
")",
"if",
"(",
"n_params",
"==",
"1",
")",
":",
"scale",
"=",
"config",
".",
"getfloat",
"(",
"'MRI Scaling'",
",",
"'scale'",
")",
"elif",
"(",
"n_params",
"==",
"3",
")",
":",
"scale_str",
"=",
"config",
".",
"get",
"(",
"'MRI Scaling'",
",",
"'scale'",
")",
"scale",
"=",
"np",
".",
"array",
"(",
"[",
"float",
"(",
"s",
")",
"for",
"s",
"in",
"scale_str",
".",
"split",
"(",
")",
"]",
")",
"else",
":",
"raise",
"ValueError",
"(",
"(",
"'Invalid n_params value in MRI cfg: %i'",
"%",
"n_params",
")",
")",
"out",
"=",
"{",
"'subject_from'",
":",
"config",
".",
"get",
"(",
"'MRI Scaling'",
",",
"'subject_from'",
")",
",",
"'n_params'",
":",
"n_params",
",",
"'scale'",
":",
"scale",
"}",
"return",
"out"
] | read information from the cfg file of a scaled mri brain . | train | false |
47,704 | def _sanitize_text_helper(text, valid_characters=valid_chars, character_map=mapped_chars, invalid_character='X'):
out = []
for c in text:
if (c in valid_characters):
out.append(c)
elif (c in character_map):
out.append(character_map[c])
else:
out.append(invalid_character)
return ''.join(out)
| [
"def",
"_sanitize_text_helper",
"(",
"text",
",",
"valid_characters",
"=",
"valid_chars",
",",
"character_map",
"=",
"mapped_chars",
",",
"invalid_character",
"=",
"'X'",
")",
":",
"out",
"=",
"[",
"]",
"for",
"c",
"in",
"text",
":",
"if",
"(",
"c",
"in",
"valid_characters",
")",
":",
"out",
".",
"append",
"(",
"c",
")",
"elif",
"(",
"c",
"in",
"character_map",
")",
":",
"out",
".",
"append",
"(",
"character_map",
"[",
"c",
"]",
")",
"else",
":",
"out",
".",
"append",
"(",
"invalid_character",
")",
"return",
"''",
".",
"join",
"(",
"out",
")"
] | restricts the characters that are allowed in a string . | train | false |
47,706 | def get_image_download_script(caller):
if (caller == 'iplot'):
check_start = "if(document.readyState == 'complete') {{"
check_end = '}}'
elif (caller == 'plot'):
check_start = ''
check_end = ''
else:
raise ValueError('caller should only be one of `iplot` or `plot`')
return (((("<script>function downloadimage(format, height, width, filename) {{var p = document.getElementById('{plot_id}');Plotly.downloadImage(p, {{format: format, height: height, width: width, filename: filename}});}};" + check_start) + "{{downloadimage('{format}', {height}, {width}, '{filename}');}}") + check_end) + '</script>')
| [
"def",
"get_image_download_script",
"(",
"caller",
")",
":",
"if",
"(",
"caller",
"==",
"'iplot'",
")",
":",
"check_start",
"=",
"\"if(document.readyState == 'complete') {{\"",
"check_end",
"=",
"'}}'",
"elif",
"(",
"caller",
"==",
"'plot'",
")",
":",
"check_start",
"=",
"''",
"check_end",
"=",
"''",
"else",
":",
"raise",
"ValueError",
"(",
"'caller should only be one of `iplot` or `plot`'",
")",
"return",
"(",
"(",
"(",
"(",
"\"<script>function downloadimage(format, height, width, filename) {{var p = document.getElementById('{plot_id}');Plotly.downloadImage(p, {{format: format, height: height, width: width, filename: filename}});}};\"",
"+",
"check_start",
")",
"+",
"\"{{downloadimage('{format}', {height}, {width}, '{filename}');}}\"",
")",
"+",
"check_end",
")",
"+",
"'</script>'",
")"
] | this function will return a script that will download an image of a plotly plot . | train | false |
47,707 | def _name(expr):
if (expr in name_dict):
return name_dict[expr]
result = base = (expr._name or '_')
if (result in seen_names):
for i in itertools.count(1):
result = ('%s_%d' % (base, i))
if (result not in seen_names):
break
seen_names.add(result)
name_dict[expr] = result
return result
| [
"def",
"_name",
"(",
"expr",
")",
":",
"if",
"(",
"expr",
"in",
"name_dict",
")",
":",
"return",
"name_dict",
"[",
"expr",
"]",
"result",
"=",
"base",
"=",
"(",
"expr",
".",
"_name",
"or",
"'_'",
")",
"if",
"(",
"result",
"in",
"seen_names",
")",
":",
"for",
"i",
"in",
"itertools",
".",
"count",
"(",
"1",
")",
":",
"result",
"=",
"(",
"'%s_%d'",
"%",
"(",
"base",
",",
"i",
")",
")",
"if",
"(",
"result",
"not",
"in",
"seen_names",
")",
":",
"break",
"seen_names",
".",
"add",
"(",
"result",
")",
"name_dict",
"[",
"expr",
"]",
"=",
"result",
"return",
"result"
] | a unique and deterministic name for an expression . | train | false |
47,708 | def _to_camel_case(label, divider='_', joiner=' '):
words = []
for entry in label.split(divider):
if (len(entry) == 0):
words.append('')
elif (len(entry) == 1):
words.append(entry.upper())
else:
words.append((entry[0].upper() + entry[1:].lower()))
return joiner.join(words)
| [
"def",
"_to_camel_case",
"(",
"label",
",",
"divider",
"=",
"'_'",
",",
"joiner",
"=",
"' '",
")",
":",
"words",
"=",
"[",
"]",
"for",
"entry",
"in",
"label",
".",
"split",
"(",
"divider",
")",
":",
"if",
"(",
"len",
"(",
"entry",
")",
"==",
"0",
")",
":",
"words",
".",
"append",
"(",
"''",
")",
"elif",
"(",
"len",
"(",
"entry",
")",
"==",
"1",
")",
":",
"words",
".",
"append",
"(",
"entry",
".",
"upper",
"(",
")",
")",
"else",
":",
"words",
".",
"append",
"(",
"(",
"entry",
"[",
"0",
"]",
".",
"upper",
"(",
")",
"+",
"entry",
"[",
"1",
":",
"]",
".",
"lower",
"(",
")",
")",
")",
"return",
"joiner",
".",
"join",
"(",
"words",
")"
] | converts the given string to camel case . | train | false |
47,709 | def _read_volume_info(fobj):
volume_info = dict()
head = np.fromfile(fobj, '>i4', 1)
if (not np.array_equal(head, [20])):
head = np.concatenate([head, np.fromfile(fobj, '>i4', 2)])
if (not np.array_equal(head, [2, 0, 20])):
warnings.warn('Unknown extension code.')
return volume_info
volume_info['head'] = head
for key in ['valid', 'filename', 'volume', 'voxelsize', 'xras', 'yras', 'zras', 'cras']:
pair = fobj.readline().decode('utf-8').split('=')
if ((pair[0].strip() != key) or (len(pair) != 2)):
raise IOError('Error parsing volume info.')
if (key in ('valid', 'filename')):
volume_info[key] = pair[1].strip()
elif (key == 'volume'):
volume_info[key] = np.array(pair[1].split()).astype(int)
else:
volume_info[key] = np.array(pair[1].split()).astype(float)
return volume_info
| [
"def",
"_read_volume_info",
"(",
"fobj",
")",
":",
"volume_info",
"=",
"dict",
"(",
")",
"head",
"=",
"np",
".",
"fromfile",
"(",
"fobj",
",",
"'>i4'",
",",
"1",
")",
"if",
"(",
"not",
"np",
".",
"array_equal",
"(",
"head",
",",
"[",
"20",
"]",
")",
")",
":",
"head",
"=",
"np",
".",
"concatenate",
"(",
"[",
"head",
",",
"np",
".",
"fromfile",
"(",
"fobj",
",",
"'>i4'",
",",
"2",
")",
"]",
")",
"if",
"(",
"not",
"np",
".",
"array_equal",
"(",
"head",
",",
"[",
"2",
",",
"0",
",",
"20",
"]",
")",
")",
":",
"warnings",
".",
"warn",
"(",
"'Unknown extension code.'",
")",
"return",
"volume_info",
"volume_info",
"[",
"'head'",
"]",
"=",
"head",
"for",
"key",
"in",
"[",
"'valid'",
",",
"'filename'",
",",
"'volume'",
",",
"'voxelsize'",
",",
"'xras'",
",",
"'yras'",
",",
"'zras'",
",",
"'cras'",
"]",
":",
"pair",
"=",
"fobj",
".",
"readline",
"(",
")",
".",
"decode",
"(",
"'utf-8'",
")",
".",
"split",
"(",
"'='",
")",
"if",
"(",
"(",
"pair",
"[",
"0",
"]",
".",
"strip",
"(",
")",
"!=",
"key",
")",
"or",
"(",
"len",
"(",
"pair",
")",
"!=",
"2",
")",
")",
":",
"raise",
"IOError",
"(",
"'Error parsing volume info.'",
")",
"if",
"(",
"key",
"in",
"(",
"'valid'",
",",
"'filename'",
")",
")",
":",
"volume_info",
"[",
"key",
"]",
"=",
"pair",
"[",
"1",
"]",
".",
"strip",
"(",
")",
"elif",
"(",
"key",
"==",
"'volume'",
")",
":",
"volume_info",
"[",
"key",
"]",
"=",
"np",
".",
"array",
"(",
"pair",
"[",
"1",
"]",
".",
"split",
"(",
")",
")",
".",
"astype",
"(",
"int",
")",
"else",
":",
"volume_info",
"[",
"key",
"]",
"=",
"np",
".",
"array",
"(",
"pair",
"[",
"1",
"]",
".",
"split",
"(",
")",
")",
".",
"astype",
"(",
"float",
")",
"return",
"volume_info"
] | an implementation of nibabel . | train | false |
47,711 | def json_method(method):
def wrapper(self, url, params=NoDefault, **kw):
content_type = 'application/json'
if (params is not NoDefault):
params = dumps(params, cls=self.JSONEncoder)
kw.update(params=params, content_type=content_type, upload_files=None)
return self._gen_request(method, url, **kw)
subst = dict(lmethod=method.lower(), method=method)
wrapper.__doc__ = (json_method.__doc__ % subst)
wrapper.__name__ = str(('%(lmethod)s_json' % subst))
return wrapper
| [
"def",
"json_method",
"(",
"method",
")",
":",
"def",
"wrapper",
"(",
"self",
",",
"url",
",",
"params",
"=",
"NoDefault",
",",
"**",
"kw",
")",
":",
"content_type",
"=",
"'application/json'",
"if",
"(",
"params",
"is",
"not",
"NoDefault",
")",
":",
"params",
"=",
"dumps",
"(",
"params",
",",
"cls",
"=",
"self",
".",
"JSONEncoder",
")",
"kw",
".",
"update",
"(",
"params",
"=",
"params",
",",
"content_type",
"=",
"content_type",
",",
"upload_files",
"=",
"None",
")",
"return",
"self",
".",
"_gen_request",
"(",
"method",
",",
"url",
",",
"**",
"kw",
")",
"subst",
"=",
"dict",
"(",
"lmethod",
"=",
"method",
".",
"lower",
"(",
")",
",",
"method",
"=",
"method",
")",
"wrapper",
".",
"__doc__",
"=",
"(",
"json_method",
".",
"__doc__",
"%",
"subst",
")",
"wrapper",
".",
"__name__",
"=",
"str",
"(",
"(",
"'%(lmethod)s_json'",
"%",
"subst",
")",
")",
"return",
"wrapper"
] | do a %s request . | train | false |
47,712 | def extendedMeasurementReport():
a = TpPd(pd=6)
b = MessageType(mesType=54)
c = ExtendedMeasurementResults()
packet = ((a / b) / c)
return packet
| [
"def",
"extendedMeasurementReport",
"(",
")",
":",
"a",
"=",
"TpPd",
"(",
"pd",
"=",
"6",
")",
"b",
"=",
"MessageType",
"(",
"mesType",
"=",
"54",
")",
"c",
"=",
"ExtendedMeasurementResults",
"(",
")",
"packet",
"=",
"(",
"(",
"a",
"/",
"b",
")",
"/",
"c",
")",
"return",
"packet"
] | extended measurement report section 9 . | train | true |
47,713 | @core_helper
def remove_url_param(key, value=None, replace=None, controller=None, action=None, extras=None, alternative_url=None):
if isinstance(key, basestring):
keys = [key]
else:
keys = key
params_nopage = [(k, v) for (k, v) in request.params.items() if (k != 'page')]
params = list(params_nopage)
if value:
params.remove((keys[0], value))
else:
for key in keys:
[params.remove((k, v)) for (k, v) in params[:] if (k == key)]
if (replace is not None):
params.append((keys[0], replace))
if alternative_url:
return _url_with_params(alternative_url, params)
return _create_url_with_params(params=params, controller=controller, action=action, extras=extras)
| [
"@",
"core_helper",
"def",
"remove_url_param",
"(",
"key",
",",
"value",
"=",
"None",
",",
"replace",
"=",
"None",
",",
"controller",
"=",
"None",
",",
"action",
"=",
"None",
",",
"extras",
"=",
"None",
",",
"alternative_url",
"=",
"None",
")",
":",
"if",
"isinstance",
"(",
"key",
",",
"basestring",
")",
":",
"keys",
"=",
"[",
"key",
"]",
"else",
":",
"keys",
"=",
"key",
"params_nopage",
"=",
"[",
"(",
"k",
",",
"v",
")",
"for",
"(",
"k",
",",
"v",
")",
"in",
"request",
".",
"params",
".",
"items",
"(",
")",
"if",
"(",
"k",
"!=",
"'page'",
")",
"]",
"params",
"=",
"list",
"(",
"params_nopage",
")",
"if",
"value",
":",
"params",
".",
"remove",
"(",
"(",
"keys",
"[",
"0",
"]",
",",
"value",
")",
")",
"else",
":",
"for",
"key",
"in",
"keys",
":",
"[",
"params",
".",
"remove",
"(",
"(",
"k",
",",
"v",
")",
")",
"for",
"(",
"k",
",",
"v",
")",
"in",
"params",
"[",
":",
"]",
"if",
"(",
"k",
"==",
"key",
")",
"]",
"if",
"(",
"replace",
"is",
"not",
"None",
")",
":",
"params",
".",
"append",
"(",
"(",
"keys",
"[",
"0",
"]",
",",
"replace",
")",
")",
"if",
"alternative_url",
":",
"return",
"_url_with_params",
"(",
"alternative_url",
",",
"params",
")",
"return",
"_create_url_with_params",
"(",
"params",
"=",
"params",
",",
"controller",
"=",
"controller",
",",
"action",
"=",
"action",
",",
"extras",
"=",
"extras",
")"
] | remove one or multiple keys from the current parameters . | train | false |
47,714 | def sync_role_definitions():
logging.info(u'Syncing role definition')
get_or_create_main_db()
create_custom_permissions()
pvms = db.session.query(ab_models.PermissionView).all()
pvms = [p for p in pvms if (p.permission and p.view_menu)]
pvms_to_delete = [p for p in pvms if (not (p.permission and p.view_menu))]
for pvm_to_delete in pvms_to_delete:
sm.get_session.delete(pvm_to_delete)
set_role(u'Admin', pvms, is_admin_pvm)
set_role(u'Alpha', pvms, is_alpha_pvm)
set_role(u'Gamma', pvms, is_gamma_pvm)
set_role(u'granter', pvms, is_granter_pvm)
set_role(u'sql_lab', pvms, is_sql_lab_pvm)
if conf.get(u'PUBLIC_ROLE_LIKE_GAMMA', False):
set_role(u'Public', pvms, is_gamma_pvm)
view_menu_set = db.session.query(models.SqlaTable).all()
create_missing_datasource_perms(view_menu_set)
create_missing_database_perms(view_menu_set)
create_missing_metrics_perm(view_menu_set)
sm.get_session.commit()
| [
"def",
"sync_role_definitions",
"(",
")",
":",
"logging",
".",
"info",
"(",
"u'Syncing role definition'",
")",
"get_or_create_main_db",
"(",
")",
"create_custom_permissions",
"(",
")",
"pvms",
"=",
"db",
".",
"session",
".",
"query",
"(",
"ab_models",
".",
"PermissionView",
")",
".",
"all",
"(",
")",
"pvms",
"=",
"[",
"p",
"for",
"p",
"in",
"pvms",
"if",
"(",
"p",
".",
"permission",
"and",
"p",
".",
"view_menu",
")",
"]",
"pvms_to_delete",
"=",
"[",
"p",
"for",
"p",
"in",
"pvms",
"if",
"(",
"not",
"(",
"p",
".",
"permission",
"and",
"p",
".",
"view_menu",
")",
")",
"]",
"for",
"pvm_to_delete",
"in",
"pvms_to_delete",
":",
"sm",
".",
"get_session",
".",
"delete",
"(",
"pvm_to_delete",
")",
"set_role",
"(",
"u'Admin'",
",",
"pvms",
",",
"is_admin_pvm",
")",
"set_role",
"(",
"u'Alpha'",
",",
"pvms",
",",
"is_alpha_pvm",
")",
"set_role",
"(",
"u'Gamma'",
",",
"pvms",
",",
"is_gamma_pvm",
")",
"set_role",
"(",
"u'granter'",
",",
"pvms",
",",
"is_granter_pvm",
")",
"set_role",
"(",
"u'sql_lab'",
",",
"pvms",
",",
"is_sql_lab_pvm",
")",
"if",
"conf",
".",
"get",
"(",
"u'PUBLIC_ROLE_LIKE_GAMMA'",
",",
"False",
")",
":",
"set_role",
"(",
"u'Public'",
",",
"pvms",
",",
"is_gamma_pvm",
")",
"view_menu_set",
"=",
"db",
".",
"session",
".",
"query",
"(",
"models",
".",
"SqlaTable",
")",
".",
"all",
"(",
")",
"create_missing_datasource_perms",
"(",
"view_menu_set",
")",
"create_missing_database_perms",
"(",
"view_menu_set",
")",
"create_missing_metrics_perm",
"(",
"view_menu_set",
")",
"sm",
".",
"get_session",
".",
"commit",
"(",
")"
] | inits the superset application with security roles and such . | train | false |
47,717 | def get_modal_alert(browser):
WebDriverWait(browser, 6).until(EC.alert_is_present())
return browser.switch_to.alert
| [
"def",
"get_modal_alert",
"(",
"browser",
")",
":",
"WebDriverWait",
"(",
"browser",
",",
"6",
")",
".",
"until",
"(",
"EC",
".",
"alert_is_present",
"(",
")",
")",
"return",
"browser",
".",
"switch_to",
".",
"alert"
] | returns instance of modal alert box shown in browser after waiting for 6 seconds . | train | false |
47,718 | @error.context_aware
def vg_create(vg_name, pv_list, force=False):
error.context(("Creating volume group '%s' by using '%s'" % (vg_name, pv_list)), logging.info)
if vg_check(vg_name):
raise error.TestError(("Volume group '%s' already exist" % vg_name))
if force:
cmd = 'vgcreate -f'
else:
cmd = 'vgcreate'
cmd += (' %s %s' % (vg_name, pv_list))
result = utils.run(cmd)
logging.info(result.stdout.rstrip())
| [
"@",
"error",
".",
"context_aware",
"def",
"vg_create",
"(",
"vg_name",
",",
"pv_list",
",",
"force",
"=",
"False",
")",
":",
"error",
".",
"context",
"(",
"(",
"\"Creating volume group '%s' by using '%s'\"",
"%",
"(",
"vg_name",
",",
"pv_list",
")",
")",
",",
"logging",
".",
"info",
")",
"if",
"vg_check",
"(",
"vg_name",
")",
":",
"raise",
"error",
".",
"TestError",
"(",
"(",
"\"Volume group '%s' already exist\"",
"%",
"vg_name",
")",
")",
"if",
"force",
":",
"cmd",
"=",
"'vgcreate -f'",
"else",
":",
"cmd",
"=",
"'vgcreate'",
"cmd",
"+=",
"(",
"' %s %s'",
"%",
"(",
"vg_name",
",",
"pv_list",
")",
")",
"result",
"=",
"utils",
".",
"run",
"(",
"cmd",
")",
"logging",
".",
"info",
"(",
"result",
".",
"stdout",
".",
"rstrip",
"(",
")",
")"
] | create a volume group by using the block special devices . | train | false |
47,719 | def url_to_path(url):
assert url.startswith('file:'), ('You can only turn file: urls into filenames (not %r)' % url)
(_, netloc, path, _, _) = urllib_parse.urlsplit(url)
if netloc:
netloc = ('\\\\' + netloc)
path = urllib_request.url2pathname((netloc + path))
return path
| [
"def",
"url_to_path",
"(",
"url",
")",
":",
"assert",
"url",
".",
"startswith",
"(",
"'file:'",
")",
",",
"(",
"'You can only turn file: urls into filenames (not %r)'",
"%",
"url",
")",
"(",
"_",
",",
"netloc",
",",
"path",
",",
"_",
",",
"_",
")",
"=",
"urllib_parse",
".",
"urlsplit",
"(",
"url",
")",
"if",
"netloc",
":",
"netloc",
"=",
"(",
"'\\\\\\\\'",
"+",
"netloc",
")",
"path",
"=",
"urllib_request",
".",
"url2pathname",
"(",
"(",
"netloc",
"+",
"path",
")",
")",
"return",
"path"
] | convert a file: url to a path . | train | true |
47,720 | def delete_subscription(topic_name, subscription_name):
pubsub_client = pubsub.Client()
topic = pubsub_client.topic(topic_name)
subscription = topic.subscription(subscription_name)
subscription.delete()
print 'Subscription {} deleted on topic {}.'.format(subscription.name, topic.name)
| [
"def",
"delete_subscription",
"(",
"topic_name",
",",
"subscription_name",
")",
":",
"pubsub_client",
"=",
"pubsub",
".",
"Client",
"(",
")",
"topic",
"=",
"pubsub_client",
".",
"topic",
"(",
"topic_name",
")",
"subscription",
"=",
"topic",
".",
"subscription",
"(",
"subscription_name",
")",
"subscription",
".",
"delete",
"(",
")",
"print",
"'Subscription {} deleted on topic {}.'",
".",
"format",
"(",
"subscription",
".",
"name",
",",
"topic",
".",
"name",
")"
] | deletes an existing pub/sub topic . | train | false |
47,721 | def get_temp_file_path(file_path):
return (file_path + '.tmp')
| [
"def",
"get_temp_file_path",
"(",
"file_path",
")",
":",
"return",
"(",
"file_path",
"+",
"'.tmp'",
")"
] | generates a temporary filename . | train | false |
47,722 | def cross_from_below(x, threshold):
x = np.asarray(x)
threshold = threshold
ind = np.nonzero(((x[:(-1)] < threshold) & (x[1:] >= threshold)))[0]
if len(ind):
return (ind + 1)
else:
return ind
| [
"def",
"cross_from_below",
"(",
"x",
",",
"threshold",
")",
":",
"x",
"=",
"np",
".",
"asarray",
"(",
"x",
")",
"threshold",
"=",
"threshold",
"ind",
"=",
"np",
".",
"nonzero",
"(",
"(",
"(",
"x",
"[",
":",
"(",
"-",
"1",
")",
"]",
"<",
"threshold",
")",
"&",
"(",
"x",
"[",
"1",
":",
"]",
">=",
"threshold",
")",
")",
")",
"[",
"0",
"]",
"if",
"len",
"(",
"ind",
")",
":",
"return",
"(",
"ind",
"+",
"1",
")",
"else",
":",
"return",
"ind"
] | return the indices into *x* where *x* crosses some threshold from below . | train | false |
47,723 | def correct_barcode_bitwise(query_seq, seq_possibilities, nt_to_bits=DEFAULT_GOLAY_NT_TO_BITS):
if (nt_to_bits is None):
nt_to_bits = DEFAULT_NT_TO_BITS
dists = []
query_seq_bits = seq_to_bits(query_seq, nt_to_bits)
for seq in seq_possibilities:
possible_seq_bits = seq_to_bits(seq, nt_to_bits)
dists.append(hamming_dist(query_seq_bits, possible_seq_bits))
min_dist = min(dists)
number_mins = dists.count(min_dist)
if (number_mins > 1):
return (None, min_dist)
else:
best_hit = seq_possibilities[dists.index(min_dist)]
return (best_hit, min_dist)
| [
"def",
"correct_barcode_bitwise",
"(",
"query_seq",
",",
"seq_possibilities",
",",
"nt_to_bits",
"=",
"DEFAULT_GOLAY_NT_TO_BITS",
")",
":",
"if",
"(",
"nt_to_bits",
"is",
"None",
")",
":",
"nt_to_bits",
"=",
"DEFAULT_NT_TO_BITS",
"dists",
"=",
"[",
"]",
"query_seq_bits",
"=",
"seq_to_bits",
"(",
"query_seq",
",",
"nt_to_bits",
")",
"for",
"seq",
"in",
"seq_possibilities",
":",
"possible_seq_bits",
"=",
"seq_to_bits",
"(",
"seq",
",",
"nt_to_bits",
")",
"dists",
".",
"append",
"(",
"hamming_dist",
"(",
"query_seq_bits",
",",
"possible_seq_bits",
")",
")",
"min_dist",
"=",
"min",
"(",
"dists",
")",
"number_mins",
"=",
"dists",
".",
"count",
"(",
"min_dist",
")",
"if",
"(",
"number_mins",
">",
"1",
")",
":",
"return",
"(",
"None",
",",
"min_dist",
")",
"else",
":",
"best_hit",
"=",
"seq_possibilities",
"[",
"dists",
".",
"index",
"(",
"min_dist",
")",
"]",
"return",
"(",
"best_hit",
",",
"min_dist",
")"
] | finds closest match to query_seq assumes: all sequences are same length no sequence appears twice in seq_possibilities returns * best_hit is closest sequence from seq_possibilities . | train | false |
47,724 | def test_saved_inner_graph():
x = tensor.tensor3()
recurrent = SimpleRecurrent(dim=3, activation=Tanh())
y = recurrent.apply(x)
application_call = get_application_call(y)
assert application_call.inner_inputs
assert application_call.inner_outputs
cg = ComputationGraph(application_call.inner_outputs)
assert (len(VariableFilter(applications=[recurrent.apply])(cg)) == 3)
assert is_same_graph(application_call.inner_outputs[0], recurrent.apply(iterate=False, *application_call.inner_inputs))
| [
"def",
"test_saved_inner_graph",
"(",
")",
":",
"x",
"=",
"tensor",
".",
"tensor3",
"(",
")",
"recurrent",
"=",
"SimpleRecurrent",
"(",
"dim",
"=",
"3",
",",
"activation",
"=",
"Tanh",
"(",
")",
")",
"y",
"=",
"recurrent",
".",
"apply",
"(",
"x",
")",
"application_call",
"=",
"get_application_call",
"(",
"y",
")",
"assert",
"application_call",
".",
"inner_inputs",
"assert",
"application_call",
".",
"inner_outputs",
"cg",
"=",
"ComputationGraph",
"(",
"application_call",
".",
"inner_outputs",
")",
"assert",
"(",
"len",
"(",
"VariableFilter",
"(",
"applications",
"=",
"[",
"recurrent",
".",
"apply",
"]",
")",
"(",
"cg",
")",
")",
"==",
"3",
")",
"assert",
"is_same_graph",
"(",
"application_call",
".",
"inner_outputs",
"[",
"0",
"]",
",",
"recurrent",
".",
"apply",
"(",
"iterate",
"=",
"False",
",",
"*",
"application_call",
".",
"inner_inputs",
")",
")"
] | make sure that the original inner graph is saved . | train | false |
47,725 | def test_basic(script, tmpdir):
expected = '--hash=sha256:2cf24dba5fb0a30e26e83b2ac5b9e29e1b161e5c1fa7425e73043362938b9824'
result = script.pip('hash', _hello_file(tmpdir))
assert (expected in str(result))
| [
"def",
"test_basic",
"(",
"script",
",",
"tmpdir",
")",
":",
"expected",
"=",
"'--hash=sha256:2cf24dba5fb0a30e26e83b2ac5b9e29e1b161e5c1fa7425e73043362938b9824'",
"result",
"=",
"script",
".",
"pip",
"(",
"'hash'",
",",
"_hello_file",
"(",
"tmpdir",
")",
")",
"assert",
"(",
"expected",
"in",
"str",
"(",
"result",
")",
")"
] | run pip hash through its default behavior . | train | false |
47,727 | def norm_rgb(r, g, b):
greatest = max([r, g, b])
if (greatest > 0):
r /= greatest
g /= greatest
b /= greatest
return (r, g, b)
| [
"def",
"norm_rgb",
"(",
"r",
",",
"g",
",",
"b",
")",
":",
"greatest",
"=",
"max",
"(",
"[",
"r",
",",
"g",
",",
"b",
"]",
")",
"if",
"(",
"greatest",
">",
"0",
")",
":",
"r",
"/=",
"greatest",
"g",
"/=",
"greatest",
"b",
"/=",
"greatest",
"return",
"(",
"r",
",",
"g",
",",
"b",
")"
] | normalise rgb components so the most intense has a value of 1 . | train | false |
47,728 | def _set_thread_safe_app():
if (_local is not None):
WSGIApplication.app = WSGIApplication.active_instance = _local('app')
WSGIApplication.request = _local('request')
| [
"def",
"_set_thread_safe_app",
"(",
")",
":",
"if",
"(",
"_local",
"is",
"not",
"None",
")",
":",
"WSGIApplication",
".",
"app",
"=",
"WSGIApplication",
".",
"active_instance",
"=",
"_local",
"(",
"'app'",
")",
"WSGIApplication",
".",
"request",
"=",
"_local",
"(",
"'request'",
")"
] | assigns wsgiapplication globals to a proxy pointing to thread-local . | train | false |
47,730 | def is_id_list(lst):
return all(map(is_gm_id, lst))
| [
"def",
"is_id_list",
"(",
"lst",
")",
":",
"return",
"all",
"(",
"map",
"(",
"is_gm_id",
",",
"lst",
")",
")"
] | returns true if the given list is made up of all strings in gm id form . | train | false |
47,731 | @cronjobs.register
def update_contributor_metrics(day=None):
if settings.STAGE:
return
update_support_forum_contributors_metric(day)
update_kb_contributors_metric(day)
update_aoa_contributors_metric(day)
| [
"@",
"cronjobs",
".",
"register",
"def",
"update_contributor_metrics",
"(",
"day",
"=",
"None",
")",
":",
"if",
"settings",
".",
"STAGE",
":",
"return",
"update_support_forum_contributors_metric",
"(",
"day",
")",
"update_kb_contributors_metric",
"(",
"day",
")",
"update_aoa_contributors_metric",
"(",
"day",
")"
] | calculate and save contributor metrics . | train | false |
47,732 | def overlay_image(img, canvas=None, left=0, top=0):
if (canvas is None):
canvas = QImage(img.size(), QImage.Format_RGB32)
canvas.fill(Qt.white)
(left, top) = (int(left), int(top))
imageops.overlay(img, canvas, left, top)
return canvas
| [
"def",
"overlay_image",
"(",
"img",
",",
"canvas",
"=",
"None",
",",
"left",
"=",
"0",
",",
"top",
"=",
"0",
")",
":",
"if",
"(",
"canvas",
"is",
"None",
")",
":",
"canvas",
"=",
"QImage",
"(",
"img",
".",
"size",
"(",
")",
",",
"QImage",
".",
"Format_RGB32",
")",
"canvas",
".",
"fill",
"(",
"Qt",
".",
"white",
")",
"(",
"left",
",",
"top",
")",
"=",
"(",
"int",
"(",
"left",
")",
",",
"int",
"(",
"top",
")",
")",
"imageops",
".",
"overlay",
"(",
"img",
",",
"canvas",
",",
"left",
",",
"top",
")",
"return",
"canvas"
] | overlay the img onto the canvas at the specified position . | train | false |
47,733 | def generate_id():
return get_id(uuid.uuid4())
| [
"def",
"generate_id",
"(",
")",
":",
"return",
"get_id",
"(",
"uuid",
".",
"uuid4",
"(",
")",
")"
] | generate a short . | train | false |
47,734 | def is_valid_imdb_title_id(value):
if (not isinstance(value, basestring)):
raise TypeError(u'is_valid_imdb_title_id expects a string but got {0}'.format(type(value)))
return (re.match(u'tt[\\d]{7}', value) is not None)
| [
"def",
"is_valid_imdb_title_id",
"(",
"value",
")",
":",
"if",
"(",
"not",
"isinstance",
"(",
"value",
",",
"basestring",
")",
")",
":",
"raise",
"TypeError",
"(",
"u'is_valid_imdb_title_id expects a string but got {0}'",
".",
"format",
"(",
"type",
"(",
"value",
")",
")",
")",
"return",
"(",
"re",
".",
"match",
"(",
"u'tt[\\\\d]{7}'",
",",
"value",
")",
"is",
"not",
"None",
")"
] | return true if value is a valid imdb id for titles . | train | false |
47,735 | def _set_binops_check_loose(self, obj):
return (isinstance(obj, (_set_binop_bases + (self.__class__,))) or (util.duck_type_collection(obj) == set))
| [
"def",
"_set_binops_check_loose",
"(",
"self",
",",
"obj",
")",
":",
"return",
"(",
"isinstance",
"(",
"obj",
",",
"(",
"_set_binop_bases",
"+",
"(",
"self",
".",
"__class__",
",",
")",
")",
")",
"or",
"(",
"util",
".",
"duck_type_collection",
"(",
"obj",
")",
"==",
"set",
")",
")"
] | allow anything set-like to participate in set binops . | train | false |
47,736 | def get_instance_availability_zone(context, instance):
host = instance.get('host')
if (not host):
az = instance.get('availability_zone')
return az
cache_key = _make_cache_key(host)
cache = _get_cache()
az = cache.get(cache_key)
az_inst = instance.get('availability_zone')
if ((az_inst is not None) and (az != az_inst)):
az = None
if (not az):
elevated = context.elevated()
az = get_host_availability_zone(elevated, host)
cache.set(cache_key, az)
return az
| [
"def",
"get_instance_availability_zone",
"(",
"context",
",",
"instance",
")",
":",
"host",
"=",
"instance",
".",
"get",
"(",
"'host'",
")",
"if",
"(",
"not",
"host",
")",
":",
"az",
"=",
"instance",
".",
"get",
"(",
"'availability_zone'",
")",
"return",
"az",
"cache_key",
"=",
"_make_cache_key",
"(",
"host",
")",
"cache",
"=",
"_get_cache",
"(",
")",
"az",
"=",
"cache",
".",
"get",
"(",
"cache_key",
")",
"az_inst",
"=",
"instance",
".",
"get",
"(",
"'availability_zone'",
")",
"if",
"(",
"(",
"az_inst",
"is",
"not",
"None",
")",
"and",
"(",
"az",
"!=",
"az_inst",
")",
")",
":",
"az",
"=",
"None",
"if",
"(",
"not",
"az",
")",
":",
"elevated",
"=",
"context",
".",
"elevated",
"(",
")",
"az",
"=",
"get_host_availability_zone",
"(",
"elevated",
",",
"host",
")",
"cache",
".",
"set",
"(",
"cache_key",
",",
"az",
")",
"return",
"az"
] | return availability zone of specified instance . | train | false |
47,737 | def BlankLine():
return Leaf(token.NEWLINE, u'')
| [
"def",
"BlankLine",
"(",
")",
":",
"return",
"Leaf",
"(",
"token",
".",
"NEWLINE",
",",
"u''",
")"
] | a blank line . | train | false |
47,738 | @set_database
def get_assessment_item_data(assessment_item_id=None, **kwargs):
try:
assessment_item = AssessmentItem.get((AssessmentItem.id == assessment_item_id))
return model_to_dict(assessment_item)
except OperationalError:
return {}
| [
"@",
"set_database",
"def",
"get_assessment_item_data",
"(",
"assessment_item_id",
"=",
"None",
",",
"**",
"kwargs",
")",
":",
"try",
":",
"assessment_item",
"=",
"AssessmentItem",
".",
"get",
"(",
"(",
"AssessmentItem",
".",
"id",
"==",
"assessment_item_id",
")",
")",
"return",
"model_to_dict",
"(",
"assessment_item",
")",
"except",
"OperationalError",
":",
"return",
"{",
"}"
] | wrapper function to return assessment_item from database as a dictionary . | train | false |
47,740 | def _is_32bit():
return ((struct.calcsize('P') * 8) == 32)
| [
"def",
"_is_32bit",
"(",
")",
":",
"return",
"(",
"(",
"struct",
".",
"calcsize",
"(",
"'P'",
")",
"*",
"8",
")",
"==",
"32",
")"
] | detect if process is 32bit python . | train | false |
47,741 | def _write_messages(message_descriptors, out):
for message in (message_descriptors or []):
(out << '')
(out << '')
(out << ('class %s(messages.Message):' % message.name))
with out.indent():
if (not (message.enum_types or message.message_types or message.fields)):
(out << '')
(out << 'pass')
else:
_write_enums(message.enum_types, out)
_write_messages(message.message_types, out)
_write_fields(message.fields, out)
| [
"def",
"_write_messages",
"(",
"message_descriptors",
",",
"out",
")",
":",
"for",
"message",
"in",
"(",
"message_descriptors",
"or",
"[",
"]",
")",
":",
"(",
"out",
"<<",
"''",
")",
"(",
"out",
"<<",
"''",
")",
"(",
"out",
"<<",
"(",
"'class %s(messages.Message):'",
"%",
"message",
".",
"name",
")",
")",
"with",
"out",
".",
"indent",
"(",
")",
":",
"if",
"(",
"not",
"(",
"message",
".",
"enum_types",
"or",
"message",
".",
"message_types",
"or",
"message",
".",
"fields",
")",
")",
":",
"(",
"out",
"<<",
"''",
")",
"(",
"out",
"<<",
"'pass'",
")",
"else",
":",
"_write_enums",
"(",
"message",
".",
"enum_types",
",",
"out",
")",
"_write_messages",
"(",
"message",
".",
"message_types",
",",
"out",
")",
"_write_fields",
"(",
"message",
".",
"fields",
",",
"out",
")"
] | write nested and non-nested message types . | train | false |
47,742 | def dirname(p):
return split(p)[0]
| [
"def",
"dirname",
"(",
"p",
")",
":",
"return",
"split",
"(",
"p",
")",
"[",
"0",
"]"
] | return the head part of a path . | train | false |
47,743 | @pytest.fixture
def header_checker(caplog, stubs):
return HeaderChecker(caplog, stubs)
| [
"@",
"pytest",
".",
"fixture",
"def",
"header_checker",
"(",
"caplog",
",",
"stubs",
")",
":",
"return",
"HeaderChecker",
"(",
"caplog",
",",
"stubs",
")"
] | fixture that provides a headerchecker class for tests . | train | false |
47,744 | def _lookup_by_id_or_name_factory(iterator, element_name, doc):
def lookup_by_id_or_name(self, ref, before=None):
u'\n Given an key *ref*, finds the first element in the iterator\n with the attribute ID == *ref* or name == *ref*. If *before*\n is provided, will stop searching at the object *before*. This\n is important, since "forward references" are not allowed in\n the VOTABLE format.\n '
for element in getattr(self, iterator)():
if (element is before):
if (ref in (element.ID, element.name)):
vo_raise(u'{} references itself'.format(element_name), element._config, element._pos, KeyError)
break
if (ref in (element.ID, element.name)):
return element
raise KeyError(u"No {} with ID or name '{}' found before the referencing {}".format(element_name, ref, element_name))
lookup_by_id_or_name.__doc__ = doc
return lookup_by_id_or_name
| [
"def",
"_lookup_by_id_or_name_factory",
"(",
"iterator",
",",
"element_name",
",",
"doc",
")",
":",
"def",
"lookup_by_id_or_name",
"(",
"self",
",",
"ref",
",",
"before",
"=",
"None",
")",
":",
"for",
"element",
"in",
"getattr",
"(",
"self",
",",
"iterator",
")",
"(",
")",
":",
"if",
"(",
"element",
"is",
"before",
")",
":",
"if",
"(",
"ref",
"in",
"(",
"element",
".",
"ID",
",",
"element",
".",
"name",
")",
")",
":",
"vo_raise",
"(",
"u'{} references itself'",
".",
"format",
"(",
"element_name",
")",
",",
"element",
".",
"_config",
",",
"element",
".",
"_pos",
",",
"KeyError",
")",
"break",
"if",
"(",
"ref",
"in",
"(",
"element",
".",
"ID",
",",
"element",
".",
"name",
")",
")",
":",
"return",
"element",
"raise",
"KeyError",
"(",
"u\"No {} with ID or name '{}' found before the referencing {}\"",
".",
"format",
"(",
"element_name",
",",
"ref",
",",
"element_name",
")",
")",
"lookup_by_id_or_name",
".",
"__doc__",
"=",
"doc",
"return",
"lookup_by_id_or_name"
] | like _lookup_by_attr_factory . | train | false |
47,745 | def p_return_stmt(p):
p[0] = ast.Return(p[2])
| [
"def",
"p_return_stmt",
"(",
"p",
")",
":",
"p",
"[",
"0",
"]",
"=",
"ast",
".",
"Return",
"(",
"p",
"[",
"2",
"]",
")"
] | return_stmt : return testlist . | train | false |
47,747 | def load_from_yahoo(indexes=None, stocks=None, start=None, end=None, adjusted=True):
data = _load_raw_yahoo_data(indexes, stocks, start, end)
if adjusted:
close_key = 'Adj Close'
else:
close_key = 'Close'
df = pd.DataFrame({key: d[close_key] for (key, d) in iteritems(data)})
df.index = df.index.tz_localize(pytz.utc)
return df
| [
"def",
"load_from_yahoo",
"(",
"indexes",
"=",
"None",
",",
"stocks",
"=",
"None",
",",
"start",
"=",
"None",
",",
"end",
"=",
"None",
",",
"adjusted",
"=",
"True",
")",
":",
"data",
"=",
"_load_raw_yahoo_data",
"(",
"indexes",
",",
"stocks",
",",
"start",
",",
"end",
")",
"if",
"adjusted",
":",
"close_key",
"=",
"'Adj Close'",
"else",
":",
"close_key",
"=",
"'Close'",
"df",
"=",
"pd",
".",
"DataFrame",
"(",
"{",
"key",
":",
"d",
"[",
"close_key",
"]",
"for",
"(",
"key",
",",
"d",
")",
"in",
"iteritems",
"(",
"data",
")",
"}",
")",
"df",
".",
"index",
"=",
"df",
".",
"index",
".",
"tz_localize",
"(",
"pytz",
".",
"utc",
")",
"return",
"df"
] | loads price data from yahoo into a dataframe for each of the indicated assets . | train | false |
47,748 | def key_pair_get(context, user_id, name):
return IMPL.key_pair_get(context, user_id, name)
| [
"def",
"key_pair_get",
"(",
"context",
",",
"user_id",
",",
"name",
")",
":",
"return",
"IMPL",
".",
"key_pair_get",
"(",
"context",
",",
"user_id",
",",
"name",
")"
] | get a key_pair or raise if it does not exist . | train | false |
47,749 | def get_install_key():
return sha1(settings.SECRET_KEY).hexdigest()
| [
"def",
"get_install_key",
"(",
")",
":",
"return",
"sha1",
"(",
"settings",
".",
"SECRET_KEY",
")",
".",
"hexdigest",
"(",
")"
] | return the installation key for this server . | train | false |
47,750 | def istask(x):
return ((type(x) is tuple) and x and callable(x[0]))
| [
"def",
"istask",
"(",
"x",
")",
":",
"return",
"(",
"(",
"type",
"(",
"x",
")",
"is",
"tuple",
")",
"and",
"x",
"and",
"callable",
"(",
"x",
"[",
"0",
"]",
")",
")"
] | is x a runnable task? a task is a tuple with a callable first argument examples . | train | false |
47,752 | def add_tooltips_columns(renderer, tooltips, group):
current_columns = renderer.data_source.data.keys()
if isinstance(tooltips[0], tuple):
tooltips_columns = [pair[1].replace('@', '') for pair in tooltips]
elif isinstance(tooltips[0], str):
tooltips_columns = tooltips
else:
tooltips_columns = []
for column in tooltips_columns:
if (column in current_columns):
continue
elif ('$' in column):
continue
renderer.data_source.add(group.get_values(column), column)
return renderer
| [
"def",
"add_tooltips_columns",
"(",
"renderer",
",",
"tooltips",
",",
"group",
")",
":",
"current_columns",
"=",
"renderer",
".",
"data_source",
".",
"data",
".",
"keys",
"(",
")",
"if",
"isinstance",
"(",
"tooltips",
"[",
"0",
"]",
",",
"tuple",
")",
":",
"tooltips_columns",
"=",
"[",
"pair",
"[",
"1",
"]",
".",
"replace",
"(",
"'@'",
",",
"''",
")",
"for",
"pair",
"in",
"tooltips",
"]",
"elif",
"isinstance",
"(",
"tooltips",
"[",
"0",
"]",
",",
"str",
")",
":",
"tooltips_columns",
"=",
"tooltips",
"else",
":",
"tooltips_columns",
"=",
"[",
"]",
"for",
"column",
"in",
"tooltips_columns",
":",
"if",
"(",
"column",
"in",
"current_columns",
")",
":",
"continue",
"elif",
"(",
"'$'",
"in",
"column",
")",
":",
"continue",
"renderer",
".",
"data_source",
".",
"add",
"(",
"group",
".",
"get_values",
"(",
"column",
")",
",",
"column",
")",
"return",
"renderer"
] | args: renderer : renderer for the glyph to be modified . | train | false |
47,755 | def getGeometryToolsPath(subName=''):
return getJoinedPath(getGeometryPath('geometry_tools'), subName)
| [
"def",
"getGeometryToolsPath",
"(",
"subName",
"=",
"''",
")",
":",
"return",
"getJoinedPath",
"(",
"getGeometryPath",
"(",
"'geometry_tools'",
")",
",",
"subName",
")"
] | get the geometry tools directory path . | train | false |
47,756 | def update_param(prefixed_name, input_values, new_value):
for key in input_values:
match = re.match((('^' + key) + '_(\\d+)\\|(.+)'), prefixed_name)
if (match and (not key.endswith('|__identifier__'))):
index = int(match.group(1))
if (isinstance(input_values[key], list) and (len(input_values[key]) > index)):
update_param(match.group(2), input_values[key][index], new_value)
else:
match = re.match((('^' + key) + '\\|(.+)'), prefixed_name)
if (isinstance(input_values[key], dict) and match):
update_param(match.group(1), input_values[key], new_value)
elif (prefixed_name == key):
input_values[key] = new_value
| [
"def",
"update_param",
"(",
"prefixed_name",
",",
"input_values",
",",
"new_value",
")",
":",
"for",
"key",
"in",
"input_values",
":",
"match",
"=",
"re",
".",
"match",
"(",
"(",
"(",
"'^'",
"+",
"key",
")",
"+",
"'_(\\\\d+)\\\\|(.+)'",
")",
",",
"prefixed_name",
")",
"if",
"(",
"match",
"and",
"(",
"not",
"key",
".",
"endswith",
"(",
"'|__identifier__'",
")",
")",
")",
":",
"index",
"=",
"int",
"(",
"match",
".",
"group",
"(",
"1",
")",
")",
"if",
"(",
"isinstance",
"(",
"input_values",
"[",
"key",
"]",
",",
"list",
")",
"and",
"(",
"len",
"(",
"input_values",
"[",
"key",
"]",
")",
">",
"index",
")",
")",
":",
"update_param",
"(",
"match",
".",
"group",
"(",
"2",
")",
",",
"input_values",
"[",
"key",
"]",
"[",
"index",
"]",
",",
"new_value",
")",
"else",
":",
"match",
"=",
"re",
".",
"match",
"(",
"(",
"(",
"'^'",
"+",
"key",
")",
"+",
"'\\\\|(.+)'",
")",
",",
"prefixed_name",
")",
"if",
"(",
"isinstance",
"(",
"input_values",
"[",
"key",
"]",
",",
"dict",
")",
"and",
"match",
")",
":",
"update_param",
"(",
"match",
".",
"group",
"(",
"1",
")",
",",
"input_values",
"[",
"key",
"]",
",",
"new_value",
")",
"elif",
"(",
"prefixed_name",
"==",
"key",
")",
":",
"input_values",
"[",
"key",
"]",
"=",
"new_value"
] | given a prefixed parameter name . | train | false |
47,758 | def client_list_entries_multi_project(client, to_delete):
PROJECT_IDS = ['one-project', 'another-project']
for entry in client.list_entries(project_ids=PROJECT_IDS):
do_something_with(entry)
| [
"def",
"client_list_entries_multi_project",
"(",
"client",
",",
"to_delete",
")",
":",
"PROJECT_IDS",
"=",
"[",
"'one-project'",
",",
"'another-project'",
"]",
"for",
"entry",
"in",
"client",
".",
"list_entries",
"(",
"project_ids",
"=",
"PROJECT_IDS",
")",
":",
"do_something_with",
"(",
"entry",
")"
] | list entries via client across multiple projects . | train | true |
47,759 | def get_project_description(project):
return _(u'{0} is translated into {1} languages using Weblate. Join the translation or start translating your own project.').format(project, project.get_language_count())
| [
"def",
"get_project_description",
"(",
"project",
")",
":",
"return",
"_",
"(",
"u'{0} is translated into {1} languages using Weblate. Join the translation or start translating your own project.'",
")",
".",
"format",
"(",
"project",
",",
"project",
".",
"get_language_count",
"(",
")",
")"
] | returns verbose description for project translation . | train | false |
47,760 | def crc_check(path, target_crc):
try:
fp = open(path, 'rb')
except:
return False
crc = binascii.crc32('')
while 1:
data = fp.read(4096)
if (not data):
break
crc = binascii.crc32(data, crc)
fp.close()
crc = ('%08x' % ((crc & 4294967295),))
return (crc.lower() == target_crc.lower())
| [
"def",
"crc_check",
"(",
"path",
",",
"target_crc",
")",
":",
"try",
":",
"fp",
"=",
"open",
"(",
"path",
",",
"'rb'",
")",
"except",
":",
"return",
"False",
"crc",
"=",
"binascii",
".",
"crc32",
"(",
"''",
")",
"while",
"1",
":",
"data",
"=",
"fp",
".",
"read",
"(",
"4096",
")",
"if",
"(",
"not",
"data",
")",
":",
"break",
"crc",
"=",
"binascii",
".",
"crc32",
"(",
"data",
",",
"crc",
")",
"fp",
".",
"close",
"(",
")",
"crc",
"=",
"(",
"'%08x'",
"%",
"(",
"(",
"crc",
"&",
"4294967295",
")",
",",
")",
")",
"return",
"(",
"crc",
".",
"lower",
"(",
")",
"==",
"target_crc",
".",
"lower",
"(",
")",
")"
] | return true if file matches crc . | train | false |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.