id_within_dataset int64 1 55.5k | snippet stringlengths 19 14.2k | tokens listlengths 6 1.63k | nl stringlengths 6 352 | split_within_dataset stringclasses 1 value | is_duplicated bool 2 classes |
|---|---|---|---|---|---|
17,517 | def load_pem(contents, pem_marker):
(pem_start, pem_end) = _markers(pem_marker)
pem_lines = []
in_pem_part = False
for line in contents.splitlines():
line = line.strip()
if (not line):
continue
if (line == pem_start):
if in_pem_part:
raise ValueError(('Seen start marker "%s" twice' % pem_start))
in_pem_part = True
continue
if (not in_pem_part):
continue
if (in_pem_part and (line == pem_end)):
in_pem_part = False
break
if (b(':') in line):
continue
pem_lines.append(line)
if (not pem_lines):
raise ValueError(('No PEM start marker "%s" found' % pem_start))
if in_pem_part:
raise ValueError(('No PEM end marker "%s" found' % pem_end))
pem = b('').join(pem_lines)
return base64.decodestring(pem)
| [
"def",
"load_pem",
"(",
"contents",
",",
"pem_marker",
")",
":",
"(",
"pem_start",
",",
"pem_end",
")",
"=",
"_markers",
"(",
"pem_marker",
")",
"pem_lines",
"=",
"[",
"]",
"in_pem_part",
"=",
"False",
"for",
"line",
"in",
"contents",
".",
"splitlines",
"(",
")",
":",
"line",
"=",
"line",
".",
"strip",
"(",
")",
"if",
"(",
"not",
"line",
")",
":",
"continue",
"if",
"(",
"line",
"==",
"pem_start",
")",
":",
"if",
"in_pem_part",
":",
"raise",
"ValueError",
"(",
"(",
"'Seen start marker \"%s\" twice'",
"%",
"pem_start",
")",
")",
"in_pem_part",
"=",
"True",
"continue",
"if",
"(",
"not",
"in_pem_part",
")",
":",
"continue",
"if",
"(",
"in_pem_part",
"and",
"(",
"line",
"==",
"pem_end",
")",
")",
":",
"in_pem_part",
"=",
"False",
"break",
"if",
"(",
"b",
"(",
"':'",
")",
"in",
"line",
")",
":",
"continue",
"pem_lines",
".",
"append",
"(",
"line",
")",
"if",
"(",
"not",
"pem_lines",
")",
":",
"raise",
"ValueError",
"(",
"(",
"'No PEM start marker \"%s\" found'",
"%",
"pem_start",
")",
")",
"if",
"in_pem_part",
":",
"raise",
"ValueError",
"(",
"(",
"'No PEM end marker \"%s\" found'",
"%",
"pem_end",
")",
")",
"pem",
"=",
"b",
"(",
"''",
")",
".",
"join",
"(",
"pem_lines",
")",
"return",
"base64",
".",
"decodestring",
"(",
"pem",
")"
] | loads a pem file . | train | false |
17,518 | def qual_missing(id_, seq, qual):
return (qual is None)
| [
"def",
"qual_missing",
"(",
"id_",
",",
"seq",
",",
"qual",
")",
":",
"return",
"(",
"qual",
"is",
"None",
")"
] | returns true if qual is none . | train | false |
17,519 | def getFloatDefaultByDictionary(defaultFloat, dictionary, key):
evaluatedFloat = None
if (key in dictionary):
evaluatedFloat = getFloatFromValue(dictionary[key])
if (evaluatedFloat == None):
return defaultFloat
return evaluatedFloat
| [
"def",
"getFloatDefaultByDictionary",
"(",
"defaultFloat",
",",
"dictionary",
",",
"key",
")",
":",
"evaluatedFloat",
"=",
"None",
"if",
"(",
"key",
"in",
"dictionary",
")",
":",
"evaluatedFloat",
"=",
"getFloatFromValue",
"(",
"dictionary",
"[",
"key",
"]",
")",
"if",
"(",
"evaluatedFloat",
"==",
"None",
")",
":",
"return",
"defaultFloat",
"return",
"evaluatedFloat"
] | get the value as a float . | train | false |
17,520 | def teardown_test_db():
from sickbeard.db import db_cons
for connection in db_cons:
db_cons[connection].commit()
| [
"def",
"teardown_test_db",
"(",
")",
":",
"from",
"sickbeard",
".",
"db",
"import",
"db_cons",
"for",
"connection",
"in",
"db_cons",
":",
"db_cons",
"[",
"connection",
"]",
".",
"commit",
"(",
")"
] | tear down the test database . | train | false |
17,522 | def nonzero(a):
return a.nonzero()
| [
"def",
"nonzero",
"(",
"a",
")",
":",
"return",
"a",
".",
"nonzero",
"(",
")"
] | return the indices of the elements that are non-zero . | train | false |
17,525 | def check_stoplimit_prices(price, label):
try:
if (not isfinite(price)):
raise BadOrderParameters(msg='Attempted to place an order with a {} price of {}.'.format(label, price))
except TypeError:
raise BadOrderParameters(msg='Attempted to place an order with a {} price of {}.'.format(label, type(price)))
if (price < 0):
raise BadOrderParameters(msg="Can't place a {} order with a negative price.".format(label))
| [
"def",
"check_stoplimit_prices",
"(",
"price",
",",
"label",
")",
":",
"try",
":",
"if",
"(",
"not",
"isfinite",
"(",
"price",
")",
")",
":",
"raise",
"BadOrderParameters",
"(",
"msg",
"=",
"'Attempted to place an order with a {} price of {}.'",
".",
"format",
"(",
"label",
",",
"price",
")",
")",
"except",
"TypeError",
":",
"raise",
"BadOrderParameters",
"(",
"msg",
"=",
"'Attempted to place an order with a {} price of {}.'",
".",
"format",
"(",
"label",
",",
"type",
"(",
"price",
")",
")",
")",
"if",
"(",
"price",
"<",
"0",
")",
":",
"raise",
"BadOrderParameters",
"(",
"msg",
"=",
"\"Can't place a {} order with a negative price.\"",
".",
"format",
"(",
"label",
")",
")"
] | check to make sure the stop/limit prices are reasonable and raise a badorderparameters exception if not . | train | true |
17,526 | def do_mark_unsafe(value):
return text_type(value)
| [
"def",
"do_mark_unsafe",
"(",
"value",
")",
":",
"return",
"text_type",
"(",
"value",
")"
] | mark a value as unsafe . | train | false |
17,527 | def add_logs(logx, logy):
if (logx < (logy + _ADD_LOGS_MAX_DIFF)):
return logy
if (logy < (logx + _ADD_LOGS_MAX_DIFF)):
return logx
base = min(logx, logy)
return (base + math.log(((2 ** (logx - base)) + (2 ** (logy - base))), 2))
| [
"def",
"add_logs",
"(",
"logx",
",",
"logy",
")",
":",
"if",
"(",
"logx",
"<",
"(",
"logy",
"+",
"_ADD_LOGS_MAX_DIFF",
")",
")",
":",
"return",
"logy",
"if",
"(",
"logy",
"<",
"(",
"logx",
"+",
"_ADD_LOGS_MAX_DIFF",
")",
")",
":",
"return",
"logx",
"base",
"=",
"min",
"(",
"logx",
",",
"logy",
")",
"return",
"(",
"base",
"+",
"math",
".",
"log",
"(",
"(",
"(",
"2",
"**",
"(",
"logx",
"-",
"base",
")",
")",
"+",
"(",
"2",
"**",
"(",
"logy",
"-",
"base",
")",
")",
")",
",",
"2",
")",
")"
] | given two numbers logx = *log(x)* and logy = *log(y)* . | train | false |
17,528 | def check_file_size(f, max_allowed_size):
if (f.size > max_allowed_size):
message = (_lazy(u'"%s" is too large (%sKB), the limit is %sKB') % (f.name, (f.size >> 10), (max_allowed_size >> 10)))
raise FileTooLargeError(message)
| [
"def",
"check_file_size",
"(",
"f",
",",
"max_allowed_size",
")",
":",
"if",
"(",
"f",
".",
"size",
">",
"max_allowed_size",
")",
":",
"message",
"=",
"(",
"_lazy",
"(",
"u'\"%s\" is too large (%sKB), the limit is %sKB'",
")",
"%",
"(",
"f",
".",
"name",
",",
"(",
"f",
".",
"size",
">>",
"10",
")",
",",
"(",
"max_allowed_size",
">>",
"10",
")",
")",
")",
"raise",
"FileTooLargeError",
"(",
"message",
")"
] | check the file size of f is less than max_allowed_size raise filetoolargeerror if the check fails . | train | false |
17,529 | def FlagCxx14Features(filename, clean_lines, linenum, error):
line = clean_lines.elided[linenum]
include = Match('\\s*#\\s*include\\s+[<"]([^<"]+)[">]', line)
if (include and (include.group(1) in ('scoped_allocator', 'shared_mutex'))):
error(filename, linenum, 'build/c++14', 5, ('<%s> is an unapproved C++14 header.' % include.group(1)))
| [
"def",
"FlagCxx14Features",
"(",
"filename",
",",
"clean_lines",
",",
"linenum",
",",
"error",
")",
":",
"line",
"=",
"clean_lines",
".",
"elided",
"[",
"linenum",
"]",
"include",
"=",
"Match",
"(",
"'\\\\s*#\\\\s*include\\\\s+[<\"]([^<\"]+)[\">]'",
",",
"line",
")",
"if",
"(",
"include",
"and",
"(",
"include",
".",
"group",
"(",
"1",
")",
"in",
"(",
"'scoped_allocator'",
",",
"'shared_mutex'",
")",
")",
")",
":",
"error",
"(",
"filename",
",",
"linenum",
",",
"'build/c++14'",
",",
"5",
",",
"(",
"'<%s> is an unapproved C++14 header.'",
"%",
"include",
".",
"group",
"(",
"1",
")",
")",
")"
] | flag those c++14 features that we restrict . | train | true |
17,530 | def entry_list_list(options):
with Session() as session:
try:
entry_list = get_list_by_exact_name(options.list_name, session=session)
except NoResultFound:
console(u'Could not find entry list with name {}'.format(options.list_name))
return
header = [u'#', u'Title', u'# of fields']
table_data = [header]
for entry in get_entries_by_list_id(entry_list.id, order_by=u'added', descending=True, session=session):
table_data.append([entry.id, entry.title, len(entry.entry)])
try:
table = TerminalTable(options.table_type, table_data)
console(table.output)
except TerminalTableError as e:
console((u'ERROR: %s' % str(e)))
| [
"def",
"entry_list_list",
"(",
"options",
")",
":",
"with",
"Session",
"(",
")",
"as",
"session",
":",
"try",
":",
"entry_list",
"=",
"get_list_by_exact_name",
"(",
"options",
".",
"list_name",
",",
"session",
"=",
"session",
")",
"except",
"NoResultFound",
":",
"console",
"(",
"u'Could not find entry list with name {}'",
".",
"format",
"(",
"options",
".",
"list_name",
")",
")",
"return",
"header",
"=",
"[",
"u'#'",
",",
"u'Title'",
",",
"u'# of fields'",
"]",
"table_data",
"=",
"[",
"header",
"]",
"for",
"entry",
"in",
"get_entries_by_list_id",
"(",
"entry_list",
".",
"id",
",",
"order_by",
"=",
"u'added'",
",",
"descending",
"=",
"True",
",",
"session",
"=",
"session",
")",
":",
"table_data",
".",
"append",
"(",
"[",
"entry",
".",
"id",
",",
"entry",
".",
"title",
",",
"len",
"(",
"entry",
".",
"entry",
")",
"]",
")",
"try",
":",
"table",
"=",
"TerminalTable",
"(",
"options",
".",
"table_type",
",",
"table_data",
")",
"console",
"(",
"table",
".",
"output",
")",
"except",
"TerminalTableError",
"as",
"e",
":",
"console",
"(",
"(",
"u'ERROR: %s'",
"%",
"str",
"(",
"e",
")",
")",
")"
] | list entry list . | train | false |
17,531 | def toResponse(stanza, stanzaType=None):
toAddr = stanza.getAttribute('from')
fromAddr = stanza.getAttribute('to')
stanzaID = stanza.getAttribute('id')
response = domish.Element((None, stanza.name))
if toAddr:
response['to'] = toAddr
if fromAddr:
response['from'] = fromAddr
if stanzaID:
response['id'] = stanzaID
if stanzaType:
response['type'] = stanzaType
return response
| [
"def",
"toResponse",
"(",
"stanza",
",",
"stanzaType",
"=",
"None",
")",
":",
"toAddr",
"=",
"stanza",
".",
"getAttribute",
"(",
"'from'",
")",
"fromAddr",
"=",
"stanza",
".",
"getAttribute",
"(",
"'to'",
")",
"stanzaID",
"=",
"stanza",
".",
"getAttribute",
"(",
"'id'",
")",
"response",
"=",
"domish",
".",
"Element",
"(",
"(",
"None",
",",
"stanza",
".",
"name",
")",
")",
"if",
"toAddr",
":",
"response",
"[",
"'to'",
"]",
"=",
"toAddr",
"if",
"fromAddr",
":",
"response",
"[",
"'from'",
"]",
"=",
"fromAddr",
"if",
"stanzaID",
":",
"response",
"[",
"'id'",
"]",
"=",
"stanzaID",
"if",
"stanzaType",
":",
"response",
"[",
"'type'",
"]",
"=",
"stanzaType",
"return",
"response"
] | create a response stanza from another stanza . | train | false |
17,532 | def _make_subclasshook(cls):
if isinstance(cls.__extra__, abc.ABCMeta):
def __extrahook__(cls, subclass):
_valid_for_check(cls)
res = cls.__extra__.__subclasshook__(subclass)
if (res is not NotImplemented):
return res
if (cls.__extra__ in getattr(subclass, u'__mro__', ())):
return True
for scls in cls.__extra__.__subclasses__():
if isinstance(scls, GenericMeta):
continue
if issubclass(subclass, scls):
return True
return NotImplemented
else:
def __extrahook__(cls, subclass):
_valid_for_check(cls)
if (cls.__extra__ and issubclass(subclass, cls.__extra__)):
return True
return NotImplemented
return classmethod(__extrahook__)
| [
"def",
"_make_subclasshook",
"(",
"cls",
")",
":",
"if",
"isinstance",
"(",
"cls",
".",
"__extra__",
",",
"abc",
".",
"ABCMeta",
")",
":",
"def",
"__extrahook__",
"(",
"cls",
",",
"subclass",
")",
":",
"_valid_for_check",
"(",
"cls",
")",
"res",
"=",
"cls",
".",
"__extra__",
".",
"__subclasshook__",
"(",
"subclass",
")",
"if",
"(",
"res",
"is",
"not",
"NotImplemented",
")",
":",
"return",
"res",
"if",
"(",
"cls",
".",
"__extra__",
"in",
"getattr",
"(",
"subclass",
",",
"u'__mro__'",
",",
"(",
")",
")",
")",
":",
"return",
"True",
"for",
"scls",
"in",
"cls",
".",
"__extra__",
".",
"__subclasses__",
"(",
")",
":",
"if",
"isinstance",
"(",
"scls",
",",
"GenericMeta",
")",
":",
"continue",
"if",
"issubclass",
"(",
"subclass",
",",
"scls",
")",
":",
"return",
"True",
"return",
"NotImplemented",
"else",
":",
"def",
"__extrahook__",
"(",
"cls",
",",
"subclass",
")",
":",
"_valid_for_check",
"(",
"cls",
")",
"if",
"(",
"cls",
".",
"__extra__",
"and",
"issubclass",
"(",
"subclass",
",",
"cls",
".",
"__extra__",
")",
")",
":",
"return",
"True",
"return",
"NotImplemented",
"return",
"classmethod",
"(",
"__extrahook__",
")"
] | construct a __subclasshook__ callable that incorporates the associated __extra__ class in subclass checks performed against cls . | train | true |
17,533 | def test_spinner_non_unicode_console():
stream = FakeTTY(u'ascii')
chars = console.Spinner._default_unicode_chars
with console.Spinner(u'Reticulating splines', file=stream, chars=chars) as s:
next(s)
| [
"def",
"test_spinner_non_unicode_console",
"(",
")",
":",
"stream",
"=",
"FakeTTY",
"(",
"u'ascii'",
")",
"chars",
"=",
"console",
".",
"Spinner",
".",
"_default_unicode_chars",
"with",
"console",
".",
"Spinner",
"(",
"u'Reticulating splines'",
",",
"file",
"=",
"stream",
",",
"chars",
"=",
"chars",
")",
"as",
"s",
":",
"next",
"(",
"s",
")"
] | regression test for #1760 ensures that the spinner can fall go into fallback mode when using the unicode spinner on a terminal whose default encoding cannot encode the unicode characters . | train | false |
17,534 | def index_to_slice(idx):
return slice(idx, (idx + 1), None)
| [
"def",
"index_to_slice",
"(",
"idx",
")",
":",
"return",
"slice",
"(",
"idx",
",",
"(",
"idx",
"+",
"1",
")",
",",
"None",
")"
] | converts an index to a slice . | train | false |
17,535 | def contentfilter(fsname, pattern):
if (pattern is None):
return True
try:
f = open(fsname)
prog = re.compile(pattern)
for line in f:
if prog.match(line):
f.close()
return True
f.close()
except:
pass
return False
| [
"def",
"contentfilter",
"(",
"fsname",
",",
"pattern",
")",
":",
"if",
"(",
"pattern",
"is",
"None",
")",
":",
"return",
"True",
"try",
":",
"f",
"=",
"open",
"(",
"fsname",
")",
"prog",
"=",
"re",
".",
"compile",
"(",
"pattern",
")",
"for",
"line",
"in",
"f",
":",
"if",
"prog",
".",
"match",
"(",
"line",
")",
":",
"f",
".",
"close",
"(",
")",
"return",
"True",
"f",
".",
"close",
"(",
")",
"except",
":",
"pass",
"return",
"False"
] | filter files which contain the given expression . | train | false |
17,537 | def update_system(version='', ruby=None, runas=None, gem_bin=None):
return _gem(['update', '--system', version], ruby, gem_bin=gem_bin, runas=runas)
| [
"def",
"update_system",
"(",
"version",
"=",
"''",
",",
"ruby",
"=",
"None",
",",
"runas",
"=",
"None",
",",
"gem_bin",
"=",
"None",
")",
":",
"return",
"_gem",
"(",
"[",
"'update'",
",",
"'--system'",
",",
"version",
"]",
",",
"ruby",
",",
"gem_bin",
"=",
"gem_bin",
",",
"runas",
"=",
"runas",
")"
] | update rubygems . | train | true |
17,538 | def _coerceToFilesystemEncoding(path, newpath, encoding=None):
if (type(path) == bytes):
return _asFilesystemBytes(newpath, encoding=encoding)
else:
return _asFilesystemText(newpath, encoding=encoding)
| [
"def",
"_coerceToFilesystemEncoding",
"(",
"path",
",",
"newpath",
",",
"encoding",
"=",
"None",
")",
":",
"if",
"(",
"type",
"(",
"path",
")",
"==",
"bytes",
")",
":",
"return",
"_asFilesystemBytes",
"(",
"newpath",
",",
"encoding",
"=",
"encoding",
")",
"else",
":",
"return",
"_asFilesystemText",
"(",
"newpath",
",",
"encoding",
"=",
"encoding",
")"
] | return a c{newpath} that is suitable for joining to c{path} . | train | false |
17,539 | def _sanitize_index(data, index, copy=False):
if (index is None):
return data
if (len(data) != len(index)):
raise ValueError('Length of values does not match length of index')
if isinstance(data, PeriodIndex):
data = data.asobject
elif isinstance(data, DatetimeIndex):
data = data._to_embed(keep_tz=True)
if copy:
data = data.copy()
elif isinstance(data, np.ndarray):
if (data.dtype.kind in ['M', 'm']):
data = _sanitize_array(data, index, copy=copy)
return data
| [
"def",
"_sanitize_index",
"(",
"data",
",",
"index",
",",
"copy",
"=",
"False",
")",
":",
"if",
"(",
"index",
"is",
"None",
")",
":",
"return",
"data",
"if",
"(",
"len",
"(",
"data",
")",
"!=",
"len",
"(",
"index",
")",
")",
":",
"raise",
"ValueError",
"(",
"'Length of values does not match length of index'",
")",
"if",
"isinstance",
"(",
"data",
",",
"PeriodIndex",
")",
":",
"data",
"=",
"data",
".",
"asobject",
"elif",
"isinstance",
"(",
"data",
",",
"DatetimeIndex",
")",
":",
"data",
"=",
"data",
".",
"_to_embed",
"(",
"keep_tz",
"=",
"True",
")",
"if",
"copy",
":",
"data",
"=",
"data",
".",
"copy",
"(",
")",
"elif",
"isinstance",
"(",
"data",
",",
"np",
".",
"ndarray",
")",
":",
"if",
"(",
"data",
".",
"dtype",
".",
"kind",
"in",
"[",
"'M'",
",",
"'m'",
"]",
")",
":",
"data",
"=",
"_sanitize_array",
"(",
"data",
",",
"index",
",",
"copy",
"=",
"copy",
")",
"return",
"data"
] | sanitize an index type to return an ndarray of the underlying . | train | false |
17,540 | def addCylinderOutputByEndStart(endZ, inradiusComplex, outputs, sides, start, topOverBottom=1.0):
inradius = Vector3(inradiusComplex.real, inradiusComplex.imag, (0.5 * abs((endZ - start.z))))
cylinderOutput = getGeometryOutput(inradius, sides, topOverBottom)
vertexes = matrix.getVertexes(cylinderOutput)
if (endZ < start.z):
for vertex in vertexes:
vertex.z = (- vertex.z)
translation = Vector3(start.x, start.y, (inradius.z + min(start.z, endZ)))
euclidean.translateVector3Path(vertexes, translation)
outputs.append(cylinderOutput)
| [
"def",
"addCylinderOutputByEndStart",
"(",
"endZ",
",",
"inradiusComplex",
",",
"outputs",
",",
"sides",
",",
"start",
",",
"topOverBottom",
"=",
"1.0",
")",
":",
"inradius",
"=",
"Vector3",
"(",
"inradiusComplex",
".",
"real",
",",
"inradiusComplex",
".",
"imag",
",",
"(",
"0.5",
"*",
"abs",
"(",
"(",
"endZ",
"-",
"start",
".",
"z",
")",
")",
")",
")",
"cylinderOutput",
"=",
"getGeometryOutput",
"(",
"inradius",
",",
"sides",
",",
"topOverBottom",
")",
"vertexes",
"=",
"matrix",
".",
"getVertexes",
"(",
"cylinderOutput",
")",
"if",
"(",
"endZ",
"<",
"start",
".",
"z",
")",
":",
"for",
"vertex",
"in",
"vertexes",
":",
"vertex",
".",
"z",
"=",
"(",
"-",
"vertex",
".",
"z",
")",
"translation",
"=",
"Vector3",
"(",
"start",
".",
"x",
",",
"start",
".",
"y",
",",
"(",
"inradius",
".",
"z",
"+",
"min",
"(",
"start",
".",
"z",
",",
"endZ",
")",
")",
")",
"euclidean",
".",
"translateVector3Path",
"(",
"vertexes",
",",
"translation",
")",
"outputs",
".",
"append",
"(",
"cylinderOutput",
")"
] | add cylinder triangle mesh by endz . | train | false |
17,541 | def get_hub():
try:
import select
if hasattr(select, 'poll'):
return 'poll'
return 'selects'
except ImportError:
return None
| [
"def",
"get_hub",
"(",
")",
":",
"try",
":",
"import",
"select",
"if",
"hasattr",
"(",
"select",
",",
"'poll'",
")",
":",
"return",
"'poll'",
"return",
"'selects'",
"except",
"ImportError",
":",
"return",
"None"
] | return the hub for the current thread . | train | false |
17,542 | def test_empty_lists(Chart):
chart = Chart()
chart.add('A', [1, 2])
chart.add('B', [])
if (not chart._dual):
chart.x_labels = ('red', 'green', 'blue')
q = chart.render_pyquery()
assert (len(q('.legend')) == 2)
| [
"def",
"test_empty_lists",
"(",
"Chart",
")",
":",
"chart",
"=",
"Chart",
"(",
")",
"chart",
".",
"add",
"(",
"'A'",
",",
"[",
"1",
",",
"2",
"]",
")",
"chart",
".",
"add",
"(",
"'B'",
",",
"[",
"]",
")",
"if",
"(",
"not",
"chart",
".",
"_dual",
")",
":",
"chart",
".",
"x_labels",
"=",
"(",
"'red'",
",",
"'green'",
",",
"'blue'",
")",
"q",
"=",
"chart",
".",
"render_pyquery",
"(",
")",
"assert",
"(",
"len",
"(",
"q",
"(",
"'.legend'",
")",
")",
"==",
"2",
")"
] | test chart rendering with an empty serie . | train | false |
17,543 | def pci_device_update(context, node_id, address, value):
return IMPL.pci_device_update(context, node_id, address, value)
| [
"def",
"pci_device_update",
"(",
"context",
",",
"node_id",
",",
"address",
",",
"value",
")",
":",
"return",
"IMPL",
".",
"pci_device_update",
"(",
"context",
",",
"node_id",
",",
"address",
",",
"value",
")"
] | update a pci device . | train | false |
17,546 | def pull_raw(url, name, verify=False):
return _pull_image('raw', url, name, verify=verify)
| [
"def",
"pull_raw",
"(",
"url",
",",
"name",
",",
"verify",
"=",
"False",
")",
":",
"return",
"_pull_image",
"(",
"'raw'",
",",
"url",
",",
"name",
",",
"verify",
"=",
"verify",
")"
] | execute a machinectl pull-raw to download a . | train | true |
17,548 | def _item_to_job(iterator, resource):
return iterator.client.job_from_resource(resource)
| [
"def",
"_item_to_job",
"(",
"iterator",
",",
"resource",
")",
":",
"return",
"iterator",
".",
"client",
".",
"job_from_resource",
"(",
"resource",
")"
] | convert a json job to the native object . | train | false |
17,549 | def a1_to_rowcol(label):
m = CELL_ADDR_RE.match(label)
if m:
column_label = m.group(1).upper()
row = int(m.group(2))
col = 0
for (i, c) in enumerate(reversed(column_label)):
col += ((ord(c) - MAGIC_NUMBER) * (26 ** i))
else:
raise IncorrectCellLabel(label)
return (row, col)
| [
"def",
"a1_to_rowcol",
"(",
"label",
")",
":",
"m",
"=",
"CELL_ADDR_RE",
".",
"match",
"(",
"label",
")",
"if",
"m",
":",
"column_label",
"=",
"m",
".",
"group",
"(",
"1",
")",
".",
"upper",
"(",
")",
"row",
"=",
"int",
"(",
"m",
".",
"group",
"(",
"2",
")",
")",
"col",
"=",
"0",
"for",
"(",
"i",
",",
"c",
")",
"in",
"enumerate",
"(",
"reversed",
"(",
"column_label",
")",
")",
":",
"col",
"+=",
"(",
"(",
"ord",
"(",
"c",
")",
"-",
"MAGIC_NUMBER",
")",
"*",
"(",
"26",
"**",
"i",
")",
")",
"else",
":",
"raise",
"IncorrectCellLabel",
"(",
"label",
")",
"return",
"(",
"row",
",",
"col",
")"
] | translates a cells address in a1 notation to a tuple of integers . | train | true |
17,550 | def find_split_rechunk(old_chunks, new_chunks, graph_size_limit):
ndim = len(old_chunks)
chunks = list(old_chunks)
for dim in range(ndim):
graph_size = estimate_graph_size(chunks, new_chunks)
if (graph_size > graph_size_limit):
break
if (len(old_chunks[dim]) > len(new_chunks[dim])):
continue
max_number = int(((len(old_chunks[dim]) * graph_size_limit) / graph_size))
c = merge_to_number(new_chunks[dim], max_number)
assert (len(c) <= max_number)
if ((len(c) >= len(old_chunks[dim])) and (max(c) <= max(old_chunks[dim]))):
chunks[dim] = c
return tuple(chunks)
| [
"def",
"find_split_rechunk",
"(",
"old_chunks",
",",
"new_chunks",
",",
"graph_size_limit",
")",
":",
"ndim",
"=",
"len",
"(",
"old_chunks",
")",
"chunks",
"=",
"list",
"(",
"old_chunks",
")",
"for",
"dim",
"in",
"range",
"(",
"ndim",
")",
":",
"graph_size",
"=",
"estimate_graph_size",
"(",
"chunks",
",",
"new_chunks",
")",
"if",
"(",
"graph_size",
">",
"graph_size_limit",
")",
":",
"break",
"if",
"(",
"len",
"(",
"old_chunks",
"[",
"dim",
"]",
")",
">",
"len",
"(",
"new_chunks",
"[",
"dim",
"]",
")",
")",
":",
"continue",
"max_number",
"=",
"int",
"(",
"(",
"(",
"len",
"(",
"old_chunks",
"[",
"dim",
"]",
")",
"*",
"graph_size_limit",
")",
"/",
"graph_size",
")",
")",
"c",
"=",
"merge_to_number",
"(",
"new_chunks",
"[",
"dim",
"]",
",",
"max_number",
")",
"assert",
"(",
"len",
"(",
"c",
")",
"<=",
"max_number",
")",
"if",
"(",
"(",
"len",
"(",
"c",
")",
">=",
"len",
"(",
"old_chunks",
"[",
"dim",
"]",
")",
")",
"and",
"(",
"max",
"(",
"c",
")",
"<=",
"max",
"(",
"old_chunks",
"[",
"dim",
"]",
")",
")",
")",
":",
"chunks",
"[",
"dim",
"]",
"=",
"c",
"return",
"tuple",
"(",
"chunks",
")"
] | find an intermediate rechunk that would split some chunks to get us nearer *new_chunks* . | train | false |
17,551 | def test_filelock_passes(tmpdir):
package_name = 'conda_file1'
tmpfile = join(tmpdir.strpath, package_name)
with FileLock(tmpfile) as lock:
path = basename(lock.lock_file_path)
assert (tmpdir.join(path).exists() and tmpdir.join(path).isfile())
assert (not tmpdir.join(path).exists())
| [
"def",
"test_filelock_passes",
"(",
"tmpdir",
")",
":",
"package_name",
"=",
"'conda_file1'",
"tmpfile",
"=",
"join",
"(",
"tmpdir",
".",
"strpath",
",",
"package_name",
")",
"with",
"FileLock",
"(",
"tmpfile",
")",
"as",
"lock",
":",
"path",
"=",
"basename",
"(",
"lock",
".",
"lock_file_path",
")",
"assert",
"(",
"tmpdir",
".",
"join",
"(",
"path",
")",
".",
"exists",
"(",
")",
"and",
"tmpdir",
".",
"join",
"(",
"path",
")",
".",
"isfile",
"(",
")",
")",
"assert",
"(",
"not",
"tmpdir",
".",
"join",
"(",
"path",
")",
".",
"exists",
"(",
")",
")"
] | normal test on file lock . | train | false |
17,553 | def lz4_decode(payload):
ctx = lz4f.createDecompContext()
data = lz4f.decompressFrame(payload, ctx)
if (data['next'] != 0):
raise RuntimeError('lz4f unable to decompress full payload')
return data['decomp']
| [
"def",
"lz4_decode",
"(",
"payload",
")",
":",
"ctx",
"=",
"lz4f",
".",
"createDecompContext",
"(",
")",
"data",
"=",
"lz4f",
".",
"decompressFrame",
"(",
"payload",
",",
"ctx",
")",
"if",
"(",
"data",
"[",
"'next'",
"]",
"!=",
"0",
")",
":",
"raise",
"RuntimeError",
"(",
"'lz4f unable to decompress full payload'",
")",
"return",
"data",
"[",
"'decomp'",
"]"
] | decode payload using interoperable lz4 framing . | train | false |
17,556 | def extract_gcs_tokens(full_object_name):
bucket_name = None
object_name = None
tokens = full_object_name.split('/')
if (len(tokens) < 3):
logging.error("Malformed GCS path '{0}'. Aborting GCS operation.".format(full_object_name))
return (bucket_name, object_name)
bucket_name = tokens[2]
object_name = ''
for token in tokens[3:(-1)]:
object_name += (token + '/')
object_name += tokens[(-1)]
return (bucket_name, object_name)
| [
"def",
"extract_gcs_tokens",
"(",
"full_object_name",
")",
":",
"bucket_name",
"=",
"None",
"object_name",
"=",
"None",
"tokens",
"=",
"full_object_name",
".",
"split",
"(",
"'/'",
")",
"if",
"(",
"len",
"(",
"tokens",
")",
"<",
"3",
")",
":",
"logging",
".",
"error",
"(",
"\"Malformed GCS path '{0}'. Aborting GCS operation.\"",
".",
"format",
"(",
"full_object_name",
")",
")",
"return",
"(",
"bucket_name",
",",
"object_name",
")",
"bucket_name",
"=",
"tokens",
"[",
"2",
"]",
"object_name",
"=",
"''",
"for",
"token",
"in",
"tokens",
"[",
"3",
":",
"(",
"-",
"1",
")",
"]",
":",
"object_name",
"+=",
"(",
"token",
"+",
"'/'",
")",
"object_name",
"+=",
"tokens",
"[",
"(",
"-",
"1",
")",
"]",
"return",
"(",
"bucket_name",
",",
"object_name",
")"
] | extracts the bucket and object name from a full gcs path . | train | false |
17,557 | def RemoveFlags(flag_values=FLAGS):
for flag_name in NamesOfDefinedFlags():
module_bar.RemoveOneFlag(flag_name, flag_values=flag_values)
module_bar.RemoveFlags(flag_values=flag_values)
| [
"def",
"RemoveFlags",
"(",
"flag_values",
"=",
"FLAGS",
")",
":",
"for",
"flag_name",
"in",
"NamesOfDefinedFlags",
"(",
")",
":",
"module_bar",
".",
"RemoveOneFlag",
"(",
"flag_name",
",",
"flag_values",
"=",
"flag_values",
")",
"module_bar",
".",
"RemoveFlags",
"(",
"flag_values",
"=",
"flag_values",
")"
] | deletes the flag definitions done by the above defineflags() . | train | false |
17,558 | def unpack_bitstring(string):
byte_count = len(string)
bits = []
for byte in range(byte_count):
value = ord(string[byte])
for _ in range(8):
bits.append(((value & 1) == 1))
value >>= 1
return bits
| [
"def",
"unpack_bitstring",
"(",
"string",
")",
":",
"byte_count",
"=",
"len",
"(",
"string",
")",
"bits",
"=",
"[",
"]",
"for",
"byte",
"in",
"range",
"(",
"byte_count",
")",
":",
"value",
"=",
"ord",
"(",
"string",
"[",
"byte",
"]",
")",
"for",
"_",
"in",
"range",
"(",
"8",
")",
":",
"bits",
".",
"append",
"(",
"(",
"(",
"value",
"&",
"1",
")",
"==",
"1",
")",
")",
"value",
">>=",
"1",
"return",
"bits"
] | creates bit array out of a string . | train | false |
17,560 | def require_user(handler):
def test_login(self, **kwargs):
'Checks if the user for the current session is logged in.'
if (not self.user_id):
self.redirect(current_user_services.create_login_url(self.request.uri))
return
return handler(self, **kwargs)
return test_login
| [
"def",
"require_user",
"(",
"handler",
")",
":",
"def",
"test_login",
"(",
"self",
",",
"**",
"kwargs",
")",
":",
"if",
"(",
"not",
"self",
".",
"user_id",
")",
":",
"self",
".",
"redirect",
"(",
"current_user_services",
".",
"create_login_url",
"(",
"self",
".",
"request",
".",
"uri",
")",
")",
"return",
"return",
"handler",
"(",
"self",
",",
"**",
"kwargs",
")",
"return",
"test_login"
] | decorator that checks if a user is associated to the current session . | train | false |
17,561 | def slip_reader(port):
partial_packet = None
in_escape = False
while True:
waiting = port.inWaiting()
read_bytes = port.read((1 if (waiting == 0) else waiting))
if (read_bytes == ''):
raise FatalError(('Timed out waiting for packet %s' % ('header' if (partial_packet is None) else 'content')))
for b in read_bytes:
if (type(b) is int):
b = bytes([b])
if (partial_packet is None):
if (b == '\xc0'):
partial_packet = ''
else:
raise FatalError(('Invalid head of packet (%r)' % b))
elif in_escape:
in_escape = False
if (b == '\xdc'):
partial_packet += '\xc0'
elif (b == '\xdd'):
partial_packet += '\xdb'
else:
raise FatalError(('Invalid SLIP escape (%r%r)' % ('\xdb', b)))
elif (b == '\xdb'):
in_escape = True
elif (b == '\xc0'):
(yield partial_packet)
partial_packet = None
else:
partial_packet += b
| [
"def",
"slip_reader",
"(",
"port",
")",
":",
"partial_packet",
"=",
"None",
"in_escape",
"=",
"False",
"while",
"True",
":",
"waiting",
"=",
"port",
".",
"inWaiting",
"(",
")",
"read_bytes",
"=",
"port",
".",
"read",
"(",
"(",
"1",
"if",
"(",
"waiting",
"==",
"0",
")",
"else",
"waiting",
")",
")",
"if",
"(",
"read_bytes",
"==",
"''",
")",
":",
"raise",
"FatalError",
"(",
"(",
"'Timed out waiting for packet %s'",
"%",
"(",
"'header'",
"if",
"(",
"partial_packet",
"is",
"None",
")",
"else",
"'content'",
")",
")",
")",
"for",
"b",
"in",
"read_bytes",
":",
"if",
"(",
"type",
"(",
"b",
")",
"is",
"int",
")",
":",
"b",
"=",
"bytes",
"(",
"[",
"b",
"]",
")",
"if",
"(",
"partial_packet",
"is",
"None",
")",
":",
"if",
"(",
"b",
"==",
"'\\xc0'",
")",
":",
"partial_packet",
"=",
"''",
"else",
":",
"raise",
"FatalError",
"(",
"(",
"'Invalid head of packet (%r)'",
"%",
"b",
")",
")",
"elif",
"in_escape",
":",
"in_escape",
"=",
"False",
"if",
"(",
"b",
"==",
"'\\xdc'",
")",
":",
"partial_packet",
"+=",
"'\\xc0'",
"elif",
"(",
"b",
"==",
"'\\xdd'",
")",
":",
"partial_packet",
"+=",
"'\\xdb'",
"else",
":",
"raise",
"FatalError",
"(",
"(",
"'Invalid SLIP escape (%r%r)'",
"%",
"(",
"'\\xdb'",
",",
"b",
")",
")",
")",
"elif",
"(",
"b",
"==",
"'\\xdb'",
")",
":",
"in_escape",
"=",
"True",
"elif",
"(",
"b",
"==",
"'\\xc0'",
")",
":",
"(",
"yield",
"partial_packet",
")",
"partial_packet",
"=",
"None",
"else",
":",
"partial_packet",
"+=",
"b"
] | generator to read slip packets from a serial port . | train | false |
17,562 | def test_brainvision_data_partially_disabled_hw_filters():
with warnings.catch_warnings(record=True) as w:
raw = _test_raw_reader(read_raw_brainvision, vhdr_fname=vhdr_partially_disabled_hw_filter_path, montage=montage, eog=eog)
trigger_warning = [('parse triggers that' in str(ww.message)) for ww in w]
lowpass_warning = [('different lowpass filters' in str(ww.message)) for ww in w]
highpass_warning = [('different highpass filters' in str(ww.message)) for ww in w]
expected_warnings = zip(trigger_warning, lowpass_warning, highpass_warning)
assert_true(all((any([trg, lp, hp]) for (trg, lp, hp) in expected_warnings)))
assert_equal(raw.info['highpass'], 0.0)
assert_equal(raw.info['lowpass'], 500.0)
| [
"def",
"test_brainvision_data_partially_disabled_hw_filters",
"(",
")",
":",
"with",
"warnings",
".",
"catch_warnings",
"(",
"record",
"=",
"True",
")",
"as",
"w",
":",
"raw",
"=",
"_test_raw_reader",
"(",
"read_raw_brainvision",
",",
"vhdr_fname",
"=",
"vhdr_partially_disabled_hw_filter_path",
",",
"montage",
"=",
"montage",
",",
"eog",
"=",
"eog",
")",
"trigger_warning",
"=",
"[",
"(",
"'parse triggers that'",
"in",
"str",
"(",
"ww",
".",
"message",
")",
")",
"for",
"ww",
"in",
"w",
"]",
"lowpass_warning",
"=",
"[",
"(",
"'different lowpass filters'",
"in",
"str",
"(",
"ww",
".",
"message",
")",
")",
"for",
"ww",
"in",
"w",
"]",
"highpass_warning",
"=",
"[",
"(",
"'different highpass filters'",
"in",
"str",
"(",
"ww",
".",
"message",
")",
")",
"for",
"ww",
"in",
"w",
"]",
"expected_warnings",
"=",
"zip",
"(",
"trigger_warning",
",",
"lowpass_warning",
",",
"highpass_warning",
")",
"assert_true",
"(",
"all",
"(",
"(",
"any",
"(",
"[",
"trg",
",",
"lp",
",",
"hp",
"]",
")",
"for",
"(",
"trg",
",",
"lp",
",",
"hp",
")",
"in",
"expected_warnings",
")",
")",
")",
"assert_equal",
"(",
"raw",
".",
"info",
"[",
"'highpass'",
"]",
",",
"0.0",
")",
"assert_equal",
"(",
"raw",
".",
"info",
"[",
"'lowpass'",
"]",
",",
"500.0",
")"
] | test heterogeneous filter settings including non-numeric values . | train | false |
17,563 | def soft_break(value, length, process=(lambda chunk: chunk)):
delimiters = re.compile('([{}]+)'.format(''.join(map(re.escape, ',.$:/+@!?()<>[]{}'))))
def soft_break_delimiter(match):
results = []
value = match.group(0)
chunks = delimiters.split(value)
for (i, chunk) in enumerate(chunks):
if ((i % 2) == 1):
results.extend([chunk, u'\u200b'])
else:
results.append(process(chunk))
return u''.join(results).rstrip(u'\u200b')
return re.sub('\\S{{{},}}'.format(length), soft_break_delimiter, value)
| [
"def",
"soft_break",
"(",
"value",
",",
"length",
",",
"process",
"=",
"(",
"lambda",
"chunk",
":",
"chunk",
")",
")",
":",
"delimiters",
"=",
"re",
".",
"compile",
"(",
"'([{}]+)'",
".",
"format",
"(",
"''",
".",
"join",
"(",
"map",
"(",
"re",
".",
"escape",
",",
"',.$:/+@!?()<>[]{}'",
")",
")",
")",
")",
"def",
"soft_break_delimiter",
"(",
"match",
")",
":",
"results",
"=",
"[",
"]",
"value",
"=",
"match",
".",
"group",
"(",
"0",
")",
"chunks",
"=",
"delimiters",
".",
"split",
"(",
"value",
")",
"for",
"(",
"i",
",",
"chunk",
")",
"in",
"enumerate",
"(",
"chunks",
")",
":",
"if",
"(",
"(",
"i",
"%",
"2",
")",
"==",
"1",
")",
":",
"results",
".",
"extend",
"(",
"[",
"chunk",
",",
"u'\\u200b'",
"]",
")",
"else",
":",
"results",
".",
"append",
"(",
"process",
"(",
"chunk",
")",
")",
"return",
"u''",
".",
"join",
"(",
"results",
")",
".",
"rstrip",
"(",
"u'\\u200b'",
")",
"return",
"re",
".",
"sub",
"(",
"'\\\\S{{{},}}'",
".",
"format",
"(",
"length",
")",
",",
"soft_break_delimiter",
",",
"value",
")"
] | encourages soft breaking of text values above a maximum length by adding zero-width spaces after common delimeters . | train | false |
17,564 | def put_tagging(Bucket, region=None, key=None, keyid=None, profile=None, **kwargs):
try:
conn = _get_conn(region=region, key=key, keyid=keyid, profile=profile)
tagslist = []
for (k, v) in six.iteritems(kwargs):
if str(k).startswith('__'):
continue
tagslist.append({'Key': str(k), 'Value': str(v)})
conn.put_bucket_tagging(Bucket=Bucket, Tagging={'TagSet': tagslist})
return {'updated': True, 'name': Bucket}
except ClientError as e:
return {'updated': False, 'error': __utils__['boto3.get_error'](e)}
| [
"def",
"put_tagging",
"(",
"Bucket",
",",
"region",
"=",
"None",
",",
"key",
"=",
"None",
",",
"keyid",
"=",
"None",
",",
"profile",
"=",
"None",
",",
"**",
"kwargs",
")",
":",
"try",
":",
"conn",
"=",
"_get_conn",
"(",
"region",
"=",
"region",
",",
"key",
"=",
"key",
",",
"keyid",
"=",
"keyid",
",",
"profile",
"=",
"profile",
")",
"tagslist",
"=",
"[",
"]",
"for",
"(",
"k",
",",
"v",
")",
"in",
"six",
".",
"iteritems",
"(",
"kwargs",
")",
":",
"if",
"str",
"(",
"k",
")",
".",
"startswith",
"(",
"'__'",
")",
":",
"continue",
"tagslist",
".",
"append",
"(",
"{",
"'Key'",
":",
"str",
"(",
"k",
")",
",",
"'Value'",
":",
"str",
"(",
"v",
")",
"}",
")",
"conn",
".",
"put_bucket_tagging",
"(",
"Bucket",
"=",
"Bucket",
",",
"Tagging",
"=",
"{",
"'TagSet'",
":",
"tagslist",
"}",
")",
"return",
"{",
"'updated'",
":",
"True",
",",
"'name'",
":",
"Bucket",
"}",
"except",
"ClientError",
"as",
"e",
":",
"return",
"{",
"'updated'",
":",
"False",
",",
"'error'",
":",
"__utils__",
"[",
"'boto3.get_error'",
"]",
"(",
"e",
")",
"}"
] | given a valid config . | train | true |
17,566 | def create_gzipped_copy(in_path, out_path, command=None):
if command:
subprocess.check_call(shlex.split(command.format(filename=in_path)))
else:
with gzip.GzipFile(out_path, 'wb+') as outf:
with open(in_path, 'rb') as inf:
outf.write(inf.read())
| [
"def",
"create_gzipped_copy",
"(",
"in_path",
",",
"out_path",
",",
"command",
"=",
"None",
")",
":",
"if",
"command",
":",
"subprocess",
".",
"check_call",
"(",
"shlex",
".",
"split",
"(",
"command",
".",
"format",
"(",
"filename",
"=",
"in_path",
")",
")",
")",
"else",
":",
"with",
"gzip",
".",
"GzipFile",
"(",
"out_path",
",",
"'wb+'",
")",
"as",
"outf",
":",
"with",
"open",
"(",
"in_path",
",",
"'rb'",
")",
"as",
"inf",
":",
"outf",
".",
"write",
"(",
"inf",
".",
"read",
"(",
")",
")"
] | create gzipped copy of in_path and save it as out_path . | train | false |
17,567 | def get_variables_with_name(name, train_only=True, printable=False):
print (' Get variables with %s' % name)
t_vars = (tf.trainable_variables() if train_only else tf.all_variables())
d_vars = [var for var in t_vars if (name in var.name)]
if printable:
for (idx, v) in enumerate(d_vars):
print ' got {:3}: {:15} {}'.format(idx, v.name, str(v.get_shape()))
return d_vars
| [
"def",
"get_variables_with_name",
"(",
"name",
",",
"train_only",
"=",
"True",
",",
"printable",
"=",
"False",
")",
":",
"print",
"(",
"' Get variables with %s'",
"%",
"name",
")",
"t_vars",
"=",
"(",
"tf",
".",
"trainable_variables",
"(",
")",
"if",
"train_only",
"else",
"tf",
".",
"all_variables",
"(",
")",
")",
"d_vars",
"=",
"[",
"var",
"for",
"var",
"in",
"t_vars",
"if",
"(",
"name",
"in",
"var",
".",
"name",
")",
"]",
"if",
"printable",
":",
"for",
"(",
"idx",
",",
"v",
")",
"in",
"enumerate",
"(",
"d_vars",
")",
":",
"print",
"' got {:3}: {:15} {}'",
".",
"format",
"(",
"idx",
",",
"v",
".",
"name",
",",
"str",
"(",
"v",
".",
"get_shape",
"(",
")",
")",
")",
"return",
"d_vars"
] | get variable list by a given name scope . | train | false |
17,568 | def p_postfix_expression_8(t):
pass
| [
"def",
"p_postfix_expression_8",
"(",
"t",
")",
":",
"pass"
] | postfix_expression : postfix_expression minusminus . | train | false |
17,569 | def get_custom_fields(data):
data_copy = deepcopy(data)
field_names = {}
for (key, value) in data.iteritems():
if (key in DEFAULT_FIELD_NAMES):
field_names[key] = data_copy.pop(key)
return (field_names, data_copy)
| [
"def",
"get_custom_fields",
"(",
"data",
")",
":",
"data_copy",
"=",
"deepcopy",
"(",
"data",
")",
"field_names",
"=",
"{",
"}",
"for",
"(",
"key",
",",
"value",
")",
"in",
"data",
".",
"iteritems",
"(",
")",
":",
"if",
"(",
"key",
"in",
"DEFAULT_FIELD_NAMES",
")",
":",
"field_names",
"[",
"key",
"]",
"=",
"data_copy",
".",
"pop",
"(",
"key",
")",
"return",
"(",
"field_names",
",",
"data_copy",
")"
] | return two dicts . | train | false |
17,570 | def _match_mismatch_counter(s1, p1, s2, p2):
if (s1[p1] == s2[p2]):
cmp_fn = operator.eq
match = True
else:
cmp_fn = operator.ne
match = False
count = 1
p1 += 1
p2 += 1
while ((p1 < len(s1)) and (p2 < len(s2)) and cmp_fn(s1[p1], s2[p2])):
count += 1
p1 += 1
p2 += 1
return (match, count)
| [
"def",
"_match_mismatch_counter",
"(",
"s1",
",",
"p1",
",",
"s2",
",",
"p2",
")",
":",
"if",
"(",
"s1",
"[",
"p1",
"]",
"==",
"s2",
"[",
"p2",
"]",
")",
":",
"cmp_fn",
"=",
"operator",
".",
"eq",
"match",
"=",
"True",
"else",
":",
"cmp_fn",
"=",
"operator",
".",
"ne",
"match",
"=",
"False",
"count",
"=",
"1",
"p1",
"+=",
"1",
"p2",
"+=",
"1",
"while",
"(",
"(",
"p1",
"<",
"len",
"(",
"s1",
")",
")",
"and",
"(",
"p2",
"<",
"len",
"(",
"s2",
")",
")",
"and",
"cmp_fn",
"(",
"s1",
"[",
"p1",
"]",
",",
"s2",
"[",
"p2",
"]",
")",
")",
":",
"count",
"+=",
"1",
"p1",
"+=",
"1",
"p2",
"+=",
"1",
"return",
"(",
"match",
",",
"count",
")"
] | count consecutive matches/mismatches between strings s1 and s2 starting at p1 and p2 . | train | false |
17,572 | def _setupTempDirectory(filename):
tmpDir = tempfile.mkdtemp()
tmpFileName = os.path.join(tmpDir, os.path.basename(filename))
return (tmpDir, tmpFileName)
| [
"def",
"_setupTempDirectory",
"(",
"filename",
")",
":",
"tmpDir",
"=",
"tempfile",
".",
"mkdtemp",
"(",
")",
"tmpFileName",
"=",
"os",
".",
"path",
".",
"join",
"(",
"tmpDir",
",",
"os",
".",
"path",
".",
"basename",
"(",
"filename",
")",
")",
"return",
"(",
"tmpDir",
",",
"tmpFileName",
")"
] | create a temp directory . | train | false |
17,574 | def LocateFileName(fileNamesString, searchPaths):
import regutil, string, os
fileNames = string.split(fileNamesString, ';')
for path in searchPaths:
for fileName in fileNames:
try:
retPath = os.path.join(path, fileName)
os.stat(retPath)
break
except os.error:
retPath = None
if retPath:
break
else:
fileName = fileNames[0]
try:
import win32ui, win32con
except ImportError:
raise error, ('Need to locate the file %s, but the win32ui module is not available\nPlease run the program again, passing as a parameter the path to this file.' % fileName)
flags = win32con.OFN_FILEMUSTEXIST
ext = os.path.splitext(fileName)[1]
filter = ('Files of requested type (*%s)|*%s||' % (ext, ext))
dlg = win32ui.CreateFileDialog(1, None, fileName, flags, filter, None)
dlg.SetOFNTitle(('Locate ' + fileName))
if (dlg.DoModal() != win32con.IDOK):
raise KeyboardInterrupt, 'User cancelled the process'
retPath = dlg.GetPathName()
return os.path.abspath(retPath)
| [
"def",
"LocateFileName",
"(",
"fileNamesString",
",",
"searchPaths",
")",
":",
"import",
"regutil",
",",
"string",
",",
"os",
"fileNames",
"=",
"string",
".",
"split",
"(",
"fileNamesString",
",",
"';'",
")",
"for",
"path",
"in",
"searchPaths",
":",
"for",
"fileName",
"in",
"fileNames",
":",
"try",
":",
"retPath",
"=",
"os",
".",
"path",
".",
"join",
"(",
"path",
",",
"fileName",
")",
"os",
".",
"stat",
"(",
"retPath",
")",
"break",
"except",
"os",
".",
"error",
":",
"retPath",
"=",
"None",
"if",
"retPath",
":",
"break",
"else",
":",
"fileName",
"=",
"fileNames",
"[",
"0",
"]",
"try",
":",
"import",
"win32ui",
",",
"win32con",
"except",
"ImportError",
":",
"raise",
"error",
",",
"(",
"'Need to locate the file %s, but the win32ui module is not available\\nPlease run the program again, passing as a parameter the path to this file.'",
"%",
"fileName",
")",
"flags",
"=",
"win32con",
".",
"OFN_FILEMUSTEXIST",
"ext",
"=",
"os",
".",
"path",
".",
"splitext",
"(",
"fileName",
")",
"[",
"1",
"]",
"filter",
"=",
"(",
"'Files of requested type (*%s)|*%s||'",
"%",
"(",
"ext",
",",
"ext",
")",
")",
"dlg",
"=",
"win32ui",
".",
"CreateFileDialog",
"(",
"1",
",",
"None",
",",
"fileName",
",",
"flags",
",",
"filter",
",",
"None",
")",
"dlg",
".",
"SetOFNTitle",
"(",
"(",
"'Locate '",
"+",
"fileName",
")",
")",
"if",
"(",
"dlg",
".",
"DoModal",
"(",
")",
"!=",
"win32con",
".",
"IDOK",
")",
":",
"raise",
"KeyboardInterrupt",
",",
"'User cancelled the process'",
"retPath",
"=",
"dlg",
".",
"GetPathName",
"(",
")",
"return",
"os",
".",
"path",
".",
"abspath",
"(",
"retPath",
")"
] | locate a file name . | train | false |
17,576 | def _gen_get_more_command(cursor_id, coll, batch_size, max_await_time_ms):
cmd = SON([('getMore', cursor_id), ('collection', coll)])
if batch_size:
cmd['batchSize'] = batch_size
if (max_await_time_ms is not None):
cmd['maxTimeMS'] = max_await_time_ms
return cmd
| [
"def",
"_gen_get_more_command",
"(",
"cursor_id",
",",
"coll",
",",
"batch_size",
",",
"max_await_time_ms",
")",
":",
"cmd",
"=",
"SON",
"(",
"[",
"(",
"'getMore'",
",",
"cursor_id",
")",
",",
"(",
"'collection'",
",",
"coll",
")",
"]",
")",
"if",
"batch_size",
":",
"cmd",
"[",
"'batchSize'",
"]",
"=",
"batch_size",
"if",
"(",
"max_await_time_ms",
"is",
"not",
"None",
")",
":",
"cmd",
"[",
"'maxTimeMS'",
"]",
"=",
"max_await_time_ms",
"return",
"cmd"
] | generate a getmore command document . | train | true |
17,577 | def library_keyword_tags_with_documentation():
pass
| [
"def",
"library_keyword_tags_with_documentation",
"(",
")",
":",
"pass"
] | summary line tags: are read only from the last line tags: one . | train | false |
17,578 | def register_plugin_class(base_class, file_path, class_name):
plugin_dir = os.path.dirname(os.path.realpath(file_path))
_register_plugin_path(plugin_dir)
module_name = _get_plugin_module(file_path)
if (module_name is None):
return None
module = imp.load_source(module_name, file_path)
klass = getattr(module, class_name, None)
if (not klass):
raise Exception(('Plugin file "%s" doesn\'t expose class named "%s"' % (file_path, class_name)))
_register_plugin(base_class, klass)
return klass
| [
"def",
"register_plugin_class",
"(",
"base_class",
",",
"file_path",
",",
"class_name",
")",
":",
"plugin_dir",
"=",
"os",
".",
"path",
".",
"dirname",
"(",
"os",
".",
"path",
".",
"realpath",
"(",
"file_path",
")",
")",
"_register_plugin_path",
"(",
"plugin_dir",
")",
"module_name",
"=",
"_get_plugin_module",
"(",
"file_path",
")",
"if",
"(",
"module_name",
"is",
"None",
")",
":",
"return",
"None",
"module",
"=",
"imp",
".",
"load_source",
"(",
"module_name",
",",
"file_path",
")",
"klass",
"=",
"getattr",
"(",
"module",
",",
"class_name",
",",
"None",
")",
"if",
"(",
"not",
"klass",
")",
":",
"raise",
"Exception",
"(",
"(",
"'Plugin file \"%s\" doesn\\'t expose class named \"%s\"'",
"%",
"(",
"file_path",
",",
"class_name",
")",
")",
")",
"_register_plugin",
"(",
"base_class",
",",
"klass",
")",
"return",
"klass"
] | retrieve a register plugin class from the provided file . | train | false |
17,579 | @requires_version('scipy', '0.16')
def test_estimate_ringing():
for kind in ('ba', 'sos'):
for (thresh, lims) in ((0.1, (30, 60)), (0.01, (300, 600)), (0.001, (3000, 6000)), (0.0001, (30000, 60000))):
n_ring = estimate_ringing_samples(butter(3, thresh, output=kind))
assert_true((lims[0] <= n_ring <= lims[1]), msg=('%s %s: %s <= %s <= %s' % (kind, thresh, lims[0], n_ring, lims[1])))
with warnings.catch_warnings(record=True) as w:
assert_equal(estimate_ringing_samples(butter(4, 1e-05)), 100000)
assert_true(any((('properly estimate' in str(ww.message)) for ww in w)))
| [
"@",
"requires_version",
"(",
"'scipy'",
",",
"'0.16'",
")",
"def",
"test_estimate_ringing",
"(",
")",
":",
"for",
"kind",
"in",
"(",
"'ba'",
",",
"'sos'",
")",
":",
"for",
"(",
"thresh",
",",
"lims",
")",
"in",
"(",
"(",
"0.1",
",",
"(",
"30",
",",
"60",
")",
")",
",",
"(",
"0.01",
",",
"(",
"300",
",",
"600",
")",
")",
",",
"(",
"0.001",
",",
"(",
"3000",
",",
"6000",
")",
")",
",",
"(",
"0.0001",
",",
"(",
"30000",
",",
"60000",
")",
")",
")",
":",
"n_ring",
"=",
"estimate_ringing_samples",
"(",
"butter",
"(",
"3",
",",
"thresh",
",",
"output",
"=",
"kind",
")",
")",
"assert_true",
"(",
"(",
"lims",
"[",
"0",
"]",
"<=",
"n_ring",
"<=",
"lims",
"[",
"1",
"]",
")",
",",
"msg",
"=",
"(",
"'%s %s: %s <= %s <= %s'",
"%",
"(",
"kind",
",",
"thresh",
",",
"lims",
"[",
"0",
"]",
",",
"n_ring",
",",
"lims",
"[",
"1",
"]",
")",
")",
")",
"with",
"warnings",
".",
"catch_warnings",
"(",
"record",
"=",
"True",
")",
"as",
"w",
":",
"assert_equal",
"(",
"estimate_ringing_samples",
"(",
"butter",
"(",
"4",
",",
"1e-05",
")",
")",
",",
"100000",
")",
"assert_true",
"(",
"any",
"(",
"(",
"(",
"'properly estimate'",
"in",
"str",
"(",
"ww",
".",
"message",
")",
")",
"for",
"ww",
"in",
"w",
")",
")",
")"
] | test our ringing estimation function . | train | false |
17,581 | @pytest.fixture
def fake_statusbar(qtbot):
container = QWidget()
qtbot.add_widget(container)
vbox = QVBoxLayout(container)
vbox.addStretch()
statusbar = FakeStatusBar(container)
statusbar.container = container
vbox.addWidget(statusbar)
with qtbot.waitExposed(container):
container.show()
return statusbar
| [
"@",
"pytest",
".",
"fixture",
"def",
"fake_statusbar",
"(",
"qtbot",
")",
":",
"container",
"=",
"QWidget",
"(",
")",
"qtbot",
".",
"add_widget",
"(",
"container",
")",
"vbox",
"=",
"QVBoxLayout",
"(",
"container",
")",
"vbox",
".",
"addStretch",
"(",
")",
"statusbar",
"=",
"FakeStatusBar",
"(",
"container",
")",
"statusbar",
".",
"container",
"=",
"container",
"vbox",
".",
"addWidget",
"(",
"statusbar",
")",
"with",
"qtbot",
".",
"waitExposed",
"(",
"container",
")",
":",
"container",
".",
"show",
"(",
")",
"return",
"statusbar"
] | fixture providing a statusbar in a container window . | train | false |
17,582 | def frag_remover(ack, load):
global pkt_frag_loads
while (len(pkt_frag_loads) > 50):
pkt_frag_loads.popitem(last=False)
copy_pkt_frag_loads = copy.deepcopy(pkt_frag_loads)
for ip_port in copy_pkt_frag_loads:
if (len(copy_pkt_frag_loads[ip_port]) > 0):
while (len(copy_pkt_frag_loads[ip_port]) > 25):
pkt_frag_loads[ip_port].popitem(last=False)
copy_pkt_frag_loads = copy.deepcopy(pkt_frag_loads)
for ip_port in copy_pkt_frag_loads:
for ack in copy_pkt_frag_loads[ip_port]:
if (len(copy_pkt_frag_loads[ip_port][ack]) > 5000):
pkt_frag_loads[ip_port][ack] = pkt_frag_loads[ip_port][ack][(-200):]
| [
"def",
"frag_remover",
"(",
"ack",
",",
"load",
")",
":",
"global",
"pkt_frag_loads",
"while",
"(",
"len",
"(",
"pkt_frag_loads",
")",
">",
"50",
")",
":",
"pkt_frag_loads",
".",
"popitem",
"(",
"last",
"=",
"False",
")",
"copy_pkt_frag_loads",
"=",
"copy",
".",
"deepcopy",
"(",
"pkt_frag_loads",
")",
"for",
"ip_port",
"in",
"copy_pkt_frag_loads",
":",
"if",
"(",
"len",
"(",
"copy_pkt_frag_loads",
"[",
"ip_port",
"]",
")",
">",
"0",
")",
":",
"while",
"(",
"len",
"(",
"copy_pkt_frag_loads",
"[",
"ip_port",
"]",
")",
">",
"25",
")",
":",
"pkt_frag_loads",
"[",
"ip_port",
"]",
".",
"popitem",
"(",
"last",
"=",
"False",
")",
"copy_pkt_frag_loads",
"=",
"copy",
".",
"deepcopy",
"(",
"pkt_frag_loads",
")",
"for",
"ip_port",
"in",
"copy_pkt_frag_loads",
":",
"for",
"ack",
"in",
"copy_pkt_frag_loads",
"[",
"ip_port",
"]",
":",
"if",
"(",
"len",
"(",
"copy_pkt_frag_loads",
"[",
"ip_port",
"]",
"[",
"ack",
"]",
")",
">",
"5000",
")",
":",
"pkt_frag_loads",
"[",
"ip_port",
"]",
"[",
"ack",
"]",
"=",
"pkt_frag_loads",
"[",
"ip_port",
"]",
"[",
"ack",
"]",
"[",
"(",
"-",
"200",
")",
":",
"]"
] | keep the filo ordereddict of frag loads from getting too large 3 points of limit: number of ip_ports < 50 number of acks per ip:port < 25 number of chars in load < 5000 . | train | false |
17,586 | def edgeNormals(v, e):
return mapEdges((lambda a, (p, n): getNormal(a, p, n)), v, e)
| [
"def",
"edgeNormals",
"(",
"v",
",",
"e",
")",
":",
"return",
"mapEdges",
"(",
"(",
"lambda",
"a",
",",
"(",
"p",
",",
"n",
")",
":",
"getNormal",
"(",
"a",
",",
"p",
",",
"n",
")",
")",
",",
"v",
",",
"e",
")"
] | assumes a mesh where each vertex has exactly least two edges . | train | false |
17,587 | def get_github_email(access_token):
request = Request(u'https://api.github.com/user/emails')
request.timeout = 1.0
request.add_header(u'User-Agent', USER_AGENT)
request.add_header(u'Authorization', u'token {0}'.format(access_token))
handle = urlopen(request)
data = json.loads(handle.read().decode(u'utf-8'))
email = None
for entry in data:
if (not entry[u'verified']):
continue
email = entry[u'email']
if entry[u'primary']:
break
return email
| [
"def",
"get_github_email",
"(",
"access_token",
")",
":",
"request",
"=",
"Request",
"(",
"u'https://api.github.com/user/emails'",
")",
"request",
".",
"timeout",
"=",
"1.0",
"request",
".",
"add_header",
"(",
"u'User-Agent'",
",",
"USER_AGENT",
")",
"request",
".",
"add_header",
"(",
"u'Authorization'",
",",
"u'token {0}'",
".",
"format",
"(",
"access_token",
")",
")",
"handle",
"=",
"urlopen",
"(",
"request",
")",
"data",
"=",
"json",
".",
"loads",
"(",
"handle",
".",
"read",
"(",
")",
".",
"decode",
"(",
"u'utf-8'",
")",
")",
"email",
"=",
"None",
"for",
"entry",
"in",
"data",
":",
"if",
"(",
"not",
"entry",
"[",
"u'verified'",
"]",
")",
":",
"continue",
"email",
"=",
"entry",
"[",
"u'email'",
"]",
"if",
"entry",
"[",
"u'primary'",
"]",
":",
"break",
"return",
"email"
] | get real email from github . | train | false |
17,588 | def _check_conditions_permissions(user, permissions, course_id, content):
def test(user, per, operator='or'):
if isinstance(per, basestring):
if (per in CONDITIONS):
return _check_condition(user, per, content)
return has_permission(user, per, course_id=course_id)
elif (isinstance(per, list) and (operator in ['and', 'or'])):
results = [test(user, x, operator='and') for x in per]
if (operator == 'or'):
return (True in results)
elif (operator == 'and'):
return (False not in results)
return test(user, permissions, operator='or')
| [
"def",
"_check_conditions_permissions",
"(",
"user",
",",
"permissions",
",",
"course_id",
",",
"content",
")",
":",
"def",
"test",
"(",
"user",
",",
"per",
",",
"operator",
"=",
"'or'",
")",
":",
"if",
"isinstance",
"(",
"per",
",",
"basestring",
")",
":",
"if",
"(",
"per",
"in",
"CONDITIONS",
")",
":",
"return",
"_check_condition",
"(",
"user",
",",
"per",
",",
"content",
")",
"return",
"has_permission",
"(",
"user",
",",
"per",
",",
"course_id",
"=",
"course_id",
")",
"elif",
"(",
"isinstance",
"(",
"per",
",",
"list",
")",
"and",
"(",
"operator",
"in",
"[",
"'and'",
",",
"'or'",
"]",
")",
")",
":",
"results",
"=",
"[",
"test",
"(",
"user",
",",
"x",
",",
"operator",
"=",
"'and'",
")",
"for",
"x",
"in",
"per",
"]",
"if",
"(",
"operator",
"==",
"'or'",
")",
":",
"return",
"(",
"True",
"in",
"results",
")",
"elif",
"(",
"operator",
"==",
"'and'",
")",
":",
"return",
"(",
"False",
"not",
"in",
"results",
")",
"return",
"test",
"(",
"user",
",",
"permissions",
",",
"operator",
"=",
"'or'",
")"
] | accepts a list of permissions and proceed if any of the permission is valid . | train | false |
17,589 | def protectResource(resource, config):
return DummyAuthResource(resource)
| [
"def",
"protectResource",
"(",
"resource",
",",
"config",
")",
":",
"return",
"DummyAuthResource",
"(",
"resource",
")"
] | dummy resource protector . | train | false |
17,590 | @register.simple_tag
def reviewer_list(review_request):
return humanize_list(([(group.display_name or group.name) for group in review_request.target_groups.all()] + [(user.get_full_name() or user.username) for user in review_request.target_people.all()]))
| [
"@",
"register",
".",
"simple_tag",
"def",
"reviewer_list",
"(",
"review_request",
")",
":",
"return",
"humanize_list",
"(",
"(",
"[",
"(",
"group",
".",
"display_name",
"or",
"group",
".",
"name",
")",
"for",
"group",
"in",
"review_request",
".",
"target_groups",
".",
"all",
"(",
")",
"]",
"+",
"[",
"(",
"user",
".",
"get_full_name",
"(",
")",
"or",
"user",
".",
"username",
")",
"for",
"user",
"in",
"review_request",
".",
"target_people",
".",
"all",
"(",
")",
"]",
")",
")"
] | returns a humanized list of target reviewers in a review request . | train | false |
17,591 | def sync_all(saltenv=None, refresh=True):
log.debug('Syncing all')
ret = {}
ret['beacons'] = sync_beacons(saltenv, False)
ret['modules'] = sync_modules(saltenv, False)
ret['states'] = sync_states(saltenv, False)
ret['sdb'] = sync_sdb(saltenv)
ret['grains'] = sync_grains(saltenv, False)
ret['renderers'] = sync_renderers(saltenv, False)
ret['returners'] = sync_returners(saltenv, False)
ret['output'] = sync_output(saltenv, False)
ret['utils'] = sync_utils(saltenv, False)
ret['log_handlers'] = sync_log_handlers(saltenv, False)
ret['proxymodules'] = sync_proxymodules(saltenv, False)
ret['engines'] = sync_engines(saltenv, False)
if (__opts__['file_client'] == 'local'):
ret['pillar'] = sync_pillar(saltenv, False)
if refresh:
refresh_modules()
refresh_pillar()
return ret
| [
"def",
"sync_all",
"(",
"saltenv",
"=",
"None",
",",
"refresh",
"=",
"True",
")",
":",
"log",
".",
"debug",
"(",
"'Syncing all'",
")",
"ret",
"=",
"{",
"}",
"ret",
"[",
"'beacons'",
"]",
"=",
"sync_beacons",
"(",
"saltenv",
",",
"False",
")",
"ret",
"[",
"'modules'",
"]",
"=",
"sync_modules",
"(",
"saltenv",
",",
"False",
")",
"ret",
"[",
"'states'",
"]",
"=",
"sync_states",
"(",
"saltenv",
",",
"False",
")",
"ret",
"[",
"'sdb'",
"]",
"=",
"sync_sdb",
"(",
"saltenv",
")",
"ret",
"[",
"'grains'",
"]",
"=",
"sync_grains",
"(",
"saltenv",
",",
"False",
")",
"ret",
"[",
"'renderers'",
"]",
"=",
"sync_renderers",
"(",
"saltenv",
",",
"False",
")",
"ret",
"[",
"'returners'",
"]",
"=",
"sync_returners",
"(",
"saltenv",
",",
"False",
")",
"ret",
"[",
"'output'",
"]",
"=",
"sync_output",
"(",
"saltenv",
",",
"False",
")",
"ret",
"[",
"'utils'",
"]",
"=",
"sync_utils",
"(",
"saltenv",
",",
"False",
")",
"ret",
"[",
"'log_handlers'",
"]",
"=",
"sync_log_handlers",
"(",
"saltenv",
",",
"False",
")",
"ret",
"[",
"'proxymodules'",
"]",
"=",
"sync_proxymodules",
"(",
"saltenv",
",",
"False",
")",
"ret",
"[",
"'engines'",
"]",
"=",
"sync_engines",
"(",
"saltenv",
",",
"False",
")",
"if",
"(",
"__opts__",
"[",
"'file_client'",
"]",
"==",
"'local'",
")",
":",
"ret",
"[",
"'pillar'",
"]",
"=",
"sync_pillar",
"(",
"saltenv",
",",
"False",
")",
"if",
"refresh",
":",
"refresh_modules",
"(",
")",
"refresh_pillar",
"(",
")",
"return",
"ret"
] | sync all custom types saltenv : base the fileserver environment from which to sync . | train | false |
17,593 | def additionalAssignment(MobileAllocation_presence=0, StartingTime_presence=0):
a = TpPd(pd=6)
b = MessageType(mesType=59)
c = ChannelDescription()
packet = ((a / b) / c)
if (MobileAllocation_presence is 1):
d = MobileAllocationHdr(ieiMA=114, eightBitMA=0)
packet = (packet / d)
if (StartingTime_presence is 1):
e = StartingTimeHdr(ieiST=124, eightBitST=0)
packet = (packet / e)
return packet
| [
"def",
"additionalAssignment",
"(",
"MobileAllocation_presence",
"=",
"0",
",",
"StartingTime_presence",
"=",
"0",
")",
":",
"a",
"=",
"TpPd",
"(",
"pd",
"=",
"6",
")",
"b",
"=",
"MessageType",
"(",
"mesType",
"=",
"59",
")",
"c",
"=",
"ChannelDescription",
"(",
")",
"packet",
"=",
"(",
"(",
"a",
"/",
"b",
")",
"/",
"c",
")",
"if",
"(",
"MobileAllocation_presence",
"is",
"1",
")",
":",
"d",
"=",
"MobileAllocationHdr",
"(",
"ieiMA",
"=",
"114",
",",
"eightBitMA",
"=",
"0",
")",
"packet",
"=",
"(",
"packet",
"/",
"d",
")",
"if",
"(",
"StartingTime_presence",
"is",
"1",
")",
":",
"e",
"=",
"StartingTimeHdr",
"(",
"ieiST",
"=",
"124",
",",
"eightBitST",
"=",
"0",
")",
"packet",
"=",
"(",
"packet",
"/",
"e",
")",
"return",
"packet"
] | additional assignment section 9 . | train | true |
17,594 | def parse_s3_url(url):
bucket = ''
path = ''
if url:
result = urlparse.urlparse(url)
bucket = result.netloc
path = result.path.strip('/')
return (bucket, path)
| [
"def",
"parse_s3_url",
"(",
"url",
")",
":",
"bucket",
"=",
"''",
"path",
"=",
"''",
"if",
"url",
":",
"result",
"=",
"urlparse",
".",
"urlparse",
"(",
"url",
")",
"bucket",
"=",
"result",
".",
"netloc",
"path",
"=",
"result",
".",
"path",
".",
"strip",
"(",
"'/'",
")",
"return",
"(",
"bucket",
",",
"path",
")"
] | parses s3 url . | train | true |
17,599 | def scans_for_fnames(fnames, keep4d=False, separate_sessions=False):
flist = None
if (not isinstance(fnames[0], list)):
if func_is_3d(fnames[0]):
fnames = [fnames]
if (separate_sessions or keep4d):
flist = np.zeros((len(fnames),), dtype=object)
for (i, f) in enumerate(fnames):
if separate_sessions:
if keep4d:
if isinstance(f, list):
flist[i] = np.array(f, dtype=object)
else:
flist[i] = np.array([f], dtype=object)
else:
flist[i] = scans_for_fname(f)
elif keep4d:
flist[i] = f
else:
scans = scans_for_fname(f)
if (flist is None):
flist = scans
else:
flist = np.concatenate((flist, scans))
return flist
| [
"def",
"scans_for_fnames",
"(",
"fnames",
",",
"keep4d",
"=",
"False",
",",
"separate_sessions",
"=",
"False",
")",
":",
"flist",
"=",
"None",
"if",
"(",
"not",
"isinstance",
"(",
"fnames",
"[",
"0",
"]",
",",
"list",
")",
")",
":",
"if",
"func_is_3d",
"(",
"fnames",
"[",
"0",
"]",
")",
":",
"fnames",
"=",
"[",
"fnames",
"]",
"if",
"(",
"separate_sessions",
"or",
"keep4d",
")",
":",
"flist",
"=",
"np",
".",
"zeros",
"(",
"(",
"len",
"(",
"fnames",
")",
",",
")",
",",
"dtype",
"=",
"object",
")",
"for",
"(",
"i",
",",
"f",
")",
"in",
"enumerate",
"(",
"fnames",
")",
":",
"if",
"separate_sessions",
":",
"if",
"keep4d",
":",
"if",
"isinstance",
"(",
"f",
",",
"list",
")",
":",
"flist",
"[",
"i",
"]",
"=",
"np",
".",
"array",
"(",
"f",
",",
"dtype",
"=",
"object",
")",
"else",
":",
"flist",
"[",
"i",
"]",
"=",
"np",
".",
"array",
"(",
"[",
"f",
"]",
",",
"dtype",
"=",
"object",
")",
"else",
":",
"flist",
"[",
"i",
"]",
"=",
"scans_for_fname",
"(",
"f",
")",
"elif",
"keep4d",
":",
"flist",
"[",
"i",
"]",
"=",
"f",
"else",
":",
"scans",
"=",
"scans_for_fname",
"(",
"f",
")",
"if",
"(",
"flist",
"is",
"None",
")",
":",
"flist",
"=",
"scans",
"else",
":",
"flist",
"=",
"np",
".",
"concatenate",
"(",
"(",
"flist",
",",
"scans",
")",
")",
"return",
"flist"
] | converts a list of files to a concatenated numpy array for each volume . | train | false |
17,600 | @staff_member_required
def admin_keywords_submit(request):
(keyword_ids, titles) = ([], [])
remove = punctuation.replace(u'-', u'')
for title in request.POST.get(u'text_keywords', u'').split(u','):
title = u''.join([c for c in title if (c not in remove)]).strip()
if title:
(kw, created) = Keyword.objects.get_or_create_iexact(title=title)
keyword_id = str(kw.id)
if (keyword_id not in keyword_ids):
keyword_ids.append(keyword_id)
titles.append(title)
return HttpResponse((u'%s|%s' % (u','.join(keyword_ids), u', '.join(titles))), content_type=u'text/plain')
| [
"@",
"staff_member_required",
"def",
"admin_keywords_submit",
"(",
"request",
")",
":",
"(",
"keyword_ids",
",",
"titles",
")",
"=",
"(",
"[",
"]",
",",
"[",
"]",
")",
"remove",
"=",
"punctuation",
".",
"replace",
"(",
"u'-'",
",",
"u''",
")",
"for",
"title",
"in",
"request",
".",
"POST",
".",
"get",
"(",
"u'text_keywords'",
",",
"u''",
")",
".",
"split",
"(",
"u','",
")",
":",
"title",
"=",
"u''",
".",
"join",
"(",
"[",
"c",
"for",
"c",
"in",
"title",
"if",
"(",
"c",
"not",
"in",
"remove",
")",
"]",
")",
".",
"strip",
"(",
")",
"if",
"title",
":",
"(",
"kw",
",",
"created",
")",
"=",
"Keyword",
".",
"objects",
".",
"get_or_create_iexact",
"(",
"title",
"=",
"title",
")",
"keyword_id",
"=",
"str",
"(",
"kw",
".",
"id",
")",
"if",
"(",
"keyword_id",
"not",
"in",
"keyword_ids",
")",
":",
"keyword_ids",
".",
"append",
"(",
"keyword_id",
")",
"titles",
".",
"append",
"(",
"title",
")",
"return",
"HttpResponse",
"(",
"(",
"u'%s|%s'",
"%",
"(",
"u','",
".",
"join",
"(",
"keyword_ids",
")",
",",
"u', '",
".",
"join",
"(",
"titles",
")",
")",
")",
",",
"content_type",
"=",
"u'text/plain'",
")"
] | adds any new given keywords from the custom keywords field in the admin . | train | false |
17,604 | def get_theano_hamiltonian_functions(model_vars, shared, logpt, potential, use_single_leapfrog=False, **theano_kwargs):
(H, q) = _theano_hamiltonian(model_vars, shared, logpt, potential)
(energy_function, p) = _theano_energy_function(H, q, **theano_kwargs)
if use_single_leapfrog:
leapfrog_integrator = _theano_single_leapfrog(H, q, p, **theano_kwargs)
else:
leapfrog_integrator = _theano_leapfrog_integrator(H, q, p, **theano_kwargs)
return (H, energy_function, leapfrog_integrator, {'q': q, 'p': p})
| [
"def",
"get_theano_hamiltonian_functions",
"(",
"model_vars",
",",
"shared",
",",
"logpt",
",",
"potential",
",",
"use_single_leapfrog",
"=",
"False",
",",
"**",
"theano_kwargs",
")",
":",
"(",
"H",
",",
"q",
")",
"=",
"_theano_hamiltonian",
"(",
"model_vars",
",",
"shared",
",",
"logpt",
",",
"potential",
")",
"(",
"energy_function",
",",
"p",
")",
"=",
"_theano_energy_function",
"(",
"H",
",",
"q",
",",
"**",
"theano_kwargs",
")",
"if",
"use_single_leapfrog",
":",
"leapfrog_integrator",
"=",
"_theano_single_leapfrog",
"(",
"H",
",",
"q",
",",
"p",
",",
"**",
"theano_kwargs",
")",
"else",
":",
"leapfrog_integrator",
"=",
"_theano_leapfrog_integrator",
"(",
"H",
",",
"q",
",",
"p",
",",
"**",
"theano_kwargs",
")",
"return",
"(",
"H",
",",
"energy_function",
",",
"leapfrog_integrator",
",",
"{",
"'q'",
":",
"q",
",",
"'p'",
":",
"p",
"}",
")"
] | construct theano functions for the hamiltonian . | train | false |
17,605 | def appendXmlTextNode(tag_name, text, parent):
el = xmlTextNode(tag_name, text)
parent.append(el)
return el
| [
"def",
"appendXmlTextNode",
"(",
"tag_name",
",",
"text",
",",
"parent",
")",
":",
"el",
"=",
"xmlTextNode",
"(",
"tag_name",
",",
"text",
")",
"parent",
".",
"append",
"(",
"el",
")",
"return",
"el"
] | creates a new <tag_name> node and sets its content to text . | train | false |
17,606 | def download_device(headers, cookies, temp_target_disk, device_url, lease_updater, total_bytes_written, total_bytes_to_write):
with open(temp_target_disk, 'wb') as handle:
response = requests.get(device_url, stream=True, headers=headers, cookies=cookies, verify=False)
if (not response.ok):
response.raise_for_status()
current_bytes_written = 0
for block in response.iter_content(chunk_size=2048):
if block:
handle.write(block)
handle.flush()
os.fsync(handle.fileno())
current_bytes_written += len(block)
written_pct = (((current_bytes_written + total_bytes_written) * 100) / total_bytes_to_write)
lease_updater.progressPercent = int(written_pct)
return current_bytes_written
| [
"def",
"download_device",
"(",
"headers",
",",
"cookies",
",",
"temp_target_disk",
",",
"device_url",
",",
"lease_updater",
",",
"total_bytes_written",
",",
"total_bytes_to_write",
")",
":",
"with",
"open",
"(",
"temp_target_disk",
",",
"'wb'",
")",
"as",
"handle",
":",
"response",
"=",
"requests",
".",
"get",
"(",
"device_url",
",",
"stream",
"=",
"True",
",",
"headers",
"=",
"headers",
",",
"cookies",
"=",
"cookies",
",",
"verify",
"=",
"False",
")",
"if",
"(",
"not",
"response",
".",
"ok",
")",
":",
"response",
".",
"raise_for_status",
"(",
")",
"current_bytes_written",
"=",
"0",
"for",
"block",
"in",
"response",
".",
"iter_content",
"(",
"chunk_size",
"=",
"2048",
")",
":",
"if",
"block",
":",
"handle",
".",
"write",
"(",
"block",
")",
"handle",
".",
"flush",
"(",
")",
"os",
".",
"fsync",
"(",
"handle",
".",
"fileno",
"(",
")",
")",
"current_bytes_written",
"+=",
"len",
"(",
"block",
")",
"written_pct",
"=",
"(",
"(",
"(",
"current_bytes_written",
"+",
"total_bytes_written",
")",
"*",
"100",
")",
"/",
"total_bytes_to_write",
")",
"lease_updater",
".",
"progressPercent",
"=",
"int",
"(",
"written_pct",
")",
"return",
"current_bytes_written"
] | download disk device of httpnfclease . | train | true |
17,607 | def create_benefit(benefit_class, **kwargs):
if (benefit_class.description is Benefit.description):
raise RuntimeError("Your custom benefit must implement its own 'description' property")
return Benefit.objects.create(proxy_class=_class_path(benefit_class), **kwargs)
| [
"def",
"create_benefit",
"(",
"benefit_class",
",",
"**",
"kwargs",
")",
":",
"if",
"(",
"benefit_class",
".",
"description",
"is",
"Benefit",
".",
"description",
")",
":",
"raise",
"RuntimeError",
"(",
"\"Your custom benefit must implement its own 'description' property\"",
")",
"return",
"Benefit",
".",
"objects",
".",
"create",
"(",
"proxy_class",
"=",
"_class_path",
"(",
"benefit_class",
")",
",",
"**",
"kwargs",
")"
] | create a custom benefit instance . | train | false |
17,609 | def is_key_line(key):
if (not key):
return False
if (key[0] == u'#'):
return False
if (key[0] == u'@'):
return False
return ((u' ssh-rsa ' in key) or (u' ecdsa-sha2-nistp256 ' in key) or (u' ssh-ed25519 ' in key))
| [
"def",
"is_key_line",
"(",
"key",
")",
":",
"if",
"(",
"not",
"key",
")",
":",
"return",
"False",
"if",
"(",
"key",
"[",
"0",
"]",
"==",
"u'#'",
")",
":",
"return",
"False",
"if",
"(",
"key",
"[",
"0",
"]",
"==",
"u'@'",
")",
":",
"return",
"False",
"return",
"(",
"(",
"u' ssh-rsa '",
"in",
"key",
")",
"or",
"(",
"u' ecdsa-sha2-nistp256 '",
"in",
"key",
")",
"or",
"(",
"u' ssh-ed25519 '",
"in",
"key",
")",
")"
] | checks whether this line looks like a valid known_hosts line . | train | false |
17,610 | def _float_or_none(x, digits=3):
if (x is None):
return str(x)
fmtstr = u'{0:.{digits}g}'.format(x, digits=digits)
return fmtstr.format(x)
| [
"def",
"_float_or_none",
"(",
"x",
",",
"digits",
"=",
"3",
")",
":",
"if",
"(",
"x",
"is",
"None",
")",
":",
"return",
"str",
"(",
"x",
")",
"fmtstr",
"=",
"u'{0:.{digits}g}'",
".",
"format",
"(",
"x",
",",
"digits",
"=",
"digits",
")",
"return",
"fmtstr",
".",
"format",
"(",
"x",
")"
] | helper function to format a variable that can be a float or none . | train | false |
17,611 | def deflood(s, n=3):
if (n == 0):
return s[0:0]
return re.sub(('((.)\\2{%s,})' % (n - 1)), (lambda m: (m.group(1)[0] * n)), s)
| [
"def",
"deflood",
"(",
"s",
",",
"n",
"=",
"3",
")",
":",
"if",
"(",
"n",
"==",
"0",
")",
":",
"return",
"s",
"[",
"0",
":",
"0",
"]",
"return",
"re",
".",
"sub",
"(",
"(",
"'((.)\\\\2{%s,})'",
"%",
"(",
"n",
"-",
"1",
")",
")",
",",
"(",
"lambda",
"m",
":",
"(",
"m",
".",
"group",
"(",
"1",
")",
"[",
"0",
"]",
"*",
"n",
")",
")",
",",
"s",
")"
] | returns the string with no more than n repeated characters . | train | true |
17,612 | def do_center(value, width=80):
return text_type(value).center(width)
| [
"def",
"do_center",
"(",
"value",
",",
"width",
"=",
"80",
")",
":",
"return",
"text_type",
"(",
"value",
")",
".",
"center",
"(",
"width",
")"
] | centers the value in a field of a given width . | train | false |
17,613 | def create_candlestick(open, high, low, close, dates=None, direction='both', **kwargs):
if (dates is not None):
utils.validate_equal_length(open, high, low, close, dates)
else:
utils.validate_equal_length(open, high, low, close)
validate_ohlc(open, high, low, close, direction, **kwargs)
if (direction is 'increasing'):
candle_incr_data = make_increasing_candle(open, high, low, close, dates, **kwargs)
data = candle_incr_data
elif (direction is 'decreasing'):
candle_decr_data = make_decreasing_candle(open, high, low, close, dates, **kwargs)
data = candle_decr_data
else:
candle_incr_data = make_increasing_candle(open, high, low, close, dates, **kwargs)
candle_decr_data = make_decreasing_candle(open, high, low, close, dates, **kwargs)
data = (candle_incr_data + candle_decr_data)
layout = graph_objs.Layout()
return graph_objs.Figure(data=data, layout=layout)
| [
"def",
"create_candlestick",
"(",
"open",
",",
"high",
",",
"low",
",",
"close",
",",
"dates",
"=",
"None",
",",
"direction",
"=",
"'both'",
",",
"**",
"kwargs",
")",
":",
"if",
"(",
"dates",
"is",
"not",
"None",
")",
":",
"utils",
".",
"validate_equal_length",
"(",
"open",
",",
"high",
",",
"low",
",",
"close",
",",
"dates",
")",
"else",
":",
"utils",
".",
"validate_equal_length",
"(",
"open",
",",
"high",
",",
"low",
",",
"close",
")",
"validate_ohlc",
"(",
"open",
",",
"high",
",",
"low",
",",
"close",
",",
"direction",
",",
"**",
"kwargs",
")",
"if",
"(",
"direction",
"is",
"'increasing'",
")",
":",
"candle_incr_data",
"=",
"make_increasing_candle",
"(",
"open",
",",
"high",
",",
"low",
",",
"close",
",",
"dates",
",",
"**",
"kwargs",
")",
"data",
"=",
"candle_incr_data",
"elif",
"(",
"direction",
"is",
"'decreasing'",
")",
":",
"candle_decr_data",
"=",
"make_decreasing_candle",
"(",
"open",
",",
"high",
",",
"low",
",",
"close",
",",
"dates",
",",
"**",
"kwargs",
")",
"data",
"=",
"candle_decr_data",
"else",
":",
"candle_incr_data",
"=",
"make_increasing_candle",
"(",
"open",
",",
"high",
",",
"low",
",",
"close",
",",
"dates",
",",
"**",
"kwargs",
")",
"candle_decr_data",
"=",
"make_decreasing_candle",
"(",
"open",
",",
"high",
",",
"low",
",",
"close",
",",
"dates",
",",
"**",
"kwargs",
")",
"data",
"=",
"(",
"candle_incr_data",
"+",
"candle_decr_data",
")",
"layout",
"=",
"graph_objs",
".",
"Layout",
"(",
")",
"return",
"graph_objs",
".",
"Figure",
"(",
"data",
"=",
"data",
",",
"layout",
"=",
"layout",
")"
] | beta function that creates a candlestick chart . | train | false |
17,614 | def _realpath(path):
if (sys.platform == 'cygwin'):
if (path[1:3] == ':\\'):
return path
elif (path[1:3] == ':/'):
path = (('/cygdrive/' + path[0]) + path[2:])
return os.path.abspath(os.path.expanduser(path))
return os.path.realpath(os.path.abspath(os.path.expanduser(path)))
| [
"def",
"_realpath",
"(",
"path",
")",
":",
"if",
"(",
"sys",
".",
"platform",
"==",
"'cygwin'",
")",
":",
"if",
"(",
"path",
"[",
"1",
":",
"3",
"]",
"==",
"':\\\\'",
")",
":",
"return",
"path",
"elif",
"(",
"path",
"[",
"1",
":",
"3",
"]",
"==",
"':/'",
")",
":",
"path",
"=",
"(",
"(",
"'/cygdrive/'",
"+",
"path",
"[",
"0",
"]",
")",
"+",
"path",
"[",
"2",
":",
"]",
")",
"return",
"os",
".",
"path",
".",
"abspath",
"(",
"os",
".",
"path",
".",
"expanduser",
"(",
"path",
")",
")",
"return",
"os",
".",
"path",
".",
"realpath",
"(",
"os",
".",
"path",
".",
"abspath",
"(",
"os",
".",
"path",
".",
"expanduser",
"(",
"path",
")",
")",
")"
] | return the real path of path is equivalent to realpath(abspath(expanduser)) . | train | true |
17,615 | def get_tests_stanza(tests, is_server, prepend=None, append=None, client_control_file=''):
assert (not (client_control_file and is_server))
if (not prepend):
prepend = []
if (not append):
append = []
raw_control_files = [read_control_file(test) for test in tests]
return _get_tests_stanza(raw_control_files, is_server, prepend, append, client_control_file=client_control_file)
| [
"def",
"get_tests_stanza",
"(",
"tests",
",",
"is_server",
",",
"prepend",
"=",
"None",
",",
"append",
"=",
"None",
",",
"client_control_file",
"=",
"''",
")",
":",
"assert",
"(",
"not",
"(",
"client_control_file",
"and",
"is_server",
")",
")",
"if",
"(",
"not",
"prepend",
")",
":",
"prepend",
"=",
"[",
"]",
"if",
"(",
"not",
"append",
")",
":",
"append",
"=",
"[",
"]",
"raw_control_files",
"=",
"[",
"read_control_file",
"(",
"test",
")",
"for",
"test",
"in",
"tests",
"]",
"return",
"_get_tests_stanza",
"(",
"raw_control_files",
",",
"is_server",
",",
"prepend",
",",
"append",
",",
"client_control_file",
"=",
"client_control_file",
")"
] | constructs the control file test step code from a list of tests . | train | false |
17,616 | def serialize_unregistered(fullname, email):
user = auth.get_user(email=email)
if (user is None):
serialized = {'fullname': fullname, 'id': None, 'registered': False, 'active': False, 'gravatar': gravatar(email, use_ssl=True, size=settings.PROFILE_IMAGE_MEDIUM), 'email': email}
else:
serialized = add_contributor_json(user)
serialized['fullname'] = fullname
serialized['email'] = email
return serialized
| [
"def",
"serialize_unregistered",
"(",
"fullname",
",",
"email",
")",
":",
"user",
"=",
"auth",
".",
"get_user",
"(",
"email",
"=",
"email",
")",
"if",
"(",
"user",
"is",
"None",
")",
":",
"serialized",
"=",
"{",
"'fullname'",
":",
"fullname",
",",
"'id'",
":",
"None",
",",
"'registered'",
":",
"False",
",",
"'active'",
":",
"False",
",",
"'gravatar'",
":",
"gravatar",
"(",
"email",
",",
"use_ssl",
"=",
"True",
",",
"size",
"=",
"settings",
".",
"PROFILE_IMAGE_MEDIUM",
")",
",",
"'email'",
":",
"email",
"}",
"else",
":",
"serialized",
"=",
"add_contributor_json",
"(",
"user",
")",
"serialized",
"[",
"'fullname'",
"]",
"=",
"fullname",
"serialized",
"[",
"'email'",
"]",
"=",
"email",
"return",
"serialized"
] | serializes an unregistered user . | train | false |
17,617 | def looks_like_a_cwl_artifact(path, classes=None):
if (not _has_extension(path, CWL_EXTENSIONS)):
return False
with open(path, 'r') as f:
try:
as_dict = yaml.safe_load(f)
except Exception:
return False
if (not isinstance(as_dict, dict)):
return False
file_class = as_dict.get('class', None)
if ((classes is not None) and (file_class not in classes)):
return False
file_cwl_version = as_dict.get('cwlVersion', None)
return (file_cwl_version is not None)
| [
"def",
"looks_like_a_cwl_artifact",
"(",
"path",
",",
"classes",
"=",
"None",
")",
":",
"if",
"(",
"not",
"_has_extension",
"(",
"path",
",",
"CWL_EXTENSIONS",
")",
")",
":",
"return",
"False",
"with",
"open",
"(",
"path",
",",
"'r'",
")",
"as",
"f",
":",
"try",
":",
"as_dict",
"=",
"yaml",
".",
"safe_load",
"(",
"f",
")",
"except",
"Exception",
":",
"return",
"False",
"if",
"(",
"not",
"isinstance",
"(",
"as_dict",
",",
"dict",
")",
")",
":",
"return",
"False",
"file_class",
"=",
"as_dict",
".",
"get",
"(",
"'class'",
",",
"None",
")",
"if",
"(",
"(",
"classes",
"is",
"not",
"None",
")",
"and",
"(",
"file_class",
"not",
"in",
"classes",
")",
")",
":",
"return",
"False",
"file_cwl_version",
"=",
"as_dict",
".",
"get",
"(",
"'cwlVersion'",
",",
"None",
")",
"return",
"(",
"file_cwl_version",
"is",
"not",
"None",
")"
] | quick check to see if a file looks like it may be a cwl artifact . | train | false |
17,618 | def _AddPropertiesForNonRepeatedCompositeField(field, cls):
proto_field_name = field.name
property_name = _PropertyName(proto_field_name)
message_type = field.message_type
def getter(self):
field_value = self._fields.get(field)
if (field_value is None):
field_value = message_type._concrete_class()
field_value._SetListener(self._listener_for_children)
field_value = self._fields.setdefault(field, field_value)
return field_value
getter.__module__ = None
getter.__doc__ = ('Getter for %s.' % proto_field_name)
def setter(self, new_value):
raise AttributeError(('Assignment not allowed to composite field "%s" in protocol message object.' % proto_field_name))
doc = ('Magic attribute generated for "%s" proto field.' % proto_field_name)
setattr(cls, property_name, property(getter, setter, doc=doc))
| [
"def",
"_AddPropertiesForNonRepeatedCompositeField",
"(",
"field",
",",
"cls",
")",
":",
"proto_field_name",
"=",
"field",
".",
"name",
"property_name",
"=",
"_PropertyName",
"(",
"proto_field_name",
")",
"message_type",
"=",
"field",
".",
"message_type",
"def",
"getter",
"(",
"self",
")",
":",
"field_value",
"=",
"self",
".",
"_fields",
".",
"get",
"(",
"field",
")",
"if",
"(",
"field_value",
"is",
"None",
")",
":",
"field_value",
"=",
"message_type",
".",
"_concrete_class",
"(",
")",
"field_value",
".",
"_SetListener",
"(",
"self",
".",
"_listener_for_children",
")",
"field_value",
"=",
"self",
".",
"_fields",
".",
"setdefault",
"(",
"field",
",",
"field_value",
")",
"return",
"field_value",
"getter",
".",
"__module__",
"=",
"None",
"getter",
".",
"__doc__",
"=",
"(",
"'Getter for %s.'",
"%",
"proto_field_name",
")",
"def",
"setter",
"(",
"self",
",",
"new_value",
")",
":",
"raise",
"AttributeError",
"(",
"(",
"'Assignment not allowed to composite field \"%s\" in protocol message object.'",
"%",
"proto_field_name",
")",
")",
"doc",
"=",
"(",
"'Magic attribute generated for \"%s\" proto field.'",
"%",
"proto_field_name",
")",
"setattr",
"(",
"cls",
",",
"property_name",
",",
"property",
"(",
"getter",
",",
"setter",
",",
"doc",
"=",
"doc",
")",
")"
] | adds a public property for a nonrepeated . | train | false |
17,620 | def _get_level(value, levels, prefix=None):
if ((value > 1) or (value < 0)):
raise ValueError(('Encountered invalid normalized alpha diversity value %s. Normalized values must be between 0 and 1.' % value))
check = [i for i in range(0, len(levels)) if (levels[i] == value)]
if len(check):
value_level = (check[0] + 2)
else:
value_level = (searchsorted(levels, value) + 1)
if (prefix is not None):
output = '{0}_{1}_of_{2}'.format(prefix, value_level, (len(levels) + 1))
else:
output = value_level
return output
| [
"def",
"_get_level",
"(",
"value",
",",
"levels",
",",
"prefix",
"=",
"None",
")",
":",
"if",
"(",
"(",
"value",
">",
"1",
")",
"or",
"(",
"value",
"<",
"0",
")",
")",
":",
"raise",
"ValueError",
"(",
"(",
"'Encountered invalid normalized alpha diversity value %s. Normalized values must be between 0 and 1.'",
"%",
"value",
")",
")",
"check",
"=",
"[",
"i",
"for",
"i",
"in",
"range",
"(",
"0",
",",
"len",
"(",
"levels",
")",
")",
"if",
"(",
"levels",
"[",
"i",
"]",
"==",
"value",
")",
"]",
"if",
"len",
"(",
"check",
")",
":",
"value_level",
"=",
"(",
"check",
"[",
"0",
"]",
"+",
"2",
")",
"else",
":",
"value_level",
"=",
"(",
"searchsorted",
"(",
"levels",
",",
"value",
")",
"+",
"1",
")",
"if",
"(",
"prefix",
"is",
"not",
"None",
")",
":",
"output",
"=",
"'{0}_{1}_of_{2}'",
".",
"format",
"(",
"prefix",
",",
"value_level",
",",
"(",
"len",
"(",
"levels",
")",
"+",
"1",
")",
")",
"else",
":",
"output",
"=",
"value_level",
"return",
"output"
] | accommodate a value into the levels list; return a string or an integer input: value: normalized value to assign a level to . | train | false |
17,621 | def _prevent_segment_delete_with_port_bound(resource, event, trigger, context, segment):
segment_id = segment['id']
query = context.session.query(models_v2.Port)
query = query.join(models.PortBindingLevel, (models.PortBindingLevel.port_id == models_v2.Port.id))
query = query.filter((models.PortBindingLevel.segment_id == segment_id))
port_ids = [p.id for p in query]
if port_ids:
reason = (_('The segment is still bound with port(s) %s') % ', '.join(port_ids))
raise seg_exc.SegmentInUse(segment_id=segment_id, reason=reason)
| [
"def",
"_prevent_segment_delete_with_port_bound",
"(",
"resource",
",",
"event",
",",
"trigger",
",",
"context",
",",
"segment",
")",
":",
"segment_id",
"=",
"segment",
"[",
"'id'",
"]",
"query",
"=",
"context",
".",
"session",
".",
"query",
"(",
"models_v2",
".",
"Port",
")",
"query",
"=",
"query",
".",
"join",
"(",
"models",
".",
"PortBindingLevel",
",",
"(",
"models",
".",
"PortBindingLevel",
".",
"port_id",
"==",
"models_v2",
".",
"Port",
".",
"id",
")",
")",
"query",
"=",
"query",
".",
"filter",
"(",
"(",
"models",
".",
"PortBindingLevel",
".",
"segment_id",
"==",
"segment_id",
")",
")",
"port_ids",
"=",
"[",
"p",
".",
"id",
"for",
"p",
"in",
"query",
"]",
"if",
"port_ids",
":",
"reason",
"=",
"(",
"_",
"(",
"'The segment is still bound with port(s) %s'",
")",
"%",
"', '",
".",
"join",
"(",
"port_ids",
")",
")",
"raise",
"seg_exc",
".",
"SegmentInUse",
"(",
"segment_id",
"=",
"segment_id",
",",
"reason",
"=",
"reason",
")"
] | raise exception if there are any ports bound with segment_id . | train | false |
17,622 | def generate_subs_from_source(speed_subs, subs_type, subs_filedata, item, language='en'):
_ = item.runtime.service(item, 'i18n').ugettext
if (subs_type.lower() != 'srt'):
raise TranscriptsGenerationException(_('We support only SubRip (*.srt) transcripts format.'))
try:
srt_subs_obj = SubRipFile.from_string(subs_filedata)
except Exception as ex:
msg = _('Something wrong with SubRip transcripts file during parsing. Inner message is {error_message}').format(error_message=ex.message)
raise TranscriptsGenerationException(msg)
if (not srt_subs_obj):
raise TranscriptsGenerationException(_('Something wrong with SubRip transcripts file during parsing.'))
sub_starts = []
sub_ends = []
sub_texts = []
for sub in srt_subs_obj:
sub_starts.append(sub.start.ordinal)
sub_ends.append(sub.end.ordinal)
sub_texts.append(sub.text.replace('\n', ' '))
subs = {'start': sub_starts, 'end': sub_ends, 'text': sub_texts}
for (speed, subs_id) in speed_subs.iteritems():
save_subs_to_store(generate_subs(speed, 1, subs), subs_id, item, language)
return subs
| [
"def",
"generate_subs_from_source",
"(",
"speed_subs",
",",
"subs_type",
",",
"subs_filedata",
",",
"item",
",",
"language",
"=",
"'en'",
")",
":",
"_",
"=",
"item",
".",
"runtime",
".",
"service",
"(",
"item",
",",
"'i18n'",
")",
".",
"ugettext",
"if",
"(",
"subs_type",
".",
"lower",
"(",
")",
"!=",
"'srt'",
")",
":",
"raise",
"TranscriptsGenerationException",
"(",
"_",
"(",
"'We support only SubRip (*.srt) transcripts format.'",
")",
")",
"try",
":",
"srt_subs_obj",
"=",
"SubRipFile",
".",
"from_string",
"(",
"subs_filedata",
")",
"except",
"Exception",
"as",
"ex",
":",
"msg",
"=",
"_",
"(",
"'Something wrong with SubRip transcripts file during parsing. Inner message is {error_message}'",
")",
".",
"format",
"(",
"error_message",
"=",
"ex",
".",
"message",
")",
"raise",
"TranscriptsGenerationException",
"(",
"msg",
")",
"if",
"(",
"not",
"srt_subs_obj",
")",
":",
"raise",
"TranscriptsGenerationException",
"(",
"_",
"(",
"'Something wrong with SubRip transcripts file during parsing.'",
")",
")",
"sub_starts",
"=",
"[",
"]",
"sub_ends",
"=",
"[",
"]",
"sub_texts",
"=",
"[",
"]",
"for",
"sub",
"in",
"srt_subs_obj",
":",
"sub_starts",
".",
"append",
"(",
"sub",
".",
"start",
".",
"ordinal",
")",
"sub_ends",
".",
"append",
"(",
"sub",
".",
"end",
".",
"ordinal",
")",
"sub_texts",
".",
"append",
"(",
"sub",
".",
"text",
".",
"replace",
"(",
"'\\n'",
",",
"' '",
")",
")",
"subs",
"=",
"{",
"'start'",
":",
"sub_starts",
",",
"'end'",
":",
"sub_ends",
",",
"'text'",
":",
"sub_texts",
"}",
"for",
"(",
"speed",
",",
"subs_id",
")",
"in",
"speed_subs",
".",
"iteritems",
"(",
")",
":",
"save_subs_to_store",
"(",
"generate_subs",
"(",
"speed",
",",
"1",
",",
"subs",
")",
",",
"subs_id",
",",
"item",
",",
"language",
")",
"return",
"subs"
] | generate transcripts from source files and save them to assets for item module . | train | false |
17,623 | def get_perm(Model, perm):
ct = ContentType.objects.get_for_model(Model)
return Permission.objects.get(content_type=ct, codename=perm)
| [
"def",
"get_perm",
"(",
"Model",
",",
"perm",
")",
":",
"ct",
"=",
"ContentType",
".",
"objects",
".",
"get_for_model",
"(",
"Model",
")",
"return",
"Permission",
".",
"objects",
".",
"get",
"(",
"content_type",
"=",
"ct",
",",
"codename",
"=",
"perm",
")"
] | return the permission object . | train | false |
17,624 | def _degree_bound_univariate(f, g):
gamma = f.ring.domain.gcd(f.LC, g.LC)
p = 1
p = nextprime(p)
while ((gamma % p) == 0):
p = nextprime(p)
fp = f.trunc_ground(p)
gp = g.trunc_ground(p)
hp = _gf_gcd(fp, gp, p)
deghp = hp.degree()
return deghp
| [
"def",
"_degree_bound_univariate",
"(",
"f",
",",
"g",
")",
":",
"gamma",
"=",
"f",
".",
"ring",
".",
"domain",
".",
"gcd",
"(",
"f",
".",
"LC",
",",
"g",
".",
"LC",
")",
"p",
"=",
"1",
"p",
"=",
"nextprime",
"(",
"p",
")",
"while",
"(",
"(",
"gamma",
"%",
"p",
")",
"==",
"0",
")",
":",
"p",
"=",
"nextprime",
"(",
"p",
")",
"fp",
"=",
"f",
".",
"trunc_ground",
"(",
"p",
")",
"gp",
"=",
"g",
".",
"trunc_ground",
"(",
"p",
")",
"hp",
"=",
"_gf_gcd",
"(",
"fp",
",",
"gp",
",",
"p",
")",
"deghp",
"=",
"hp",
".",
"degree",
"(",
")",
"return",
"deghp"
] | compute an upper bound for the degree of the gcd of two univariate integer polynomials f and g . | train | false |
17,626 | @receiver(score_reset)
def submissions_score_reset_handler(sender, **kwargs):
course_id = kwargs['course_id']
usage_id = kwargs['item_id']
user = user_by_anonymous_id(kwargs['anonymous_user_id'])
if (user is None):
return
PROBLEM_WEIGHTED_SCORE_CHANGED.send(sender=None, weighted_earned=0, weighted_possible=0, user_id=user.id, anonymous_user_id=kwargs['anonymous_user_id'], course_id=course_id, usage_id=usage_id, modified=kwargs['created_at'], score_db_table=ScoreDatabaseTableEnum.submissions)
| [
"@",
"receiver",
"(",
"score_reset",
")",
"def",
"submissions_score_reset_handler",
"(",
"sender",
",",
"**",
"kwargs",
")",
":",
"course_id",
"=",
"kwargs",
"[",
"'course_id'",
"]",
"usage_id",
"=",
"kwargs",
"[",
"'item_id'",
"]",
"user",
"=",
"user_by_anonymous_id",
"(",
"kwargs",
"[",
"'anonymous_user_id'",
"]",
")",
"if",
"(",
"user",
"is",
"None",
")",
":",
"return",
"PROBLEM_WEIGHTED_SCORE_CHANGED",
".",
"send",
"(",
"sender",
"=",
"None",
",",
"weighted_earned",
"=",
"0",
",",
"weighted_possible",
"=",
"0",
",",
"user_id",
"=",
"user",
".",
"id",
",",
"anonymous_user_id",
"=",
"kwargs",
"[",
"'anonymous_user_id'",
"]",
",",
"course_id",
"=",
"course_id",
",",
"usage_id",
"=",
"usage_id",
",",
"modified",
"=",
"kwargs",
"[",
"'created_at'",
"]",
",",
"score_db_table",
"=",
"ScoreDatabaseTableEnum",
".",
"submissions",
")"
] | consume the score_reset signal defined in the submissions api . | train | false |
17,627 | @profiler.trace
def policy_create(request, **kwargs):
body = {'firewall_policy': kwargs}
policy = neutronclient(request).create_firewall_policy(body).get('firewall_policy')
return Policy(policy)
| [
"@",
"profiler",
".",
"trace",
"def",
"policy_create",
"(",
"request",
",",
"**",
"kwargs",
")",
":",
"body",
"=",
"{",
"'firewall_policy'",
":",
"kwargs",
"}",
"policy",
"=",
"neutronclient",
"(",
"request",
")",
".",
"create_firewall_policy",
"(",
"body",
")",
".",
"get",
"(",
"'firewall_policy'",
")",
"return",
"Policy",
"(",
"policy",
")"
] | create a firewall policy . | train | false |
17,629 | def helmert(n, full=False):
H = (np.tril(np.ones((n, n)), (-1)) - np.diag(np.arange(n)))
d = (np.arange(n) * np.arange(1, (n + 1)))
H[0] = 1
d[0] = n
H_full = (H / np.sqrt(d)[:, np.newaxis])
if full:
return H_full
else:
return H_full[1:]
| [
"def",
"helmert",
"(",
"n",
",",
"full",
"=",
"False",
")",
":",
"H",
"=",
"(",
"np",
".",
"tril",
"(",
"np",
".",
"ones",
"(",
"(",
"n",
",",
"n",
")",
")",
",",
"(",
"-",
"1",
")",
")",
"-",
"np",
".",
"diag",
"(",
"np",
".",
"arange",
"(",
"n",
")",
")",
")",
"d",
"=",
"(",
"np",
".",
"arange",
"(",
"n",
")",
"*",
"np",
".",
"arange",
"(",
"1",
",",
"(",
"n",
"+",
"1",
")",
")",
")",
"H",
"[",
"0",
"]",
"=",
"1",
"d",
"[",
"0",
"]",
"=",
"n",
"H_full",
"=",
"(",
"H",
"/",
"np",
".",
"sqrt",
"(",
"d",
")",
"[",
":",
",",
"np",
".",
"newaxis",
"]",
")",
"if",
"full",
":",
"return",
"H_full",
"else",
":",
"return",
"H_full",
"[",
"1",
":",
"]"
] | create a helmert matrix of order n . | train | false |
17,630 | def FillUsersInQuery(filters):
for filter in filters:
for property in filter.property_list():
FillUser(property)
| [
"def",
"FillUsersInQuery",
"(",
"filters",
")",
":",
"for",
"filter",
"in",
"filters",
":",
"for",
"property",
"in",
"filter",
".",
"property_list",
"(",
")",
":",
"FillUser",
"(",
"property",
")"
] | fill in a synthetic user id for all user properties in a set of filters . | train | false |
17,634 | @utils.arg('server', metavar='<server>', help=_('Name or ID of server.'))
@utils.arg('name', metavar='<name>', help=_('Name of the backup image.'))
@utils.arg('backup_type', metavar='<backup-type>', help=_('The backup type, like "daily" or "weekly".'))
@utils.arg('rotation', metavar='<rotation>', help=_('Int parameter representing how many backups to keep around.'))
def do_backup(cs, args):
_find_server(cs, args.server).backup(args.name, args.backup_type, args.rotation)
| [
"@",
"utils",
".",
"arg",
"(",
"'server'",
",",
"metavar",
"=",
"'<server>'",
",",
"help",
"=",
"_",
"(",
"'Name or ID of server.'",
")",
")",
"@",
"utils",
".",
"arg",
"(",
"'name'",
",",
"metavar",
"=",
"'<name>'",
",",
"help",
"=",
"_",
"(",
"'Name of the backup image.'",
")",
")",
"@",
"utils",
".",
"arg",
"(",
"'backup_type'",
",",
"metavar",
"=",
"'<backup-type>'",
",",
"help",
"=",
"_",
"(",
"'The backup type, like \"daily\" or \"weekly\".'",
")",
")",
"@",
"utils",
".",
"arg",
"(",
"'rotation'",
",",
"metavar",
"=",
"'<rotation>'",
",",
"help",
"=",
"_",
"(",
"'Int parameter representing how many backups to keep around.'",
")",
")",
"def",
"do_backup",
"(",
"cs",
",",
"args",
")",
":",
"_find_server",
"(",
"cs",
",",
"args",
".",
"server",
")",
".",
"backup",
"(",
"args",
".",
"name",
",",
"args",
".",
"backup_type",
",",
"args",
".",
"rotation",
")"
] | backup a server by creating a backup type snapshot . | train | false |
17,636 | def extended_gcd(p, q):
(a, b) = (p, q)
if (a < 0):
a = ((-1) * a)
if (b < 0):
b = ((-1) * b)
x0 = 0
y1 = 0
x1 = 1
y0 = 1
while (b != 0):
quotient = (a // b)
(a, b) = (b, (a % b))
(x1, x0) = ((x0 - (quotient * x1)), x1)
(y1, y0) = ((y0 - (quotient * y1)), y1)
if (p < 0):
y0 = ((-1) * y0)
if (q < 0):
x0 = ((-1) * x0)
return (y0, x0)
| [
"def",
"extended_gcd",
"(",
"p",
",",
"q",
")",
":",
"(",
"a",
",",
"b",
")",
"=",
"(",
"p",
",",
"q",
")",
"if",
"(",
"a",
"<",
"0",
")",
":",
"a",
"=",
"(",
"(",
"-",
"1",
")",
"*",
"a",
")",
"if",
"(",
"b",
"<",
"0",
")",
":",
"b",
"=",
"(",
"(",
"-",
"1",
")",
"*",
"b",
")",
"x0",
"=",
"0",
"y1",
"=",
"0",
"x1",
"=",
"1",
"y0",
"=",
"1",
"while",
"(",
"b",
"!=",
"0",
")",
":",
"quotient",
"=",
"(",
"a",
"//",
"b",
")",
"(",
"a",
",",
"b",
")",
"=",
"(",
"b",
",",
"(",
"a",
"%",
"b",
")",
")",
"(",
"x1",
",",
"x0",
")",
"=",
"(",
"(",
"x0",
"-",
"(",
"quotient",
"*",
"x1",
")",
")",
",",
"x1",
")",
"(",
"y1",
",",
"y0",
")",
"=",
"(",
"(",
"y0",
"-",
"(",
"quotient",
"*",
"y1",
")",
")",
",",
"y1",
")",
"if",
"(",
"p",
"<",
"0",
")",
":",
"y0",
"=",
"(",
"(",
"-",
"1",
")",
"*",
"y0",
")",
"if",
"(",
"q",
"<",
"0",
")",
":",
"x0",
"=",
"(",
"(",
"-",
"1",
")",
"*",
"x0",
")",
"return",
"(",
"y0",
",",
"x0",
")"
] | returns a tuple such that r = gcd = ia + jb . | train | false |
17,637 | def _gpi10iterator(handle):
for inline in handle:
if (inline[0] == '!'):
continue
inrec = inline.rstrip('\n').split(' DCTB ')
if (len(inrec) == 1):
continue
inrec[5] = inrec[5].split('|')
inrec[8] = inrec[8].split('|')
(yield dict(zip(GPI10FIELDS, inrec)))
| [
"def",
"_gpi10iterator",
"(",
"handle",
")",
":",
"for",
"inline",
"in",
"handle",
":",
"if",
"(",
"inline",
"[",
"0",
"]",
"==",
"'!'",
")",
":",
"continue",
"inrec",
"=",
"inline",
".",
"rstrip",
"(",
"'\\n'",
")",
".",
"split",
"(",
"' DCTB '",
")",
"if",
"(",
"len",
"(",
"inrec",
")",
"==",
"1",
")",
":",
"continue",
"inrec",
"[",
"5",
"]",
"=",
"inrec",
"[",
"5",
"]",
".",
"split",
"(",
"'|'",
")",
"inrec",
"[",
"8",
"]",
"=",
"inrec",
"[",
"8",
"]",
".",
"split",
"(",
"'|'",
")",
"(",
"yield",
"dict",
"(",
"zip",
"(",
"GPI10FIELDS",
",",
"inrec",
")",
")",
")"
] | read gpi 1 . | train | false |
17,638 | def assert_aws_environ():
try:
import boto
except ImportError as e:
raise SkipTest(str(e))
if ('AWS_ACCESS_KEY_ID' not in os.environ):
raise SkipTest('AWS keys not found')
| [
"def",
"assert_aws_environ",
"(",
")",
":",
"try",
":",
"import",
"boto",
"except",
"ImportError",
"as",
"e",
":",
"raise",
"SkipTest",
"(",
"str",
"(",
"e",
")",
")",
"if",
"(",
"'AWS_ACCESS_KEY_ID'",
"not",
"in",
"os",
".",
"environ",
")",
":",
"raise",
"SkipTest",
"(",
"'AWS keys not found'",
")"
] | asserts the current environment is suitable for running aws testsi . | train | false |
17,639 | @permission_required('questions.tag_question')
@require_POST
def remove_tag_async(request, question_id):
name = request.POST.get('name')
if name:
question = get_object_or_404(Question, pk=question_id)
question.tags.remove(name)
question.clear_cached_tags()
return HttpResponse('{}', content_type='application/json')
return HttpResponseBadRequest(json.dumps({'error': unicode(NO_TAG)}), content_type='application/json')
| [
"@",
"permission_required",
"(",
"'questions.tag_question'",
")",
"@",
"require_POST",
"def",
"remove_tag_async",
"(",
"request",
",",
"question_id",
")",
":",
"name",
"=",
"request",
".",
"POST",
".",
"get",
"(",
"'name'",
")",
"if",
"name",
":",
"question",
"=",
"get_object_or_404",
"(",
"Question",
",",
"pk",
"=",
"question_id",
")",
"question",
".",
"tags",
".",
"remove",
"(",
"name",
")",
"question",
".",
"clear_cached_tags",
"(",
")",
"return",
"HttpResponse",
"(",
"'{}'",
",",
"content_type",
"=",
"'application/json'",
")",
"return",
"HttpResponseBadRequest",
"(",
"json",
".",
"dumps",
"(",
"{",
"'error'",
":",
"unicode",
"(",
"NO_TAG",
")",
"}",
")",
",",
"content_type",
"=",
"'application/json'",
")"
] | remove a tag from question . | train | false |
17,640 | def _deactivate_invalidation(certificate):
try:
certificate_invalidation = CertificateInvalidation.objects.get(generated_certificate=certificate, active=True)
certificate_invalidation.deactivate()
except CertificateInvalidation.DoesNotExist:
pass
| [
"def",
"_deactivate_invalidation",
"(",
"certificate",
")",
":",
"try",
":",
"certificate_invalidation",
"=",
"CertificateInvalidation",
".",
"objects",
".",
"get",
"(",
"generated_certificate",
"=",
"certificate",
",",
"active",
"=",
"True",
")",
"certificate_invalidation",
".",
"deactivate",
"(",
")",
"except",
"CertificateInvalidation",
".",
"DoesNotExist",
":",
"pass"
] | deactivate certificate invalidation by setting active to false . | train | false |
17,641 | @memoize_default(None, evaluator_is_first_arg=True)
def follow_param(evaluator, param):
def eval_docstring(docstr):
param_str = str(param.name)
return set([p for string in _search_param_in_docstr(docstr, param_str) for p in _evaluate_for_statement_string(evaluator, string, module)])
func = param.parent_function
module = param.get_parent_until()
docstr = func.raw_doc
types = eval_docstring(docstr)
if (func.name.value == '__init__'):
cls = func.get_parent_until(Class)
if (cls.type == 'classdef'):
types |= eval_docstring(cls.raw_doc)
return types
| [
"@",
"memoize_default",
"(",
"None",
",",
"evaluator_is_first_arg",
"=",
"True",
")",
"def",
"follow_param",
"(",
"evaluator",
",",
"param",
")",
":",
"def",
"eval_docstring",
"(",
"docstr",
")",
":",
"param_str",
"=",
"str",
"(",
"param",
".",
"name",
")",
"return",
"set",
"(",
"[",
"p",
"for",
"string",
"in",
"_search_param_in_docstr",
"(",
"docstr",
",",
"param_str",
")",
"for",
"p",
"in",
"_evaluate_for_statement_string",
"(",
"evaluator",
",",
"string",
",",
"module",
")",
"]",
")",
"func",
"=",
"param",
".",
"parent_function",
"module",
"=",
"param",
".",
"get_parent_until",
"(",
")",
"docstr",
"=",
"func",
".",
"raw_doc",
"types",
"=",
"eval_docstring",
"(",
"docstr",
")",
"if",
"(",
"func",
".",
"name",
".",
"value",
"==",
"'__init__'",
")",
":",
"cls",
"=",
"func",
".",
"get_parent_until",
"(",
"Class",
")",
"if",
"(",
"cls",
".",
"type",
"==",
"'classdef'",
")",
":",
"types",
"|=",
"eval_docstring",
"(",
"cls",
".",
"raw_doc",
")",
"return",
"types"
] | determines a set of potential types for param using docstring hints :type evaluator: jedi . | train | false |
17,642 | def setup_logrotate(app_name, watch, log_size):
app_logrotate_script = '{0}/appscale-{1}'.format(LOGROTATE_CONFIG_DIR, app_name)
contents = '/var/log/appscale/{watch}*.log {{\n size {size}\n missingok\n rotate 7\n compress\n delaycompress\n notifempty\n copytruncate\n}}\n'.format(watch=watch, size=log_size)
logging.debug('Logrotate file: {} - Contents:\n{}'.format(app_logrotate_script, contents))
with open(app_logrotate_script, 'w') as app_logrotate_fd:
app_logrotate_fd.write(contents)
return True
| [
"def",
"setup_logrotate",
"(",
"app_name",
",",
"watch",
",",
"log_size",
")",
":",
"app_logrotate_script",
"=",
"'{0}/appscale-{1}'",
".",
"format",
"(",
"LOGROTATE_CONFIG_DIR",
",",
"app_name",
")",
"contents",
"=",
"'/var/log/appscale/{watch}*.log {{\\n size {size}\\n missingok\\n rotate 7\\n compress\\n delaycompress\\n notifempty\\n copytruncate\\n}}\\n'",
".",
"format",
"(",
"watch",
"=",
"watch",
",",
"size",
"=",
"log_size",
")",
"logging",
".",
"debug",
"(",
"'Logrotate file: {} - Contents:\\n{}'",
".",
"format",
"(",
"app_logrotate_script",
",",
"contents",
")",
")",
"with",
"open",
"(",
"app_logrotate_script",
",",
"'w'",
")",
"as",
"app_logrotate_fd",
":",
"app_logrotate_fd",
".",
"write",
"(",
"contents",
")",
"return",
"True"
] | creates a logrotate script for the logs that the given application will create . | train | false |
17,643 | def render_string_template(template_string, context=None, context_instance=None):
if (context is None):
context = {}
template = Template(template_string)
if (('user' not in context) and context_instance):
if ('request' in context_instance):
context.update({'user': context_instance['request']})
return template.render(context)
| [
"def",
"render_string_template",
"(",
"template_string",
",",
"context",
"=",
"None",
",",
"context_instance",
"=",
"None",
")",
":",
"if",
"(",
"context",
"is",
"None",
")",
":",
"context",
"=",
"{",
"}",
"template",
"=",
"Template",
"(",
"template_string",
")",
"if",
"(",
"(",
"'user'",
"not",
"in",
"context",
")",
"and",
"context_instance",
")",
":",
"if",
"(",
"'request'",
"in",
"context_instance",
")",
":",
"context",
".",
"update",
"(",
"{",
"'user'",
":",
"context_instance",
"[",
"'request'",
"]",
"}",
")",
"return",
"template",
".",
"render",
"(",
"context",
")"
] | performs rendering using template_string instead of a file . | train | false |
17,644 | def mul_sign(lh_expr, rh_expr):
is_pos = ((lh_expr.is_zero() or rh_expr.is_zero()) or (lh_expr.is_positive() and rh_expr.is_positive()) or (lh_expr.is_negative() and rh_expr.is_negative()))
is_neg = ((lh_expr.is_zero() or rh_expr.is_zero()) or (lh_expr.is_positive() and rh_expr.is_negative()) or (lh_expr.is_negative() and rh_expr.is_positive()))
return (is_pos, is_neg)
| [
"def",
"mul_sign",
"(",
"lh_expr",
",",
"rh_expr",
")",
":",
"is_pos",
"=",
"(",
"(",
"lh_expr",
".",
"is_zero",
"(",
")",
"or",
"rh_expr",
".",
"is_zero",
"(",
")",
")",
"or",
"(",
"lh_expr",
".",
"is_positive",
"(",
")",
"and",
"rh_expr",
".",
"is_positive",
"(",
")",
")",
"or",
"(",
"lh_expr",
".",
"is_negative",
"(",
")",
"and",
"rh_expr",
".",
"is_negative",
"(",
")",
")",
")",
"is_neg",
"=",
"(",
"(",
"lh_expr",
".",
"is_zero",
"(",
")",
"or",
"rh_expr",
".",
"is_zero",
"(",
")",
")",
"or",
"(",
"lh_expr",
".",
"is_positive",
"(",
")",
"and",
"rh_expr",
".",
"is_negative",
"(",
")",
")",
"or",
"(",
"lh_expr",
".",
"is_negative",
"(",
")",
"and",
"rh_expr",
".",
"is_positive",
"(",
")",
")",
")",
"return",
"(",
"is_pos",
",",
"is_neg",
")"
] | give the sign resulting from multiplying two expressions . | train | false |
17,646 | def clean_text_by_sentences(text):
original_sentences = split_sentences(text)
filtered_sentences = [join_words(sentence) for sentence in preprocess_documents(original_sentences)]
return merge_syntactic_units(original_sentences, filtered_sentences)
| [
"def",
"clean_text_by_sentences",
"(",
"text",
")",
":",
"original_sentences",
"=",
"split_sentences",
"(",
"text",
")",
"filtered_sentences",
"=",
"[",
"join_words",
"(",
"sentence",
")",
"for",
"sentence",
"in",
"preprocess_documents",
"(",
"original_sentences",
")",
"]",
"return",
"merge_syntactic_units",
"(",
"original_sentences",
",",
"filtered_sentences",
")"
] | tokenizes a given text into sentences . | train | false |
17,648 | def unpack_tarball(tar_filename, dest_path):
if os.path.exists(tar_filename):
if file_access_rights(dest_path, os.W_OK, check_above=False):
try:
this_tar_file = tarfile.open(tar_filename, 'r:bz2')
except Exception as e:
raise IOError(("[tar] cannot open '%s'" % tar_filename))
else:
try:
this_tar_file.extractall(dest_path)
except Exception as e:
raise IOError(("[tar] error while extracting '%s'" % tar_filename))
else:
pass
else:
raise IOError(("[tar] no right access to '%s'" % dest_path))
else:
raise IOError(("'%s' not found" % tar_filename))
| [
"def",
"unpack_tarball",
"(",
"tar_filename",
",",
"dest_path",
")",
":",
"if",
"os",
".",
"path",
".",
"exists",
"(",
"tar_filename",
")",
":",
"if",
"file_access_rights",
"(",
"dest_path",
",",
"os",
".",
"W_OK",
",",
"check_above",
"=",
"False",
")",
":",
"try",
":",
"this_tar_file",
"=",
"tarfile",
".",
"open",
"(",
"tar_filename",
",",
"'r:bz2'",
")",
"except",
"Exception",
"as",
"e",
":",
"raise",
"IOError",
"(",
"(",
"\"[tar] cannot open '%s'\"",
"%",
"tar_filename",
")",
")",
"else",
":",
"try",
":",
"this_tar_file",
".",
"extractall",
"(",
"dest_path",
")",
"except",
"Exception",
"as",
"e",
":",
"raise",
"IOError",
"(",
"(",
"\"[tar] error while extracting '%s'\"",
"%",
"tar_filename",
")",
")",
"else",
":",
"pass",
"else",
":",
"raise",
"IOError",
"(",
"(",
"\"[tar] no right access to '%s'\"",
"%",
"dest_path",
")",
")",
"else",
":",
"raise",
"IOError",
"(",
"(",
"\"'%s' not found\"",
"%",
"tar_filename",
")",
")"
] | unpacks a tarball to a destination directory . | train | false |
17,650 | def get_errstring(url, base='Invalid URL'):
url_error = url.errorString()
if url_error:
return (base + ' - {}'.format(url_error))
else:
return base
| [
"def",
"get_errstring",
"(",
"url",
",",
"base",
"=",
"'Invalid URL'",
")",
":",
"url_error",
"=",
"url",
".",
"errorString",
"(",
")",
"if",
"url_error",
":",
"return",
"(",
"base",
"+",
"' - {}'",
".",
"format",
"(",
"url_error",
")",
")",
"else",
":",
"return",
"base"
] | get an error string for a url . | train | false |
17,651 | def has_course_ended(end_date):
return ((datetime.now(utc) > end_date) if (end_date is not None) else False)
| [
"def",
"has_course_ended",
"(",
"end_date",
")",
":",
"return",
"(",
"(",
"datetime",
".",
"now",
"(",
"utc",
")",
">",
"end_date",
")",
"if",
"(",
"end_date",
"is",
"not",
"None",
")",
"else",
"False",
")"
] | given a courses end datetime . | train | false |
17,652 | def active(display_progress=False):
ret = {}
client = salt.client.get_local_client(__opts__['conf_file'])
try:
active_ = client.cmd('*', 'saltutil.running', timeout=__opts__['timeout'])
except SaltClientError as client_error:
print(client_error)
return ret
if display_progress:
__jid_event__.fire_event({'message': 'Attempting to contact minions: {0}'.format(list(active_.keys()))}, 'progress')
for (minion, data) in six.iteritems(active_):
if display_progress:
__jid_event__.fire_event({'message': 'Received reply from minion {0}'.format(minion)}, 'progress')
if (not isinstance(data, list)):
continue
for job in data:
if (not (job['jid'] in ret)):
ret[job['jid']] = _format_jid_instance(job['jid'], job)
ret[job['jid']].update({'Running': [{minion: job.get('pid', None)}], 'Returned': []})
else:
ret[job['jid']]['Running'].append({minion: job['pid']})
mminion = salt.minion.MasterMinion(__opts__)
for jid in ret:
returner = _get_returner((__opts__['ext_job_cache'], __opts__['master_job_cache']))
data = mminion.returners['{0}.get_jid'.format(returner)](jid)
for minion in data:
if (minion not in ret[jid]['Returned']):
ret[jid]['Returned'].append(minion)
return ret
| [
"def",
"active",
"(",
"display_progress",
"=",
"False",
")",
":",
"ret",
"=",
"{",
"}",
"client",
"=",
"salt",
".",
"client",
".",
"get_local_client",
"(",
"__opts__",
"[",
"'conf_file'",
"]",
")",
"try",
":",
"active_",
"=",
"client",
".",
"cmd",
"(",
"'*'",
",",
"'saltutil.running'",
",",
"timeout",
"=",
"__opts__",
"[",
"'timeout'",
"]",
")",
"except",
"SaltClientError",
"as",
"client_error",
":",
"print",
"(",
"client_error",
")",
"return",
"ret",
"if",
"display_progress",
":",
"__jid_event__",
".",
"fire_event",
"(",
"{",
"'message'",
":",
"'Attempting to contact minions: {0}'",
".",
"format",
"(",
"list",
"(",
"active_",
".",
"keys",
"(",
")",
")",
")",
"}",
",",
"'progress'",
")",
"for",
"(",
"minion",
",",
"data",
")",
"in",
"six",
".",
"iteritems",
"(",
"active_",
")",
":",
"if",
"display_progress",
":",
"__jid_event__",
".",
"fire_event",
"(",
"{",
"'message'",
":",
"'Received reply from minion {0}'",
".",
"format",
"(",
"minion",
")",
"}",
",",
"'progress'",
")",
"if",
"(",
"not",
"isinstance",
"(",
"data",
",",
"list",
")",
")",
":",
"continue",
"for",
"job",
"in",
"data",
":",
"if",
"(",
"not",
"(",
"job",
"[",
"'jid'",
"]",
"in",
"ret",
")",
")",
":",
"ret",
"[",
"job",
"[",
"'jid'",
"]",
"]",
"=",
"_format_jid_instance",
"(",
"job",
"[",
"'jid'",
"]",
",",
"job",
")",
"ret",
"[",
"job",
"[",
"'jid'",
"]",
"]",
".",
"update",
"(",
"{",
"'Running'",
":",
"[",
"{",
"minion",
":",
"job",
".",
"get",
"(",
"'pid'",
",",
"None",
")",
"}",
"]",
",",
"'Returned'",
":",
"[",
"]",
"}",
")",
"else",
":",
"ret",
"[",
"job",
"[",
"'jid'",
"]",
"]",
"[",
"'Running'",
"]",
".",
"append",
"(",
"{",
"minion",
":",
"job",
"[",
"'pid'",
"]",
"}",
")",
"mminion",
"=",
"salt",
".",
"minion",
".",
"MasterMinion",
"(",
"__opts__",
")",
"for",
"jid",
"in",
"ret",
":",
"returner",
"=",
"_get_returner",
"(",
"(",
"__opts__",
"[",
"'ext_job_cache'",
"]",
",",
"__opts__",
"[",
"'master_job_cache'",
"]",
")",
")",
"data",
"=",
"mminion",
".",
"returners",
"[",
"'{0}.get_jid'",
".",
"format",
"(",
"returner",
")",
"]",
"(",
"jid",
")",
"for",
"minion",
"in",
"data",
":",
"if",
"(",
"minion",
"not",
"in",
"ret",
"[",
"jid",
"]",
"[",
"'Returned'",
"]",
")",
":",
"ret",
"[",
"jid",
"]",
"[",
"'Returned'",
"]",
".",
"append",
"(",
"minion",
")",
"return",
"ret"
] | list of tasks currently being executed . | train | true |
17,653 | def _group_lengths(grouping):
from itertools import chain, repeat
if (not grouping):
return []
elif ((grouping[(-1)] == 0) and (len(grouping) >= 2)):
return chain(grouping[:(-1)], repeat(grouping[(-2)]))
elif (grouping[(-1)] == _locale.CHAR_MAX):
return grouping[:(-1)]
else:
raise ValueError('unrecognised format for grouping')
| [
"def",
"_group_lengths",
"(",
"grouping",
")",
":",
"from",
"itertools",
"import",
"chain",
",",
"repeat",
"if",
"(",
"not",
"grouping",
")",
":",
"return",
"[",
"]",
"elif",
"(",
"(",
"grouping",
"[",
"(",
"-",
"1",
")",
"]",
"==",
"0",
")",
"and",
"(",
"len",
"(",
"grouping",
")",
">=",
"2",
")",
")",
":",
"return",
"chain",
"(",
"grouping",
"[",
":",
"(",
"-",
"1",
")",
"]",
",",
"repeat",
"(",
"grouping",
"[",
"(",
"-",
"2",
")",
"]",
")",
")",
"elif",
"(",
"grouping",
"[",
"(",
"-",
"1",
")",
"]",
"==",
"_locale",
".",
"CHAR_MAX",
")",
":",
"return",
"grouping",
"[",
":",
"(",
"-",
"1",
")",
"]",
"else",
":",
"raise",
"ValueError",
"(",
"'unrecognised format for grouping'",
")"
] | convert a localeconv-style grouping into a iterable of integers representing group lengths . | train | false |
17,654 | def translate_matrix(m, v):
(a, b, c, d, e, f) = m
(x, y) = v
return (a, b, c, d, (((x * a) + (y * c)) + e), (((x * b) + (y * d)) + f))
| [
"def",
"translate_matrix",
"(",
"m",
",",
"v",
")",
":",
"(",
"a",
",",
"b",
",",
"c",
",",
"d",
",",
"e",
",",
"f",
")",
"=",
"m",
"(",
"x",
",",
"y",
")",
"=",
"v",
"return",
"(",
"a",
",",
"b",
",",
"c",
",",
"d",
",",
"(",
"(",
"(",
"x",
"*",
"a",
")",
"+",
"(",
"y",
"*",
"c",
")",
")",
"+",
"e",
")",
",",
"(",
"(",
"(",
"x",
"*",
"b",
")",
"+",
"(",
"y",
"*",
"d",
")",
")",
"+",
"f",
")",
")"
] | translates a matrix by . | train | true |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.