id_within_dataset int64 1 55.5k | snippet stringlengths 19 14.2k | tokens listlengths 6 1.63k | nl stringlengths 6 352 | split_within_dataset stringclasses 1 value | is_duplicated bool 2 classes |
|---|---|---|---|---|---|
19,828 | def select(rlist, wlist, xlist, timeout=None):
allevents = []
timeout = Timeout.start_new(timeout)
result = SelectResult()
try:
try:
for readfd in rlist:
allevents.append(core.read_event(get_fileno(readfd), result.update, arg=readfd))
for writefd in wlist:
allevents.append(core.write_event(get_fileno(writefd), result.update, arg=writefd))
except IOError as ex:
raise error(*ex.args)
result.event.wait(timeout=timeout)
return (result.read, result.write, [])
finally:
for evt in allevents:
evt.cancel()
timeout.cancel()
| [
"def",
"select",
"(",
"rlist",
",",
"wlist",
",",
"xlist",
",",
"timeout",
"=",
"None",
")",
":",
"allevents",
"=",
"[",
"]",
"timeout",
"=",
"Timeout",
".",
"start_new",
"(",
"timeout",
")",
"result",
"=",
"SelectResult",
"(",
")",
"try",
":",
"try",
":",
"for",
"readfd",
"in",
"rlist",
":",
"allevents",
".",
"append",
"(",
"core",
".",
"read_event",
"(",
"get_fileno",
"(",
"readfd",
")",
",",
"result",
".",
"update",
",",
"arg",
"=",
"readfd",
")",
")",
"for",
"writefd",
"in",
"wlist",
":",
"allevents",
".",
"append",
"(",
"core",
".",
"write_event",
"(",
"get_fileno",
"(",
"writefd",
")",
",",
"result",
".",
"update",
",",
"arg",
"=",
"writefd",
")",
")",
"except",
"IOError",
"as",
"ex",
":",
"raise",
"error",
"(",
"*",
"ex",
".",
"args",
")",
"result",
".",
"event",
".",
"wait",
"(",
"timeout",
"=",
"timeout",
")",
"return",
"(",
"result",
".",
"read",
",",
"result",
".",
"write",
",",
"[",
"]",
")",
"finally",
":",
"for",
"evt",
"in",
"allevents",
":",
"evt",
".",
"cancel",
"(",
")",
"timeout",
".",
"cancel",
"(",
")"
] | returns the ith order statistic in array a in linear time . | train | false |
19,829 | def _rescale_data(X, y, sample_weight):
n_samples = X.shape[0]
sample_weight = (sample_weight * np.ones(n_samples))
sample_weight = np.sqrt(sample_weight)
sw_matrix = sparse.dia_matrix((sample_weight, 0), shape=(n_samples, n_samples))
X = safe_sparse_dot(sw_matrix, X)
y = safe_sparse_dot(sw_matrix, y)
return (X, y)
| [
"def",
"_rescale_data",
"(",
"X",
",",
"y",
",",
"sample_weight",
")",
":",
"n_samples",
"=",
"X",
".",
"shape",
"[",
"0",
"]",
"sample_weight",
"=",
"(",
"sample_weight",
"*",
"np",
".",
"ones",
"(",
"n_samples",
")",
")",
"sample_weight",
"=",
"np",
".",
"sqrt",
"(",
"sample_weight",
")",
"sw_matrix",
"=",
"sparse",
".",
"dia_matrix",
"(",
"(",
"sample_weight",
",",
"0",
")",
",",
"shape",
"=",
"(",
"n_samples",
",",
"n_samples",
")",
")",
"X",
"=",
"safe_sparse_dot",
"(",
"sw_matrix",
",",
"X",
")",
"y",
"=",
"safe_sparse_dot",
"(",
"sw_matrix",
",",
"y",
")",
"return",
"(",
"X",
",",
"y",
")"
] | rescale data so as to support sample_weight . | train | false |
19,830 | def copy_files_to(address, client, username, password, port, local_path, remote_path, limit='', log_filename=None, verbose=False, timeout=600, interface=None):
if (client == 'scp'):
scp_to_remote(address, port, username, password, local_path, remote_path, limit, log_filename, timeout, interface=interface)
elif (client == 'rss'):
log_func = None
if verbose:
log_func = logging.debug
c = rss_client.FileUploadClient(address, port, log_func)
c.upload(local_path, remote_path, timeout)
c.close()
| [
"def",
"copy_files_to",
"(",
"address",
",",
"client",
",",
"username",
",",
"password",
",",
"port",
",",
"local_path",
",",
"remote_path",
",",
"limit",
"=",
"''",
",",
"log_filename",
"=",
"None",
",",
"verbose",
"=",
"False",
",",
"timeout",
"=",
"600",
",",
"interface",
"=",
"None",
")",
":",
"if",
"(",
"client",
"==",
"'scp'",
")",
":",
"scp_to_remote",
"(",
"address",
",",
"port",
",",
"username",
",",
"password",
",",
"local_path",
",",
"remote_path",
",",
"limit",
",",
"log_filename",
",",
"timeout",
",",
"interface",
"=",
"interface",
")",
"elif",
"(",
"client",
"==",
"'rss'",
")",
":",
"log_func",
"=",
"None",
"if",
"verbose",
":",
"log_func",
"=",
"logging",
".",
"debug",
"c",
"=",
"rss_client",
".",
"FileUploadClient",
"(",
"address",
",",
"port",
",",
"log_func",
")",
"c",
".",
"upload",
"(",
"local_path",
",",
"remote_path",
",",
"timeout",
")",
"c",
".",
"close",
"(",
")"
] | copy files to a remote host using the selected client . | train | false |
19,832 | def make_dir_obsolete(directory):
p = directory.parent
if ((p is not None) and (p.child_dirs.filter(obsolete=False).count() == 1)):
make_dir_obsolete(p)
directory.obsolete = True
directory.save()
| [
"def",
"make_dir_obsolete",
"(",
"directory",
")",
":",
"p",
"=",
"directory",
".",
"parent",
"if",
"(",
"(",
"p",
"is",
"not",
"None",
")",
"and",
"(",
"p",
".",
"child_dirs",
".",
"filter",
"(",
"obsolete",
"=",
"False",
")",
".",
"count",
"(",
")",
"==",
"1",
")",
")",
":",
"make_dir_obsolete",
"(",
"p",
")",
"directory",
".",
"obsolete",
"=",
"True",
"directory",
".",
"save",
"(",
")"
] | make directory and its parents obsolete if a parent contains one empty directory only . | train | false |
19,833 | def test_undefined_function(caplog):
with caplog.at_level(logging.ERROR):
data = jinja.render('undef.html')
assert ('There was an error while rendering undef.html' in data)
assert ("'does_not_exist' is undefined" in data)
assert data.startswith('<!DOCTYPE html>')
assert (len(caplog.records) == 1)
assert (caplog.records[0].msg == 'UndefinedError while rendering undef.html')
| [
"def",
"test_undefined_function",
"(",
"caplog",
")",
":",
"with",
"caplog",
".",
"at_level",
"(",
"logging",
".",
"ERROR",
")",
":",
"data",
"=",
"jinja",
".",
"render",
"(",
"'undef.html'",
")",
"assert",
"(",
"'There was an error while rendering undef.html'",
"in",
"data",
")",
"assert",
"(",
"\"'does_not_exist' is undefined\"",
"in",
"data",
")",
"assert",
"data",
".",
"startswith",
"(",
"'<!DOCTYPE html>'",
")",
"assert",
"(",
"len",
"(",
"caplog",
".",
"records",
")",
"==",
"1",
")",
"assert",
"(",
"caplog",
".",
"records",
"[",
"0",
"]",
".",
"msg",
"==",
"'UndefinedError while rendering undef.html'",
")"
] | make sure we dont crash if an undefined function is called . | train | false |
19,836 | def detect_logos(path):
vision_client = vision.Client()
with io.open(path, 'rb') as image_file:
content = image_file.read()
image = vision_client.image(content=content)
logos = image.detect_logos()
print 'Logos:'
for logo in logos:
print logo.description
| [
"def",
"detect_logos",
"(",
"path",
")",
":",
"vision_client",
"=",
"vision",
".",
"Client",
"(",
")",
"with",
"io",
".",
"open",
"(",
"path",
",",
"'rb'",
")",
"as",
"image_file",
":",
"content",
"=",
"image_file",
".",
"read",
"(",
")",
"image",
"=",
"vision_client",
".",
"image",
"(",
"content",
"=",
"content",
")",
"logos",
"=",
"image",
".",
"detect_logos",
"(",
")",
"print",
"'Logos:'",
"for",
"logo",
"in",
"logos",
":",
"print",
"logo",
".",
"description"
] | detects logos in the file . | train | false |
19,838 | def check_bcs_lengths(header, mapping_data, warnings):
len_counts = defaultdict(int)
header_field_to_check = 'BarcodeSequence'
try:
check_ix = header.index(header_field_to_check)
except ValueError:
return warnings
for curr_data in range(len(mapping_data)):
len_counts[len(mapping_data[curr_data][check_ix])] += 1
expected_bc_len = max(len_counts.iteritems(), key=itemgetter(1))[0]
correction_ix = 1
for curr_data in range(len(mapping_data)):
if (len(mapping_data[curr_data][check_ix]) != expected_bc_len):
warnings.append(('Barcode %s differs than length %d DCTB %d,%d' % (mapping_data[curr_data][check_ix], expected_bc_len, (curr_data + correction_ix), check_ix)))
return warnings
| [
"def",
"check_bcs_lengths",
"(",
"header",
",",
"mapping_data",
",",
"warnings",
")",
":",
"len_counts",
"=",
"defaultdict",
"(",
"int",
")",
"header_field_to_check",
"=",
"'BarcodeSequence'",
"try",
":",
"check_ix",
"=",
"header",
".",
"index",
"(",
"header_field_to_check",
")",
"except",
"ValueError",
":",
"return",
"warnings",
"for",
"curr_data",
"in",
"range",
"(",
"len",
"(",
"mapping_data",
")",
")",
":",
"len_counts",
"[",
"len",
"(",
"mapping_data",
"[",
"curr_data",
"]",
"[",
"check_ix",
"]",
")",
"]",
"+=",
"1",
"expected_bc_len",
"=",
"max",
"(",
"len_counts",
".",
"iteritems",
"(",
")",
",",
"key",
"=",
"itemgetter",
"(",
"1",
")",
")",
"[",
"0",
"]",
"correction_ix",
"=",
"1",
"for",
"curr_data",
"in",
"range",
"(",
"len",
"(",
"mapping_data",
")",
")",
":",
"if",
"(",
"len",
"(",
"mapping_data",
"[",
"curr_data",
"]",
"[",
"check_ix",
"]",
")",
"!=",
"expected_bc_len",
")",
":",
"warnings",
".",
"append",
"(",
"(",
"'Barcode %s differs than length %d DCTB %d,%d'",
"%",
"(",
"mapping_data",
"[",
"curr_data",
"]",
"[",
"check_ix",
"]",
",",
"expected_bc_len",
",",
"(",
"curr_data",
"+",
"correction_ix",
")",
",",
"check_ix",
")",
")",
")",
"return",
"warnings"
] | adds warnings if barcodes have different lengths as this is mostly intended to find typos in barcodes . | train | false |
19,840 | def sum_layout_dimensions(dimensions):
min = sum([d.min for d in dimensions if (d.min is not None)])
max = sum([d.max for d in dimensions if (d.max is not None)])
preferred = sum([d.preferred for d in dimensions])
return LayoutDimension(min=min, max=max, preferred=preferred)
| [
"def",
"sum_layout_dimensions",
"(",
"dimensions",
")",
":",
"min",
"=",
"sum",
"(",
"[",
"d",
".",
"min",
"for",
"d",
"in",
"dimensions",
"if",
"(",
"d",
".",
"min",
"is",
"not",
"None",
")",
"]",
")",
"max",
"=",
"sum",
"(",
"[",
"d",
".",
"max",
"for",
"d",
"in",
"dimensions",
"if",
"(",
"d",
".",
"max",
"is",
"not",
"None",
")",
"]",
")",
"preferred",
"=",
"sum",
"(",
"[",
"d",
".",
"preferred",
"for",
"d",
"in",
"dimensions",
"]",
")",
"return",
"LayoutDimension",
"(",
"min",
"=",
"min",
",",
"max",
"=",
"max",
",",
"preferred",
"=",
"preferred",
")"
] | sum a list of :class: . | train | true |
19,841 | def _adapt_smtp_secure(value):
if isinstance(value, basestring):
return (value,)
if isinstance(value, config.Config):
assert (set(value.keys()) <= set(['keyfile', 'certfile']))
return (value.keyfile, value.certfile)
if value:
return ()
| [
"def",
"_adapt_smtp_secure",
"(",
"value",
")",
":",
"if",
"isinstance",
"(",
"value",
",",
"basestring",
")",
":",
"return",
"(",
"value",
",",
")",
"if",
"isinstance",
"(",
"value",
",",
"config",
".",
"Config",
")",
":",
"assert",
"(",
"set",
"(",
"value",
".",
"keys",
"(",
")",
")",
"<=",
"set",
"(",
"[",
"'keyfile'",
",",
"'certfile'",
"]",
")",
")",
"return",
"(",
"value",
".",
"keyfile",
",",
"value",
".",
"certfile",
")",
"if",
"value",
":",
"return",
"(",
")"
] | adapt the value to arguments of smtp . | train | false |
19,843 | def test_entity_id():
schema = vol.Schema(cv.entity_id)
with pytest.raises(vol.MultipleInvalid):
schema('invalid_entity')
assert (schema('sensor.LIGHT') == 'sensor.light')
| [
"def",
"test_entity_id",
"(",
")",
":",
"schema",
"=",
"vol",
".",
"Schema",
"(",
"cv",
".",
"entity_id",
")",
"with",
"pytest",
".",
"raises",
"(",
"vol",
".",
"MultipleInvalid",
")",
":",
"schema",
"(",
"'invalid_entity'",
")",
"assert",
"(",
"schema",
"(",
"'sensor.LIGHT'",
")",
"==",
"'sensor.light'",
")"
] | test entity id validation . | train | false |
19,844 | def texture_image(canvas, texture):
if canvas.hasAlphaChannel():
canvas = blend_image(canvas)
return imageops.texture_image(canvas, texture)
| [
"def",
"texture_image",
"(",
"canvas",
",",
"texture",
")",
":",
"if",
"canvas",
".",
"hasAlphaChannel",
"(",
")",
":",
"canvas",
"=",
"blend_image",
"(",
"canvas",
")",
"return",
"imageops",
".",
"texture_image",
"(",
"canvas",
",",
"texture",
")"
] | repeatedly tile the image texture across and down the image canvas . | train | false |
19,845 | def _read_ssh_ed25519_pubkey(keydata):
if (type(keydata) != six.text_type):
raise Exception('invalid type {} for keydata'.format(type(keydata)))
parts = keydata.strip().split()
if (len(parts) != 3):
raise Exception('invalid SSH Ed25519 public key')
(algo, keydata, comment) = parts
if (algo != u'ssh-ed25519'):
raise Exception('not a Ed25519 SSH public key (but {})'.format(algo))
blob = binascii.a2b_base64(keydata)
try:
key = _unpack(blob)[1]
except Exception as e:
raise Exception('could not parse key ({})'.format(e))
if (len(key) != 32):
raise Exception('invalid length {} for embedded raw key (must be 32 bytes)'.format(len(key)))
return (key, comment)
| [
"def",
"_read_ssh_ed25519_pubkey",
"(",
"keydata",
")",
":",
"if",
"(",
"type",
"(",
"keydata",
")",
"!=",
"six",
".",
"text_type",
")",
":",
"raise",
"Exception",
"(",
"'invalid type {} for keydata'",
".",
"format",
"(",
"type",
"(",
"keydata",
")",
")",
")",
"parts",
"=",
"keydata",
".",
"strip",
"(",
")",
".",
"split",
"(",
")",
"if",
"(",
"len",
"(",
"parts",
")",
"!=",
"3",
")",
":",
"raise",
"Exception",
"(",
"'invalid SSH Ed25519 public key'",
")",
"(",
"algo",
",",
"keydata",
",",
"comment",
")",
"=",
"parts",
"if",
"(",
"algo",
"!=",
"u'ssh-ed25519'",
")",
":",
"raise",
"Exception",
"(",
"'not a Ed25519 SSH public key (but {})'",
".",
"format",
"(",
"algo",
")",
")",
"blob",
"=",
"binascii",
".",
"a2b_base64",
"(",
"keydata",
")",
"try",
":",
"key",
"=",
"_unpack",
"(",
"blob",
")",
"[",
"1",
"]",
"except",
"Exception",
"as",
"e",
":",
"raise",
"Exception",
"(",
"'could not parse key ({})'",
".",
"format",
"(",
"e",
")",
")",
"if",
"(",
"len",
"(",
"key",
")",
"!=",
"32",
")",
":",
"raise",
"Exception",
"(",
"'invalid length {} for embedded raw key (must be 32 bytes)'",
".",
"format",
"(",
"len",
"(",
"key",
")",
")",
")",
"return",
"(",
"key",
",",
"comment",
")"
] | parse an openssh ed25519 public key from a string into a raw public key . | train | false |
19,846 | def sysprint(command):
print_debug(('command: %s' % command))
output = subprocess.check_output(command, shell=True, stderr=subprocess.STDOUT)
for line in output.splitlines():
if line.endswith('command not found'):
raise NotFound(output)
elif (line == 'Warning: Using a password on the command line interface can be insecure.'):
pass
else:
print line
| [
"def",
"sysprint",
"(",
"command",
")",
":",
"print_debug",
"(",
"(",
"'command: %s'",
"%",
"command",
")",
")",
"output",
"=",
"subprocess",
".",
"check_output",
"(",
"command",
",",
"shell",
"=",
"True",
",",
"stderr",
"=",
"subprocess",
".",
"STDOUT",
")",
"for",
"line",
"in",
"output",
".",
"splitlines",
"(",
")",
":",
"if",
"line",
".",
"endswith",
"(",
"'command not found'",
")",
":",
"raise",
"NotFound",
"(",
"output",
")",
"elif",
"(",
"line",
"==",
"'Warning: Using a password on the command line interface can be insecure.'",
")",
":",
"pass",
"else",
":",
"print",
"line"
] | helper to print all system commands in debug mode . | train | false |
19,847 | def ErrorDocuments(app, global_conf, mapper, **kw):
if (global_conf is None):
global_conf = {}
return RecursiveMiddleware(StatusBasedForward(app, global_conf=global_conf, mapper=mapper, **kw))
| [
"def",
"ErrorDocuments",
"(",
"app",
",",
"global_conf",
",",
"mapper",
",",
"**",
"kw",
")",
":",
"if",
"(",
"global_conf",
"is",
"None",
")",
":",
"global_conf",
"=",
"{",
"}",
"return",
"RecursiveMiddleware",
"(",
"StatusBasedForward",
"(",
"app",
",",
"global_conf",
"=",
"global_conf",
",",
"mapper",
"=",
"mapper",
",",
"**",
"kw",
")",
")"
] | wraps the app in error docs using paste recursivemiddleware and errordocumentsmiddleware . | train | false |
19,849 | def nat_gateway_exists(nat_gateway_id=None, subnet_id=None, subnet_name=None, vpc_id=None, vpc_name=None, states=('pending', 'available'), region=None, key=None, keyid=None, profile=None):
return bool(_find_nat_gateways(nat_gateway_id=nat_gateway_id, subnet_id=subnet_id, subnet_name=subnet_name, vpc_id=vpc_id, vpc_name=vpc_name, states=states, region=region, key=key, keyid=keyid, profile=profile))
| [
"def",
"nat_gateway_exists",
"(",
"nat_gateway_id",
"=",
"None",
",",
"subnet_id",
"=",
"None",
",",
"subnet_name",
"=",
"None",
",",
"vpc_id",
"=",
"None",
",",
"vpc_name",
"=",
"None",
",",
"states",
"=",
"(",
"'pending'",
",",
"'available'",
")",
",",
"region",
"=",
"None",
",",
"key",
"=",
"None",
",",
"keyid",
"=",
"None",
",",
"profile",
"=",
"None",
")",
":",
"return",
"bool",
"(",
"_find_nat_gateways",
"(",
"nat_gateway_id",
"=",
"nat_gateway_id",
",",
"subnet_id",
"=",
"subnet_id",
",",
"subnet_name",
"=",
"subnet_name",
",",
"vpc_id",
"=",
"vpc_id",
",",
"vpc_name",
"=",
"vpc_name",
",",
"states",
"=",
"states",
",",
"region",
"=",
"region",
",",
"key",
"=",
"key",
",",
"keyid",
"=",
"keyid",
",",
"profile",
"=",
"profile",
")",
")"
] | checks if a nat gateway exists . | train | true |
19,851 | def improve_memory_error_message(error, msg=''):
assert isinstance(error, MemoryError)
if str(error):
raise error
else:
raise TypicalMemoryError(msg)
| [
"def",
"improve_memory_error_message",
"(",
"error",
",",
"msg",
"=",
"''",
")",
":",
"assert",
"isinstance",
"(",
"error",
",",
"MemoryError",
")",
"if",
"str",
"(",
"error",
")",
":",
"raise",
"error",
"else",
":",
"raise",
"TypicalMemoryError",
"(",
"msg",
")"
] | raises a typicalmemoryerror if the memoryerror has no messages parameters error: memoryerror an instance of memoryerror msg: string a message explaining what possibly happened . | train | false |
19,853 | def get_arn(name, region=None, key=None, keyid=None, profile=None):
if name.startswith('arn:aws:sns:'):
return name
account_id = __salt__['boto_iam.get_account_id'](region=region, key=key, keyid=keyid, profile=profile)
return 'arn:aws:sns:{0}:{1}:{2}'.format(_get_region(region, profile), account_id, name)
| [
"def",
"get_arn",
"(",
"name",
",",
"region",
"=",
"None",
",",
"key",
"=",
"None",
",",
"keyid",
"=",
"None",
",",
"profile",
"=",
"None",
")",
":",
"if",
"name",
".",
"startswith",
"(",
"'arn:aws:sns:'",
")",
":",
"return",
"name",
"account_id",
"=",
"__salt__",
"[",
"'boto_iam.get_account_id'",
"]",
"(",
"region",
"=",
"region",
",",
"key",
"=",
"key",
",",
"keyid",
"=",
"keyid",
",",
"profile",
"=",
"profile",
")",
"return",
"'arn:aws:sns:{0}:{1}:{2}'",
".",
"format",
"(",
"_get_region",
"(",
"region",
",",
"profile",
")",
",",
"account_id",
",",
"name",
")"
] | returns the full arn for a given topic name . | train | true |
19,854 | def test_can_get_sentence_from_string():
step = Step.from_string(I_HAVE_TASTY_BEVERAGES)
assert isinstance(step, Step)
assert_equals(step.sentence, string.split(I_HAVE_TASTY_BEVERAGES, '\n')[0])
| [
"def",
"test_can_get_sentence_from_string",
"(",
")",
":",
"step",
"=",
"Step",
".",
"from_string",
"(",
"I_HAVE_TASTY_BEVERAGES",
")",
"assert",
"isinstance",
"(",
"step",
",",
"Step",
")",
"assert_equals",
"(",
"step",
".",
"sentence",
",",
"string",
".",
"split",
"(",
"I_HAVE_TASTY_BEVERAGES",
",",
"'\\n'",
")",
"[",
"0",
"]",
")"
] | it should extract the sentence string from the whole step . | train | false |
19,855 | def add_dicts(*args):
counters = [Counter(arg) for arg in args]
return dict(reduce(operator.add, counters))
| [
"def",
"add_dicts",
"(",
"*",
"args",
")",
":",
"counters",
"=",
"[",
"Counter",
"(",
"arg",
")",
"for",
"arg",
"in",
"args",
"]",
"return",
"dict",
"(",
"reduce",
"(",
"operator",
".",
"add",
",",
"counters",
")",
")"
] | adds two or more dicts together . | train | true |
19,856 | def ShiftedGompertz(name, b, eta):
return rv(name, ShiftedGompertzDistribution, (b, eta))
| [
"def",
"ShiftedGompertz",
"(",
"name",
",",
"b",
",",
"eta",
")",
":",
"return",
"rv",
"(",
"name",
",",
"ShiftedGompertzDistribution",
",",
"(",
"b",
",",
"eta",
")",
")"
] | create a continuous random variable with a shifted gompertz distribution . | train | false |
19,857 | @pytest.mark.django_db
def test_linkcolumn_non_field_based():
class Table(tables.Table, ):
first_name = tables.Column()
delete_link = tables.LinkColumn(u'person_delete', text=u'delete', kwargs={u'pk': tables.A(u'id')})
willem = Person.objects.create(first_name=u'Willem', last_name=u'Wever')
html = Table(Person.objects.all()).as_html(build_request())
expected = u'<td class="delete_link"><a href="{}">delete</a></td>'.format(reverse(u'person_delete', kwargs={u'pk': willem.pk}))
assert (expected in html)
| [
"@",
"pytest",
".",
"mark",
".",
"django_db",
"def",
"test_linkcolumn_non_field_based",
"(",
")",
":",
"class",
"Table",
"(",
"tables",
".",
"Table",
",",
")",
":",
"first_name",
"=",
"tables",
".",
"Column",
"(",
")",
"delete_link",
"=",
"tables",
".",
"LinkColumn",
"(",
"u'person_delete'",
",",
"text",
"=",
"u'delete'",
",",
"kwargs",
"=",
"{",
"u'pk'",
":",
"tables",
".",
"A",
"(",
"u'id'",
")",
"}",
")",
"willem",
"=",
"Person",
".",
"objects",
".",
"create",
"(",
"first_name",
"=",
"u'Willem'",
",",
"last_name",
"=",
"u'Wever'",
")",
"html",
"=",
"Table",
"(",
"Person",
".",
"objects",
".",
"all",
"(",
")",
")",
".",
"as_html",
"(",
"build_request",
"(",
")",
")",
"expected",
"=",
"u'<td class=\"delete_link\"><a href=\"{}\">delete</a></td>'",
".",
"format",
"(",
"reverse",
"(",
"u'person_delete'",
",",
"kwargs",
"=",
"{",
"u'pk'",
":",
"willem",
".",
"pk",
"}",
")",
")",
"assert",
"(",
"expected",
"in",
"html",
")"
] | test for issue 257 . | train | false |
19,858 | def get_master_event(opts, sock_dir, listen=True, io_loop=None, raise_errors=False):
if (opts['transport'] in ('zeromq', 'tcp', 'detect')):
return MasterEvent(sock_dir, opts, listen=listen, io_loop=io_loop, raise_errors=raise_errors)
elif (opts['transport'] == 'raet'):
import salt.utils.raetevent
return salt.utils.raetevent.MasterEvent(opts=opts, sock_dir=sock_dir, listen=listen)
| [
"def",
"get_master_event",
"(",
"opts",
",",
"sock_dir",
",",
"listen",
"=",
"True",
",",
"io_loop",
"=",
"None",
",",
"raise_errors",
"=",
"False",
")",
":",
"if",
"(",
"opts",
"[",
"'transport'",
"]",
"in",
"(",
"'zeromq'",
",",
"'tcp'",
",",
"'detect'",
")",
")",
":",
"return",
"MasterEvent",
"(",
"sock_dir",
",",
"opts",
",",
"listen",
"=",
"listen",
",",
"io_loop",
"=",
"io_loop",
",",
"raise_errors",
"=",
"raise_errors",
")",
"elif",
"(",
"opts",
"[",
"'transport'",
"]",
"==",
"'raet'",
")",
":",
"import",
"salt",
".",
"utils",
".",
"raetevent",
"return",
"salt",
".",
"utils",
".",
"raetevent",
".",
"MasterEvent",
"(",
"opts",
"=",
"opts",
",",
"sock_dir",
"=",
"sock_dir",
",",
"listen",
"=",
"listen",
")"
] | return an event object suitable for the named transport . | train | false |
19,860 | def make_executable(script_path):
status = os.stat(script_path)
os.chmod(script_path, (status.st_mode | stat.S_IEXEC))
| [
"def",
"make_executable",
"(",
"script_path",
")",
":",
"status",
"=",
"os",
".",
"stat",
"(",
"script_path",
")",
"os",
".",
"chmod",
"(",
"script_path",
",",
"(",
"status",
".",
"st_mode",
"|",
"stat",
".",
"S_IEXEC",
")",
")"
] | makes script_path executable . | train | true |
19,861 | def _CalculateWritesForCompositeIndex(index, unique_old_properties, unique_new_properties, common_properties):
old_count = 1
new_count = 1
common_count = 1
for prop in index.property_list():
old_count *= len(unique_old_properties[prop.name()])
new_count *= len(unique_new_properties[prop.name()])
common_count *= common_properties[prop.name()]
return ((old_count - common_count) + (new_count - common_count))
| [
"def",
"_CalculateWritesForCompositeIndex",
"(",
"index",
",",
"unique_old_properties",
",",
"unique_new_properties",
",",
"common_properties",
")",
":",
"old_count",
"=",
"1",
"new_count",
"=",
"1",
"common_count",
"=",
"1",
"for",
"prop",
"in",
"index",
".",
"property_list",
"(",
")",
":",
"old_count",
"*=",
"len",
"(",
"unique_old_properties",
"[",
"prop",
".",
"name",
"(",
")",
"]",
")",
"new_count",
"*=",
"len",
"(",
"unique_new_properties",
"[",
"prop",
".",
"name",
"(",
")",
"]",
")",
"common_count",
"*=",
"common_properties",
"[",
"prop",
".",
"name",
"(",
")",
"]",
"return",
"(",
"(",
"old_count",
"-",
"common_count",
")",
"+",
"(",
"new_count",
"-",
"common_count",
")",
")"
] | calculate the number of writes required to maintain a specific index . | train | false |
19,863 | def get_internaldate(date, received):
if (date is None):
(other, date) = received.split(';')
parsed_date = parsedate_tz(date)
timestamp = mktime_tz(parsed_date)
dt = datetime.utcfromtimestamp(timestamp)
return dt
| [
"def",
"get_internaldate",
"(",
"date",
",",
"received",
")",
":",
"if",
"(",
"date",
"is",
"None",
")",
":",
"(",
"other",
",",
"date",
")",
"=",
"received",
".",
"split",
"(",
"';'",
")",
"parsed_date",
"=",
"parsedate_tz",
"(",
"date",
")",
"timestamp",
"=",
"mktime_tz",
"(",
"parsed_date",
")",
"dt",
"=",
"datetime",
".",
"utcfromtimestamp",
"(",
"timestamp",
")",
"return",
"dt"
] | get the date from the headers . | train | false |
19,864 | def getVersionFileName():
return getFabmetheusUtilitiesPath('version.txt')
| [
"def",
"getVersionFileName",
"(",
")",
":",
"return",
"getFabmetheusUtilitiesPath",
"(",
"'version.txt'",
")"
] | get the file name of the version date . | train | false |
19,865 | def full_scorers_processors():
scorers = [fuzz.ratio]
processors = [(lambda x: x), partial(utils.full_process, force_ascii=False), partial(utils.full_process, force_ascii=True)]
splist = list(product(scorers, processors))
splist.extend([(fuzz.WRatio, partial(utils.full_process, force_ascii=True)), (fuzz.QRatio, partial(utils.full_process, force_ascii=True)), (fuzz.UWRatio, partial(utils.full_process, force_ascii=False)), (fuzz.UQRatio, partial(utils.full_process, force_ascii=False))])
return splist
| [
"def",
"full_scorers_processors",
"(",
")",
":",
"scorers",
"=",
"[",
"fuzz",
".",
"ratio",
"]",
"processors",
"=",
"[",
"(",
"lambda",
"x",
":",
"x",
")",
",",
"partial",
"(",
"utils",
".",
"full_process",
",",
"force_ascii",
"=",
"False",
")",
",",
"partial",
"(",
"utils",
".",
"full_process",
",",
"force_ascii",
"=",
"True",
")",
"]",
"splist",
"=",
"list",
"(",
"product",
"(",
"scorers",
",",
"processors",
")",
")",
"splist",
".",
"extend",
"(",
"[",
"(",
"fuzz",
".",
"WRatio",
",",
"partial",
"(",
"utils",
".",
"full_process",
",",
"force_ascii",
"=",
"True",
")",
")",
",",
"(",
"fuzz",
".",
"QRatio",
",",
"partial",
"(",
"utils",
".",
"full_process",
",",
"force_ascii",
"=",
"True",
")",
")",
",",
"(",
"fuzz",
".",
"UWRatio",
",",
"partial",
"(",
"utils",
".",
"full_process",
",",
"force_ascii",
"=",
"False",
")",
")",
",",
"(",
"fuzz",
".",
"UQRatio",
",",
"partial",
"(",
"utils",
".",
"full_process",
",",
"force_ascii",
"=",
"False",
")",
")",
"]",
")",
"return",
"splist"
] | generate a list of pairs for testing for scorers that use the full string only :return: [ . | train | false |
19,868 | def monkeypatch_django():
from django.contrib.staticfiles.management.commands.runserver import Command as StaticRunserverCommand
from .management.commands.runserver import Command as RunserverCommand
StaticRunserverCommand.__bases__ = (RunserverCommand,)
| [
"def",
"monkeypatch_django",
"(",
")",
":",
"from",
"django",
".",
"contrib",
".",
"staticfiles",
".",
"management",
".",
"commands",
".",
"runserver",
"import",
"Command",
"as",
"StaticRunserverCommand",
"from",
".",
"management",
".",
"commands",
".",
"runserver",
"import",
"Command",
"as",
"RunserverCommand",
"StaticRunserverCommand",
".",
"__bases__",
"=",
"(",
"RunserverCommand",
",",
")"
] | monkeypatches support for us into parts of django . | train | false |
19,871 | def aitchison_aitken_reg(h, Xi, x):
kernel_value = np.ones(Xi.size)
ix = (Xi != x)
inDom = (ix * h)
kernel_value[ix] = inDom[ix]
return kernel_value
| [
"def",
"aitchison_aitken_reg",
"(",
"h",
",",
"Xi",
",",
"x",
")",
":",
"kernel_value",
"=",
"np",
".",
"ones",
"(",
"Xi",
".",
"size",
")",
"ix",
"=",
"(",
"Xi",
"!=",
"x",
")",
"inDom",
"=",
"(",
"ix",
"*",
"h",
")",
"kernel_value",
"[",
"ix",
"]",
"=",
"inDom",
"[",
"ix",
"]",
"return",
"kernel_value"
] | a version for the aitchison-aitken kernel for nonparametric regression . | train | false |
19,872 | @testing.requires_testing_data
def test_read_source_spaces():
src = read_source_spaces(fname, patch_stats=True)
lh_points = src[0]['rr']
lh_faces = src[0]['tris']
lh_use_faces = src[0]['use_tris']
rh_points = src[1]['rr']
rh_faces = src[1]['tris']
rh_use_faces = src[1]['use_tris']
assert_true((lh_faces.min() == 0))
assert_true((lh_faces.max() == (lh_points.shape[0] - 1)))
assert_true((lh_use_faces.min() >= 0))
assert_true((lh_use_faces.max() <= (lh_points.shape[0] - 1)))
assert_true((rh_faces.min() == 0))
assert_true((rh_faces.max() == (rh_points.shape[0] - 1)))
assert_true((rh_use_faces.min() >= 0))
assert_true((rh_use_faces.max() <= (rh_points.shape[0] - 1)))
| [
"@",
"testing",
".",
"requires_testing_data",
"def",
"test_read_source_spaces",
"(",
")",
":",
"src",
"=",
"read_source_spaces",
"(",
"fname",
",",
"patch_stats",
"=",
"True",
")",
"lh_points",
"=",
"src",
"[",
"0",
"]",
"[",
"'rr'",
"]",
"lh_faces",
"=",
"src",
"[",
"0",
"]",
"[",
"'tris'",
"]",
"lh_use_faces",
"=",
"src",
"[",
"0",
"]",
"[",
"'use_tris'",
"]",
"rh_points",
"=",
"src",
"[",
"1",
"]",
"[",
"'rr'",
"]",
"rh_faces",
"=",
"src",
"[",
"1",
"]",
"[",
"'tris'",
"]",
"rh_use_faces",
"=",
"src",
"[",
"1",
"]",
"[",
"'use_tris'",
"]",
"assert_true",
"(",
"(",
"lh_faces",
".",
"min",
"(",
")",
"==",
"0",
")",
")",
"assert_true",
"(",
"(",
"lh_faces",
".",
"max",
"(",
")",
"==",
"(",
"lh_points",
".",
"shape",
"[",
"0",
"]",
"-",
"1",
")",
")",
")",
"assert_true",
"(",
"(",
"lh_use_faces",
".",
"min",
"(",
")",
">=",
"0",
")",
")",
"assert_true",
"(",
"(",
"lh_use_faces",
".",
"max",
"(",
")",
"<=",
"(",
"lh_points",
".",
"shape",
"[",
"0",
"]",
"-",
"1",
")",
")",
")",
"assert_true",
"(",
"(",
"rh_faces",
".",
"min",
"(",
")",
"==",
"0",
")",
")",
"assert_true",
"(",
"(",
"rh_faces",
".",
"max",
"(",
")",
"==",
"(",
"rh_points",
".",
"shape",
"[",
"0",
"]",
"-",
"1",
")",
")",
")",
"assert_true",
"(",
"(",
"rh_use_faces",
".",
"min",
"(",
")",
">=",
"0",
")",
")",
"assert_true",
"(",
"(",
"rh_use_faces",
".",
"max",
"(",
")",
"<=",
"(",
"rh_points",
".",
"shape",
"[",
"0",
"]",
"-",
"1",
")",
")",
")"
] | test reading of source space meshes . | train | false |
19,874 | def _unicode_output(cursor, name, default_type, size, precision, scale):
if (default_type in (cx_Oracle.STRING, cx_Oracle.LONG_STRING, cx_Oracle.FIXED_CHAR, cx_Oracle.CLOB)):
return cursor.var(six.text_type, size, cursor.arraysize)
| [
"def",
"_unicode_output",
"(",
"cursor",
",",
"name",
",",
"default_type",
",",
"size",
",",
"precision",
",",
"scale",
")",
":",
"if",
"(",
"default_type",
"in",
"(",
"cx_Oracle",
".",
"STRING",
",",
"cx_Oracle",
".",
"LONG_STRING",
",",
"cx_Oracle",
".",
"FIXED_CHAR",
",",
"cx_Oracle",
".",
"CLOB",
")",
")",
":",
"return",
"cursor",
".",
"var",
"(",
"six",
".",
"text_type",
",",
"size",
",",
"cursor",
".",
"arraysize",
")"
] | return strings values as python unicode string URL . | train | true |
19,875 | def _loop_lift_modify_blocks(func_ir, loopinfo, blocks, typingctx, targetctx, flags, locals):
from numba.dispatcher import LiftedLoop
loop = loopinfo.loop
loopblockkeys = ((set(loop.body) | set(loop.entries)) | set(loop.exits))
loopblocks = dict(((k, blocks[k].copy()) for k in loopblockkeys))
_loop_lift_prepare_loop_func(loopinfo, loopblocks)
lifted_ir = func_ir.derive(blocks=loopblocks, arg_names=tuple(loopinfo.inputs), arg_count=len(loopinfo.inputs), force_non_generator=True)
liftedloop = LiftedLoop(lifted_ir, typingctx, targetctx, flags, locals)
callblock = _loop_lift_modify_call_block(liftedloop, blocks[loopinfo.callfrom], loopinfo.inputs, loopinfo.outputs, loopinfo.returnto)
for k in loopblockkeys:
del blocks[k]
blocks[loopinfo.callfrom] = callblock
return liftedloop
| [
"def",
"_loop_lift_modify_blocks",
"(",
"func_ir",
",",
"loopinfo",
",",
"blocks",
",",
"typingctx",
",",
"targetctx",
",",
"flags",
",",
"locals",
")",
":",
"from",
"numba",
".",
"dispatcher",
"import",
"LiftedLoop",
"loop",
"=",
"loopinfo",
".",
"loop",
"loopblockkeys",
"=",
"(",
"(",
"set",
"(",
"loop",
".",
"body",
")",
"|",
"set",
"(",
"loop",
".",
"entries",
")",
")",
"|",
"set",
"(",
"loop",
".",
"exits",
")",
")",
"loopblocks",
"=",
"dict",
"(",
"(",
"(",
"k",
",",
"blocks",
"[",
"k",
"]",
".",
"copy",
"(",
")",
")",
"for",
"k",
"in",
"loopblockkeys",
")",
")",
"_loop_lift_prepare_loop_func",
"(",
"loopinfo",
",",
"loopblocks",
")",
"lifted_ir",
"=",
"func_ir",
".",
"derive",
"(",
"blocks",
"=",
"loopblocks",
",",
"arg_names",
"=",
"tuple",
"(",
"loopinfo",
".",
"inputs",
")",
",",
"arg_count",
"=",
"len",
"(",
"loopinfo",
".",
"inputs",
")",
",",
"force_non_generator",
"=",
"True",
")",
"liftedloop",
"=",
"LiftedLoop",
"(",
"lifted_ir",
",",
"typingctx",
",",
"targetctx",
",",
"flags",
",",
"locals",
")",
"callblock",
"=",
"_loop_lift_modify_call_block",
"(",
"liftedloop",
",",
"blocks",
"[",
"loopinfo",
".",
"callfrom",
"]",
",",
"loopinfo",
".",
"inputs",
",",
"loopinfo",
".",
"outputs",
",",
"loopinfo",
".",
"returnto",
")",
"for",
"k",
"in",
"loopblockkeys",
":",
"del",
"blocks",
"[",
"k",
"]",
"blocks",
"[",
"loopinfo",
".",
"callfrom",
"]",
"=",
"callblock",
"return",
"liftedloop"
] | modify the block inplace to call to the lifted-loop . | train | false |
19,876 | def print_inventory(a_device):
fields = ['device_name', 'ip_address', 'device_class', 'ssh_port', 'api_port', 'vendor', 'model', 'device_type', 'os_version', 'serial_number', 'uptime_seconds']
print
print ('#' * 80)
for a_field in fields:
value = getattr(a_device, a_field)
print '{:>15s}: {:<65s}'.format(a_field, str(value))
print ('#' * 80)
print
| [
"def",
"print_inventory",
"(",
"a_device",
")",
":",
"fields",
"=",
"[",
"'device_name'",
",",
"'ip_address'",
",",
"'device_class'",
",",
"'ssh_port'",
",",
"'api_port'",
",",
"'vendor'",
",",
"'model'",
",",
"'device_type'",
",",
"'os_version'",
",",
"'serial_number'",
",",
"'uptime_seconds'",
"]",
"print",
"print",
"(",
"'#'",
"*",
"80",
")",
"for",
"a_field",
"in",
"fields",
":",
"value",
"=",
"getattr",
"(",
"a_device",
",",
"a_field",
")",
"print",
"'{:>15s}: {:<65s}'",
".",
"format",
"(",
"a_field",
",",
"str",
"(",
"value",
")",
")",
"print",
"(",
"'#'",
"*",
"80",
")",
"print"
] | print network device inventory information . | train | false |
19,877 | def sentMessage(ignored, group, avatar):
l = group.leave()
l.addCallback(leftGroup, avatar)
return l
| [
"def",
"sentMessage",
"(",
"ignored",
",",
"group",
",",
"avatar",
")",
":",
"l",
"=",
"group",
".",
"leave",
"(",
")",
"l",
".",
"addCallback",
"(",
"leftGroup",
",",
"avatar",
")",
"return",
"l"
] | sent the message successfully . | train | false |
19,878 | def revdep_rebuild(lib=None):
cmd = 'revdep-rebuild -i --quiet --no-progress'
if (lib is not None):
cmd += ' --library={0}'.format(lib)
return (__salt__['cmd.retcode'](cmd, python_shell=False) == 0)
| [
"def",
"revdep_rebuild",
"(",
"lib",
"=",
"None",
")",
":",
"cmd",
"=",
"'revdep-rebuild -i --quiet --no-progress'",
"if",
"(",
"lib",
"is",
"not",
"None",
")",
":",
"cmd",
"+=",
"' --library={0}'",
".",
"format",
"(",
"lib",
")",
"return",
"(",
"__salt__",
"[",
"'cmd.retcode'",
"]",
"(",
"cmd",
",",
"python_shell",
"=",
"False",
")",
"==",
"0",
")"
] | fix up broken reverse dependencies lib search for reverse dependencies for a particular library rather than every library on the system . | train | true |
19,879 | @Profiler.profile
def test_orm_flush(n):
session = Session(bind=engine)
for chunk in range(0, n, 1000):
customers = session.query(Customer).filter(Customer.id.between(chunk, (chunk + 1000))).all()
for customer in customers:
customer.description += 'updated'
session.flush()
session.commit()
| [
"@",
"Profiler",
".",
"profile",
"def",
"test_orm_flush",
"(",
"n",
")",
":",
"session",
"=",
"Session",
"(",
"bind",
"=",
"engine",
")",
"for",
"chunk",
"in",
"range",
"(",
"0",
",",
"n",
",",
"1000",
")",
":",
"customers",
"=",
"session",
".",
"query",
"(",
"Customer",
")",
".",
"filter",
"(",
"Customer",
".",
"id",
".",
"between",
"(",
"chunk",
",",
"(",
"chunk",
"+",
"1000",
")",
")",
")",
".",
"all",
"(",
")",
"for",
"customer",
"in",
"customers",
":",
"customer",
".",
"description",
"+=",
"'updated'",
"session",
".",
"flush",
"(",
")",
"session",
".",
"commit",
"(",
")"
] | update statements via the orm flush process . | train | false |
19,880 | def to_255(vals_01):
try:
ret = [(v * 255) for v in vals_01]
if (type(vals_01) is tuple):
return tuple(ret)
else:
return ret
except TypeError:
return (vals_01 * 255)
| [
"def",
"to_255",
"(",
"vals_01",
")",
":",
"try",
":",
"ret",
"=",
"[",
"(",
"v",
"*",
"255",
")",
"for",
"v",
"in",
"vals_01",
"]",
"if",
"(",
"type",
"(",
"vals_01",
")",
"is",
"tuple",
")",
":",
"return",
"tuple",
"(",
"ret",
")",
"else",
":",
"return",
"ret",
"except",
"TypeError",
":",
"return",
"(",
"vals_01",
"*",
"255",
")"
] | convert vals in [0 . | train | false |
19,881 | def getNewRepository():
return ExportRepository()
| [
"def",
"getNewRepository",
"(",
")",
":",
"return",
"ExportRepository",
"(",
")"
] | get the repository constructor . | train | false |
19,882 | def DeleteEntity(key):
if (key.kind() in MAPREDUCE_OBJECTS):
entity = datastore.Get(key)
if (entity and (not entity['active'])):
(yield operation.db.Delete(key))
elif (key.kind() == utils.DatastoreAdminOperation.kind()):
entity = datastore.Get(key)
if (entity and (not entity['active_jobs'])):
(yield operation.db.Delete(key))
else:
(yield operation.db.Delete(key))
| [
"def",
"DeleteEntity",
"(",
"key",
")",
":",
"if",
"(",
"key",
".",
"kind",
"(",
")",
"in",
"MAPREDUCE_OBJECTS",
")",
":",
"entity",
"=",
"datastore",
".",
"Get",
"(",
"key",
")",
"if",
"(",
"entity",
"and",
"(",
"not",
"entity",
"[",
"'active'",
"]",
")",
")",
":",
"(",
"yield",
"operation",
".",
"db",
".",
"Delete",
"(",
"key",
")",
")",
"elif",
"(",
"key",
".",
"kind",
"(",
")",
"==",
"utils",
".",
"DatastoreAdminOperation",
".",
"kind",
"(",
")",
")",
":",
"entity",
"=",
"datastore",
".",
"Get",
"(",
"key",
")",
"if",
"(",
"entity",
"and",
"(",
"not",
"entity",
"[",
"'active_jobs'",
"]",
")",
")",
":",
"(",
"yield",
"operation",
".",
"db",
".",
"Delete",
"(",
"key",
")",
")",
"else",
":",
"(",
"yield",
"operation",
".",
"db",
".",
"Delete",
"(",
"key",
")",
")"
] | delete function which deletes all processed entities . | train | false |
19,883 | def map_subproject_slug(view_func):
@wraps(view_func)
def inner_view(request, subproject=None, subproject_slug=None, *args, **kwargs):
if ((subproject is None) and subproject_slug):
try:
subproject = Project.objects.get(slug=subproject_slug)
except Project.DoesNotExist:
try:
rel = ProjectRelationship.objects.get(parent=kwargs['project'], alias=subproject_slug)
subproject = rel.child
except (ProjectRelationship.DoesNotExist, KeyError):
raise Http404
return view_func(request, subproject=subproject, *args, **kwargs)
return inner_view
| [
"def",
"map_subproject_slug",
"(",
"view_func",
")",
":",
"@",
"wraps",
"(",
"view_func",
")",
"def",
"inner_view",
"(",
"request",
",",
"subproject",
"=",
"None",
",",
"subproject_slug",
"=",
"None",
",",
"*",
"args",
",",
"**",
"kwargs",
")",
":",
"if",
"(",
"(",
"subproject",
"is",
"None",
")",
"and",
"subproject_slug",
")",
":",
"try",
":",
"subproject",
"=",
"Project",
".",
"objects",
".",
"get",
"(",
"slug",
"=",
"subproject_slug",
")",
"except",
"Project",
".",
"DoesNotExist",
":",
"try",
":",
"rel",
"=",
"ProjectRelationship",
".",
"objects",
".",
"get",
"(",
"parent",
"=",
"kwargs",
"[",
"'project'",
"]",
",",
"alias",
"=",
"subproject_slug",
")",
"subproject",
"=",
"rel",
".",
"child",
"except",
"(",
"ProjectRelationship",
".",
"DoesNotExist",
",",
"KeyError",
")",
":",
"raise",
"Http404",
"return",
"view_func",
"(",
"request",
",",
"subproject",
"=",
"subproject",
",",
"*",
"args",
",",
"**",
"kwargs",
")",
"return",
"inner_view"
] | a decorator that maps a subproject_slug url param into a project . | train | false |
19,884 | def is_opentype_cff_font(filename):
if (os.path.splitext(filename)[1].lower() == u'.otf'):
result = _is_opentype_cff_font_cache.get(filename)
if (result is None):
with open(filename, u'rb') as fd:
tag = fd.read(4)
result = (tag == 'OTTO')
_is_opentype_cff_font_cache[filename] = result
return result
return False
| [
"def",
"is_opentype_cff_font",
"(",
"filename",
")",
":",
"if",
"(",
"os",
".",
"path",
".",
"splitext",
"(",
"filename",
")",
"[",
"1",
"]",
".",
"lower",
"(",
")",
"==",
"u'.otf'",
")",
":",
"result",
"=",
"_is_opentype_cff_font_cache",
".",
"get",
"(",
"filename",
")",
"if",
"(",
"result",
"is",
"None",
")",
":",
"with",
"open",
"(",
"filename",
",",
"u'rb'",
")",
"as",
"fd",
":",
"tag",
"=",
"fd",
".",
"read",
"(",
"4",
")",
"result",
"=",
"(",
"tag",
"==",
"'OTTO'",
")",
"_is_opentype_cff_font_cache",
"[",
"filename",
"]",
"=",
"result",
"return",
"result",
"return",
"False"
] | returns true if the given font is a postscript compact font format font embedded in an opentype wrapper . | train | false |
19,885 | def get_distribution(dist):
if isinstance(dist, basestring):
dist = Requirement.parse(dist)
if isinstance(dist, Requirement):
dist = get_provider(dist)
if (not isinstance(dist, Distribution)):
raise TypeError('Expected string, Requirement, or Distribution', dist)
return dist
| [
"def",
"get_distribution",
"(",
"dist",
")",
":",
"if",
"isinstance",
"(",
"dist",
",",
"basestring",
")",
":",
"dist",
"=",
"Requirement",
".",
"parse",
"(",
"dist",
")",
"if",
"isinstance",
"(",
"dist",
",",
"Requirement",
")",
":",
"dist",
"=",
"get_provider",
"(",
"dist",
")",
"if",
"(",
"not",
"isinstance",
"(",
"dist",
",",
"Distribution",
")",
")",
":",
"raise",
"TypeError",
"(",
"'Expected string, Requirement, or Distribution'",
",",
"dist",
")",
"return",
"dist"
] | return a current distribution object for a requirement or string . | train | false |
19,887 | def get_git_name(repository_dir, git_dir):
if git_dir.startswith(repository_dir):
git_name = git_dir[len(repository_dir):]
else:
git_name = os.path.split(git_dir)[1]
git_name = git_name.strip('\\/')
if git_name.endswith('.git'):
git_name = git_name[:(-4)]
return git_name
| [
"def",
"get_git_name",
"(",
"repository_dir",
",",
"git_dir",
")",
":",
"if",
"git_dir",
".",
"startswith",
"(",
"repository_dir",
")",
":",
"git_name",
"=",
"git_dir",
"[",
"len",
"(",
"repository_dir",
")",
":",
"]",
"else",
":",
"git_name",
"=",
"os",
".",
"path",
".",
"split",
"(",
"git_dir",
")",
"[",
"1",
"]",
"git_name",
"=",
"git_name",
".",
"strip",
"(",
"'\\\\/'",
")",
"if",
"git_name",
".",
"endswith",
"(",
"'.git'",
")",
":",
"git_name",
"=",
"git_name",
"[",
":",
"(",
"-",
"4",
")",
"]",
"return",
"git_name"
] | guess the name of the repository used in gitosis . | train | false |
19,889 | def _object_matcher(obj):
if isinstance(obj, TreeElement):
return _identity_matcher(obj)
if isinstance(obj, type):
return _class_matcher(obj)
if isinstance(obj, basestring):
return _string_matcher(obj)
if isinstance(obj, dict):
return _attribute_matcher(obj)
if callable(obj):
return _function_matcher(obj)
raise ValueError(('%s (type %s) is not a valid type for comparison.' % (obj, type(obj))))
| [
"def",
"_object_matcher",
"(",
"obj",
")",
":",
"if",
"isinstance",
"(",
"obj",
",",
"TreeElement",
")",
":",
"return",
"_identity_matcher",
"(",
"obj",
")",
"if",
"isinstance",
"(",
"obj",
",",
"type",
")",
":",
"return",
"_class_matcher",
"(",
"obj",
")",
"if",
"isinstance",
"(",
"obj",
",",
"basestring",
")",
":",
"return",
"_string_matcher",
"(",
"obj",
")",
"if",
"isinstance",
"(",
"obj",
",",
"dict",
")",
":",
"return",
"_attribute_matcher",
"(",
"obj",
")",
"if",
"callable",
"(",
"obj",
")",
":",
"return",
"_function_matcher",
"(",
"obj",
")",
"raise",
"ValueError",
"(",
"(",
"'%s (type %s) is not a valid type for comparison.'",
"%",
"(",
"obj",
",",
"type",
"(",
"obj",
")",
")",
")",
")"
] | retrieve a matcher function by passing an arbitrary object . | train | false |
19,892 | def iter_default_settings():
for name in dir(default_settings):
if name.isupper():
(yield (name, getattr(default_settings, name)))
| [
"def",
"iter_default_settings",
"(",
")",
":",
"for",
"name",
"in",
"dir",
"(",
"default_settings",
")",
":",
"if",
"name",
".",
"isupper",
"(",
")",
":",
"(",
"yield",
"(",
"name",
",",
"getattr",
"(",
"default_settings",
",",
"name",
")",
")",
")"
] | return the default settings as an iterator of tuples . | train | false |
19,894 | def cast_bytes(s, encoding='utf8', errors='strict'):
if isinstance(s, bytes):
return s
elif isinstance(s, unicode):
return s.encode(encoding, errors)
else:
raise TypeError(('Expected unicode or bytes, got %r' % s))
| [
"def",
"cast_bytes",
"(",
"s",
",",
"encoding",
"=",
"'utf8'",
",",
"errors",
"=",
"'strict'",
")",
":",
"if",
"isinstance",
"(",
"s",
",",
"bytes",
")",
":",
"return",
"s",
"elif",
"isinstance",
"(",
"s",
",",
"unicode",
")",
":",
"return",
"s",
".",
"encode",
"(",
"encoding",
",",
"errors",
")",
"else",
":",
"raise",
"TypeError",
"(",
"(",
"'Expected unicode or bytes, got %r'",
"%",
"s",
")",
")"
] | cast unicode or bytes to bytes . | train | true |
19,895 | def getVector3ByDictionaryListValue(value, vector3):
if ((value.__class__ == Vector3) or (value.__class__.__name__ == 'Vector3Index')):
return value
if (value.__class__ == dict):
return getVector3ByDictionary(value, vector3)
if (value.__class__ == list):
return getVector3ByFloatList(value, vector3)
floatFromValue = euclidean.getFloatFromValue(value)
if (floatFromValue == None):
return vector3
vector3.setToXYZ(floatFromValue, floatFromValue, floatFromValue)
return vector3
| [
"def",
"getVector3ByDictionaryListValue",
"(",
"value",
",",
"vector3",
")",
":",
"if",
"(",
"(",
"value",
".",
"__class__",
"==",
"Vector3",
")",
"or",
"(",
"value",
".",
"__class__",
".",
"__name__",
"==",
"'Vector3Index'",
")",
")",
":",
"return",
"value",
"if",
"(",
"value",
".",
"__class__",
"==",
"dict",
")",
":",
"return",
"getVector3ByDictionary",
"(",
"value",
",",
"vector3",
")",
"if",
"(",
"value",
".",
"__class__",
"==",
"list",
")",
":",
"return",
"getVector3ByFloatList",
"(",
"value",
",",
"vector3",
")",
"floatFromValue",
"=",
"euclidean",
".",
"getFloatFromValue",
"(",
"value",
")",
"if",
"(",
"floatFromValue",
"==",
"None",
")",
":",
"return",
"vector3",
"vector3",
".",
"setToXYZ",
"(",
"floatFromValue",
",",
"floatFromValue",
",",
"floatFromValue",
")",
"return",
"vector3"
] | get vector3 by dictionary . | train | false |
19,896 | def test_batch_normalized_mlp_learn_scale_propagated_at_alloc():
mlp = BatchNormalizedMLP([Tanh(), Tanh()], [5, 7, 9], learn_scale=False)
assert (not mlp.learn_scale)
assert all((act.children[0].learn_scale for act in mlp.activations))
mlp.allocate()
assert (not any((act.children[0].learn_scale for act in mlp.activations)))
| [
"def",
"test_batch_normalized_mlp_learn_scale_propagated_at_alloc",
"(",
")",
":",
"mlp",
"=",
"BatchNormalizedMLP",
"(",
"[",
"Tanh",
"(",
")",
",",
"Tanh",
"(",
")",
"]",
",",
"[",
"5",
",",
"7",
",",
"9",
"]",
",",
"learn_scale",
"=",
"False",
")",
"assert",
"(",
"not",
"mlp",
".",
"learn_scale",
")",
"assert",
"all",
"(",
"(",
"act",
".",
"children",
"[",
"0",
"]",
".",
"learn_scale",
"for",
"act",
"in",
"mlp",
".",
"activations",
")",
")",
"mlp",
".",
"allocate",
"(",
")",
"assert",
"(",
"not",
"any",
"(",
"(",
"act",
".",
"children",
"[",
"0",
"]",
".",
"learn_scale",
"for",
"act",
"in",
"mlp",
".",
"activations",
")",
")",
")"
] | test that setting learn_scale on a batchnormalizedmlp works . | train | false |
19,897 | def batchScoreIAT(path='.', write_file=False):
files = glob.glob(os.path.join(path, '*.csv'))
for f in files:
scoreIAT(f, write_file=write_file)
| [
"def",
"batchScoreIAT",
"(",
"path",
"=",
"'.'",
",",
"write_file",
"=",
"False",
")",
":",
"files",
"=",
"glob",
".",
"glob",
"(",
"os",
".",
"path",
".",
"join",
"(",
"path",
",",
"'*.csv'",
")",
")",
"for",
"f",
"in",
"files",
":",
"scoreIAT",
"(",
"f",
",",
"write_file",
"=",
"write_file",
")"
] | call scoreiat() on all csv files in path . | train | false |
19,898 | def __grant_generate(grant, database, user, host='localhost', grant_option=False, escape=True, ssl_option=False):
grant = re.sub('\\s*,\\s*', ', ', grant).upper()
grant = __grant_normalize(grant)
db_part = database.rpartition('.')
dbc = db_part[0]
table = db_part[2]
if escape:
if (dbc is not '*'):
dbc = quote_identifier(dbc, for_grants=(table is '*'))
if (table is not '*'):
table = quote_identifier(table)
qry = 'GRANT {0} ON {1}.{2} TO %(user)s@%(host)s'.format(grant, dbc, table)
args = {}
args['user'] = user
args['host'] = host
if (isinstance(ssl_option, list) and len(ssl_option)):
qry += __ssl_option_sanitize(ssl_option)
if salt.utils.is_true(grant_option):
qry += ' WITH GRANT OPTION'
log.debug('Grant Query generated: {0} args {1}'.format(qry, repr(args)))
return {'qry': qry, 'args': args}
| [
"def",
"__grant_generate",
"(",
"grant",
",",
"database",
",",
"user",
",",
"host",
"=",
"'localhost'",
",",
"grant_option",
"=",
"False",
",",
"escape",
"=",
"True",
",",
"ssl_option",
"=",
"False",
")",
":",
"grant",
"=",
"re",
".",
"sub",
"(",
"'\\\\s*,\\\\s*'",
",",
"', '",
",",
"grant",
")",
".",
"upper",
"(",
")",
"grant",
"=",
"__grant_normalize",
"(",
"grant",
")",
"db_part",
"=",
"database",
".",
"rpartition",
"(",
"'.'",
")",
"dbc",
"=",
"db_part",
"[",
"0",
"]",
"table",
"=",
"db_part",
"[",
"2",
"]",
"if",
"escape",
":",
"if",
"(",
"dbc",
"is",
"not",
"'*'",
")",
":",
"dbc",
"=",
"quote_identifier",
"(",
"dbc",
",",
"for_grants",
"=",
"(",
"table",
"is",
"'*'",
")",
")",
"if",
"(",
"table",
"is",
"not",
"'*'",
")",
":",
"table",
"=",
"quote_identifier",
"(",
"table",
")",
"qry",
"=",
"'GRANT {0} ON {1}.{2} TO %(user)s@%(host)s'",
".",
"format",
"(",
"grant",
",",
"dbc",
",",
"table",
")",
"args",
"=",
"{",
"}",
"args",
"[",
"'user'",
"]",
"=",
"user",
"args",
"[",
"'host'",
"]",
"=",
"host",
"if",
"(",
"isinstance",
"(",
"ssl_option",
",",
"list",
")",
"and",
"len",
"(",
"ssl_option",
")",
")",
":",
"qry",
"+=",
"__ssl_option_sanitize",
"(",
"ssl_option",
")",
"if",
"salt",
".",
"utils",
".",
"is_true",
"(",
"grant_option",
")",
":",
"qry",
"+=",
"' WITH GRANT OPTION'",
"log",
".",
"debug",
"(",
"'Grant Query generated: {0} args {1}'",
".",
"format",
"(",
"qry",
",",
"repr",
"(",
"args",
")",
")",
")",
"return",
"{",
"'qry'",
":",
"qry",
",",
"'args'",
":",
"args",
"}"
] | validate grants and build the query that could set the given grants note that this query contains arguments for user and host but not for grants or database . | train | true |
19,899 | def get_todays_events(as_list=False):
from frappe.desk.doctype.event.event import get_events
from frappe.utils import nowdate
today = nowdate()
events = get_events(today, today)
return (events if as_list else len(events))
| [
"def",
"get_todays_events",
"(",
"as_list",
"=",
"False",
")",
":",
"from",
"frappe",
".",
"desk",
".",
"doctype",
".",
"event",
".",
"event",
"import",
"get_events",
"from",
"frappe",
".",
"utils",
"import",
"nowdate",
"today",
"=",
"nowdate",
"(",
")",
"events",
"=",
"get_events",
"(",
"today",
",",
"today",
")",
"return",
"(",
"events",
"if",
"as_list",
"else",
"len",
"(",
"events",
")",
")"
] | returns a count of todays events in calendar . | train | false |
19,900 | def cmp_to_key(mycmp):
class KeyClass(object, ):
'Key class'
def __init__(self, obj, *args):
self.obj = obj
def __lt__(self, other):
return (mycmp(self.obj, other.obj) < 0)
def __gt__(self, other):
return (mycmp(self.obj, other.obj) > 0)
def __eq__(self, other):
return (mycmp(self.obj, other.obj) == 0)
def __le__(self, other):
return (mycmp(self.obj, other.obj) <= 0)
def __ge__(self, other):
return (mycmp(self.obj, other.obj) >= 0)
def __ne__(self, other):
return (mycmp(self.obj, other.obj) != 0)
return KeyClass
| [
"def",
"cmp_to_key",
"(",
"mycmp",
")",
":",
"class",
"KeyClass",
"(",
"object",
",",
")",
":",
"def",
"__init__",
"(",
"self",
",",
"obj",
",",
"*",
"args",
")",
":",
"self",
".",
"obj",
"=",
"obj",
"def",
"__lt__",
"(",
"self",
",",
"other",
")",
":",
"return",
"(",
"mycmp",
"(",
"self",
".",
"obj",
",",
"other",
".",
"obj",
")",
"<",
"0",
")",
"def",
"__gt__",
"(",
"self",
",",
"other",
")",
":",
"return",
"(",
"mycmp",
"(",
"self",
".",
"obj",
",",
"other",
".",
"obj",
")",
">",
"0",
")",
"def",
"__eq__",
"(",
"self",
",",
"other",
")",
":",
"return",
"(",
"mycmp",
"(",
"self",
".",
"obj",
",",
"other",
".",
"obj",
")",
"==",
"0",
")",
"def",
"__le__",
"(",
"self",
",",
"other",
")",
":",
"return",
"(",
"mycmp",
"(",
"self",
".",
"obj",
",",
"other",
".",
"obj",
")",
"<=",
"0",
")",
"def",
"__ge__",
"(",
"self",
",",
"other",
")",
":",
"return",
"(",
"mycmp",
"(",
"self",
".",
"obj",
",",
"other",
".",
"obj",
")",
">=",
"0",
")",
"def",
"__ne__",
"(",
"self",
",",
"other",
")",
":",
"return",
"(",
"mycmp",
"(",
"self",
".",
"obj",
",",
"other",
".",
"obj",
")",
"!=",
"0",
")",
"return",
"KeyClass"
] | convert a cmp= function into a key= function . | train | false |
19,901 | def test_iforest_parallel_regression():
rng = check_random_state(0)
(X_train, X_test, y_train, y_test) = train_test_split(boston.data, boston.target, random_state=rng)
ensemble = IsolationForest(n_jobs=3, random_state=0).fit(X_train)
ensemble.set_params(n_jobs=1)
y1 = ensemble.predict(X_test)
ensemble.set_params(n_jobs=2)
y2 = ensemble.predict(X_test)
assert_array_almost_equal(y1, y2)
ensemble = IsolationForest(n_jobs=1, random_state=0).fit(X_train)
y3 = ensemble.predict(X_test)
assert_array_almost_equal(y1, y3)
| [
"def",
"test_iforest_parallel_regression",
"(",
")",
":",
"rng",
"=",
"check_random_state",
"(",
"0",
")",
"(",
"X_train",
",",
"X_test",
",",
"y_train",
",",
"y_test",
")",
"=",
"train_test_split",
"(",
"boston",
".",
"data",
",",
"boston",
".",
"target",
",",
"random_state",
"=",
"rng",
")",
"ensemble",
"=",
"IsolationForest",
"(",
"n_jobs",
"=",
"3",
",",
"random_state",
"=",
"0",
")",
".",
"fit",
"(",
"X_train",
")",
"ensemble",
".",
"set_params",
"(",
"n_jobs",
"=",
"1",
")",
"y1",
"=",
"ensemble",
".",
"predict",
"(",
"X_test",
")",
"ensemble",
".",
"set_params",
"(",
"n_jobs",
"=",
"2",
")",
"y2",
"=",
"ensemble",
".",
"predict",
"(",
"X_test",
")",
"assert_array_almost_equal",
"(",
"y1",
",",
"y2",
")",
"ensemble",
"=",
"IsolationForest",
"(",
"n_jobs",
"=",
"1",
",",
"random_state",
"=",
"0",
")",
".",
"fit",
"(",
"X_train",
")",
"y3",
"=",
"ensemble",
".",
"predict",
"(",
"X_test",
")",
"assert_array_almost_equal",
"(",
"y1",
",",
"y3",
")"
] | check parallel regression . | train | false |
19,902 | @testing.requires_testing_data
def test_time_index():
raw_fname = op.join(op.dirname(__file__), '..', '..', 'io', 'tests', 'data', 'test_raw.fif')
raw = read_raw_fif(raw_fname)
orig_inds = raw.time_as_index(raw.times)
assert (len(set(orig_inds)) != len(orig_inds))
new_inds = raw.time_as_index(raw.times, use_rounding=True)
assert (len(set(new_inds)) == len(new_inds))
| [
"@",
"testing",
".",
"requires_testing_data",
"def",
"test_time_index",
"(",
")",
":",
"raw_fname",
"=",
"op",
".",
"join",
"(",
"op",
".",
"dirname",
"(",
"__file__",
")",
",",
"'..'",
",",
"'..'",
",",
"'io'",
",",
"'tests'",
",",
"'data'",
",",
"'test_raw.fif'",
")",
"raw",
"=",
"read_raw_fif",
"(",
"raw_fname",
")",
"orig_inds",
"=",
"raw",
".",
"time_as_index",
"(",
"raw",
".",
"times",
")",
"assert",
"(",
"len",
"(",
"set",
"(",
"orig_inds",
")",
")",
"!=",
"len",
"(",
"orig_inds",
")",
")",
"new_inds",
"=",
"raw",
".",
"time_as_index",
"(",
"raw",
".",
"times",
",",
"use_rounding",
"=",
"True",
")",
"assert",
"(",
"len",
"(",
"set",
"(",
"new_inds",
")",
")",
"==",
"len",
"(",
"new_inds",
")",
")"
] | test indexing of raw times . | train | false |
19,906 | def reverse_readline(fh, start_byte=0, buf_size=8192):
segment = None
offset = 0
if start_byte:
fh.seek(start_byte)
else:
fh.seek(0, os.SEEK_END)
total_size = remaining_size = fh.tell()
while (remaining_size > 0):
offset = min(total_size, (offset + buf_size))
fh.seek((- offset), os.SEEK_END)
buf = fh.read(min(remaining_size, buf_size))
remaining_size -= buf_size
lines = buf.decode(sys.getfilesystemencoding()).split(u'\n')
if (segment is not None):
if (buf[(-1)] is not u'\n'):
lines[(-1)] += segment
else:
(yield segment)
segment = lines[0]
for index in range((len(lines) - 1), 0, (-1)):
if len(lines[index]):
(yield lines[index])
(yield segment)
| [
"def",
"reverse_readline",
"(",
"fh",
",",
"start_byte",
"=",
"0",
",",
"buf_size",
"=",
"8192",
")",
":",
"segment",
"=",
"None",
"offset",
"=",
"0",
"if",
"start_byte",
":",
"fh",
".",
"seek",
"(",
"start_byte",
")",
"else",
":",
"fh",
".",
"seek",
"(",
"0",
",",
"os",
".",
"SEEK_END",
")",
"total_size",
"=",
"remaining_size",
"=",
"fh",
".",
"tell",
"(",
")",
"while",
"(",
"remaining_size",
">",
"0",
")",
":",
"offset",
"=",
"min",
"(",
"total_size",
",",
"(",
"offset",
"+",
"buf_size",
")",
")",
"fh",
".",
"seek",
"(",
"(",
"-",
"offset",
")",
",",
"os",
".",
"SEEK_END",
")",
"buf",
"=",
"fh",
".",
"read",
"(",
"min",
"(",
"remaining_size",
",",
"buf_size",
")",
")",
"remaining_size",
"-=",
"buf_size",
"lines",
"=",
"buf",
".",
"decode",
"(",
"sys",
".",
"getfilesystemencoding",
"(",
")",
")",
".",
"split",
"(",
"u'\\n'",
")",
"if",
"(",
"segment",
"is",
"not",
"None",
")",
":",
"if",
"(",
"buf",
"[",
"(",
"-",
"1",
")",
"]",
"is",
"not",
"u'\\n'",
")",
":",
"lines",
"[",
"(",
"-",
"1",
")",
"]",
"+=",
"segment",
"else",
":",
"(",
"yield",
"segment",
")",
"segment",
"=",
"lines",
"[",
"0",
"]",
"for",
"index",
"in",
"range",
"(",
"(",
"len",
"(",
"lines",
")",
"-",
"1",
")",
",",
"0",
",",
"(",
"-",
"1",
")",
")",
":",
"if",
"len",
"(",
"lines",
"[",
"index",
"]",
")",
":",
"(",
"yield",
"lines",
"[",
"index",
"]",
")",
"(",
"yield",
"segment",
")"
] | a generator that returns the lines of a file in reverse order . | train | false |
19,909 | def expireat(key, timestamp, host=None, port=None, db=None, password=None):
server = _connect(host, port, db, password)
return server.expireat(key, timestamp)
| [
"def",
"expireat",
"(",
"key",
",",
"timestamp",
",",
"host",
"=",
"None",
",",
"port",
"=",
"None",
",",
"db",
"=",
"None",
",",
"password",
"=",
"None",
")",
":",
"server",
"=",
"_connect",
"(",
"host",
",",
"port",
",",
"db",
",",
"password",
")",
"return",
"server",
".",
"expireat",
"(",
"key",
",",
"timestamp",
")"
] | set a keys expire at given unix time cli example: . | train | true |
19,910 | def logparser(registry, xml_parent, data):
clog = XML.SubElement(xml_parent, 'hudson.plugins.logparser.LogParserPublisher')
clog.set('plugin', 'log-parser')
mappings = [('unstable-on-warning', 'unstableOnWarning', False), ('fail-on-error', 'failBuildOnError', False), ('parse-rules', 'parsingRulesPath', '')]
helpers.convert_mapping_to_xml(clog, data, mappings, fail_required=True)
| [
"def",
"logparser",
"(",
"registry",
",",
"xml_parent",
",",
"data",
")",
":",
"clog",
"=",
"XML",
".",
"SubElement",
"(",
"xml_parent",
",",
"'hudson.plugins.logparser.LogParserPublisher'",
")",
"clog",
".",
"set",
"(",
"'plugin'",
",",
"'log-parser'",
")",
"mappings",
"=",
"[",
"(",
"'unstable-on-warning'",
",",
"'unstableOnWarning'",
",",
"False",
")",
",",
"(",
"'fail-on-error'",
",",
"'failBuildOnError'",
",",
"False",
")",
",",
"(",
"'parse-rules'",
",",
"'parsingRulesPath'",
",",
"''",
")",
"]",
"helpers",
".",
"convert_mapping_to_xml",
"(",
"clog",
",",
"data",
",",
"mappings",
",",
"fail_required",
"=",
"True",
")"
] | yaml: logparser requires the jenkins :jenkins-wiki:log parser plugin <log+parser+plugin> . | train | false |
19,914 | def has_known_categories(x):
x = getattr(x, '_meta', x)
if isinstance(x, pd.Series):
return (UNKNOWN_CATEGORIES not in x.cat.categories)
elif isinstance(x, pd.CategoricalIndex):
return (UNKNOWN_CATEGORIES not in x.categories)
raise TypeError('Expected Series or CategoricalIndex')
| [
"def",
"has_known_categories",
"(",
"x",
")",
":",
"x",
"=",
"getattr",
"(",
"x",
",",
"'_meta'",
",",
"x",
")",
"if",
"isinstance",
"(",
"x",
",",
"pd",
".",
"Series",
")",
":",
"return",
"(",
"UNKNOWN_CATEGORIES",
"not",
"in",
"x",
".",
"cat",
".",
"categories",
")",
"elif",
"isinstance",
"(",
"x",
",",
"pd",
".",
"CategoricalIndex",
")",
":",
"return",
"(",
"UNKNOWN_CATEGORIES",
"not",
"in",
"x",
".",
"categories",
")",
"raise",
"TypeError",
"(",
"'Expected Series or CategoricalIndex'",
")"
] | returns whether the categories in x are known . | train | false |
19,915 | def getNewRepository():
return ExportRepository()
| [
"def",
"getNewRepository",
"(",
")",
":",
"return",
"ExportRepository",
"(",
")"
] | get new repository . | train | false |
19,916 | def is_gentoo():
return os.path.exists('/etc/gentoo-release')
| [
"def",
"is_gentoo",
"(",
")",
":",
"return",
"os",
".",
"path",
".",
"exists",
"(",
"'/etc/gentoo-release'",
")"
] | checks if were running on gentoo . | train | false |
19,918 | def getInsetPoint(loop, tinyRadius):
pointIndex = getWideAnglePointIndex(loop)
point = loop[(pointIndex % len(loop))]
afterPoint = loop[((pointIndex + 1) % len(loop))]
beforePoint = loop[((pointIndex - 1) % len(loop))]
afterSegmentNormalized = euclidean.getNormalized((afterPoint - point))
beforeSegmentNormalized = euclidean.getNormalized((beforePoint - point))
afterClockwise = complex(afterSegmentNormalized.imag, (- afterSegmentNormalized.real))
beforeWiddershins = complex((- beforeSegmentNormalized.imag), beforeSegmentNormalized.real)
midpoint = (afterClockwise + beforeWiddershins)
midpointNormalized = (midpoint / abs(midpoint))
return (point + (midpointNormalized * tinyRadius))
| [
"def",
"getInsetPoint",
"(",
"loop",
",",
"tinyRadius",
")",
":",
"pointIndex",
"=",
"getWideAnglePointIndex",
"(",
"loop",
")",
"point",
"=",
"loop",
"[",
"(",
"pointIndex",
"%",
"len",
"(",
"loop",
")",
")",
"]",
"afterPoint",
"=",
"loop",
"[",
"(",
"(",
"pointIndex",
"+",
"1",
")",
"%",
"len",
"(",
"loop",
")",
")",
"]",
"beforePoint",
"=",
"loop",
"[",
"(",
"(",
"pointIndex",
"-",
"1",
")",
"%",
"len",
"(",
"loop",
")",
")",
"]",
"afterSegmentNormalized",
"=",
"euclidean",
".",
"getNormalized",
"(",
"(",
"afterPoint",
"-",
"point",
")",
")",
"beforeSegmentNormalized",
"=",
"euclidean",
".",
"getNormalized",
"(",
"(",
"beforePoint",
"-",
"point",
")",
")",
"afterClockwise",
"=",
"complex",
"(",
"afterSegmentNormalized",
".",
"imag",
",",
"(",
"-",
"afterSegmentNormalized",
".",
"real",
")",
")",
"beforeWiddershins",
"=",
"complex",
"(",
"(",
"-",
"beforeSegmentNormalized",
".",
"imag",
")",
",",
"beforeSegmentNormalized",
".",
"real",
")",
"midpoint",
"=",
"(",
"afterClockwise",
"+",
"beforeWiddershins",
")",
"midpointNormalized",
"=",
"(",
"midpoint",
"/",
"abs",
"(",
"midpoint",
")",
")",
"return",
"(",
"point",
"+",
"(",
"midpointNormalized",
"*",
"tinyRadius",
")",
")"
] | get the inset vertex . | train | false |
19,921 | def add_html_link(app, pagename, templatename, context, doctree):
base_url = app.config['html_theme_options'].get('base_url', '')
if base_url:
app.sitemap_links.append(((base_url + pagename) + '.html'))
| [
"def",
"add_html_link",
"(",
"app",
",",
"pagename",
",",
"templatename",
",",
"context",
",",
"doctree",
")",
":",
"base_url",
"=",
"app",
".",
"config",
"[",
"'html_theme_options'",
"]",
".",
"get",
"(",
"'base_url'",
",",
"''",
")",
"if",
"base_url",
":",
"app",
".",
"sitemap_links",
".",
"append",
"(",
"(",
"(",
"base_url",
"+",
"pagename",
")",
"+",
"'.html'",
")",
")"
] | as each page is built . | train | true |
19,922 | def check_params(module):
if ((module.params.get('name') is None) and (module.params.get('id') is None)):
module.fail_json(msg='"name" or "id" is required')
| [
"def",
"check_params",
"(",
"module",
")",
":",
"if",
"(",
"(",
"module",
".",
"params",
".",
"get",
"(",
"'name'",
")",
"is",
"None",
")",
"and",
"(",
"module",
".",
"params",
".",
"get",
"(",
"'id'",
")",
"is",
"None",
")",
")",
":",
"module",
".",
"fail_json",
"(",
"msg",
"=",
"'\"name\" or \"id\" is required'",
")"
] | most modules must have either name or id specified . | train | false |
19,923 | def build_discarder(registry, xml_parent, data):
base_sub = XML.SubElement(xml_parent, 'jenkins.model.BuildDiscarderProperty')
strategy = XML.SubElement(base_sub, 'strategy')
strategy.set('class', 'hudson.tasks.LogRotator')
mappings = [('days-to-keep', 'daysToKeep', (-1)), ('num-to-keep', 'numToKeep', (-1)), ('artifact-days-to-keep', 'artifactDaysToKeep', (-1)), ('artifact-num-to-keep', 'artifactNumToKeep', (-1))]
helpers.convert_mapping_to_xml(strategy, data, mappings, fail_required=True)
| [
"def",
"build_discarder",
"(",
"registry",
",",
"xml_parent",
",",
"data",
")",
":",
"base_sub",
"=",
"XML",
".",
"SubElement",
"(",
"xml_parent",
",",
"'jenkins.model.BuildDiscarderProperty'",
")",
"strategy",
"=",
"XML",
".",
"SubElement",
"(",
"base_sub",
",",
"'strategy'",
")",
"strategy",
".",
"set",
"(",
"'class'",
",",
"'hudson.tasks.LogRotator'",
")",
"mappings",
"=",
"[",
"(",
"'days-to-keep'",
",",
"'daysToKeep'",
",",
"(",
"-",
"1",
")",
")",
",",
"(",
"'num-to-keep'",
",",
"'numToKeep'",
",",
"(",
"-",
"1",
")",
")",
",",
"(",
"'artifact-days-to-keep'",
",",
"'artifactDaysToKeep'",
",",
"(",
"-",
"1",
")",
")",
",",
"(",
"'artifact-num-to-keep'",
",",
"'artifactNumToKeep'",
",",
"(",
"-",
"1",
")",
")",
"]",
"helpers",
".",
"convert_mapping_to_xml",
"(",
"strategy",
",",
"data",
",",
"mappings",
",",
"fail_required",
"=",
"True",
")"
] | yaml: build-discarder :arg int days-to-keep: number of days to keep builds for :arg int num-to-keep: number of builds to keep :arg int artifact-days-to-keep: number of days to keep builds with artifacts :arg int artifact-num-to-keep: number of builds with artifacts to keep example: . | train | false |
19,924 | def _renew_by(name, window=None):
expiry = _expires(name)
if (window is not None):
expiry = (expiry - datetime.timedelta(days=window))
return expiry
| [
"def",
"_renew_by",
"(",
"name",
",",
"window",
"=",
"None",
")",
":",
"expiry",
"=",
"_expires",
"(",
"name",
")",
"if",
"(",
"window",
"is",
"not",
"None",
")",
":",
"expiry",
"=",
"(",
"expiry",
"-",
"datetime",
".",
"timedelta",
"(",
"days",
"=",
"window",
")",
")",
"return",
"expiry"
] | date before a certificate should be renewed . | train | true |
19,925 | def from_text(textring):
keyring = {}
for keytext in textring:
keyname = dns.name.from_text(keytext)
secret = base64.decodestring(textring[keytext])
keyring[keyname] = secret
return keyring
| [
"def",
"from_text",
"(",
"textring",
")",
":",
"keyring",
"=",
"{",
"}",
"for",
"keytext",
"in",
"textring",
":",
"keyname",
"=",
"dns",
".",
"name",
".",
"from_text",
"(",
"keytext",
")",
"secret",
"=",
"base64",
".",
"decodestring",
"(",
"textring",
"[",
"keytext",
"]",
")",
"keyring",
"[",
"keyname",
"]",
"=",
"secret",
"return",
"keyring"
] | convert the text form of a ttl to an integer . | train | true |
19,926 | def _igam(a, x):
ax = math.exp((((a * math.log(x)) - x) - math.lgamma(a)))
r = a
c = 1.0
ans = 1.0
while True:
r += 1.0
c *= (x / r)
ans += c
if ((c / ans) <= MACHEP):
return ((ans * ax) / a)
| [
"def",
"_igam",
"(",
"a",
",",
"x",
")",
":",
"ax",
"=",
"math",
".",
"exp",
"(",
"(",
"(",
"(",
"a",
"*",
"math",
".",
"log",
"(",
"x",
")",
")",
"-",
"x",
")",
"-",
"math",
".",
"lgamma",
"(",
"a",
")",
")",
")",
"r",
"=",
"a",
"c",
"=",
"1.0",
"ans",
"=",
"1.0",
"while",
"True",
":",
"r",
"+=",
"1.0",
"c",
"*=",
"(",
"x",
"/",
"r",
")",
"ans",
"+=",
"c",
"if",
"(",
"(",
"c",
"/",
"ans",
")",
"<=",
"MACHEP",
")",
":",
"return",
"(",
"(",
"ans",
"*",
"ax",
")",
"/",
"a",
")"
] | left tail of incomplete gamma function . | train | true |
19,927 | def get_components(principal):
if (not principal):
return None
return re.split('[\\/@]', str(principal))
| [
"def",
"get_components",
"(",
"principal",
")",
":",
"if",
"(",
"not",
"principal",
")",
":",
"return",
"None",
"return",
"re",
".",
"split",
"(",
"'[\\\\/@]'",
",",
"str",
"(",
"principal",
")",
")"
] | get_components -> (short name . | train | false |
19,928 | def stat(pid):
with open(('/proc/%d/stat' % pid)) as fd:
s = fd.read()
i = s.find('(')
j = s.rfind(')')
name = s[(i + 1):j]
return ((s[:i].split() + [name]) + s[(j + 1):].split())
| [
"def",
"stat",
"(",
"pid",
")",
":",
"with",
"open",
"(",
"(",
"'/proc/%d/stat'",
"%",
"pid",
")",
")",
"as",
"fd",
":",
"s",
"=",
"fd",
".",
"read",
"(",
")",
"i",
"=",
"s",
".",
"find",
"(",
"'('",
")",
"j",
"=",
"s",
".",
"rfind",
"(",
"')'",
")",
"name",
"=",
"s",
"[",
"(",
"i",
"+",
"1",
")",
":",
"j",
"]",
"return",
"(",
"(",
"s",
"[",
":",
"i",
"]",
".",
"split",
"(",
")",
"+",
"[",
"name",
"]",
")",
"+",
"s",
"[",
"(",
"j",
"+",
"1",
")",
":",
"]",
".",
"split",
"(",
")",
")"
] | get status of a finalized file given its full path filename . | train | false |
19,929 | def getNetworkOperatorName():
try:
mContext = autoclass('android.content.Context')
pythonActivity = autoclass('org.renpy.android.PythonService')
telephonyManager = cast('android.telephony.TelephonyManager', pythonActivity.mService.getSystemService(mContext.TELEPHONY_SERVICE))
networkOperatorName = telephonyManager.getNetworkOperatorName()
return networkOperatorName
except Exception as e:
return None
| [
"def",
"getNetworkOperatorName",
"(",
")",
":",
"try",
":",
"mContext",
"=",
"autoclass",
"(",
"'android.content.Context'",
")",
"pythonActivity",
"=",
"autoclass",
"(",
"'org.renpy.android.PythonService'",
")",
"telephonyManager",
"=",
"cast",
"(",
"'android.telephony.TelephonyManager'",
",",
"pythonActivity",
".",
"mService",
".",
"getSystemService",
"(",
"mContext",
".",
"TELEPHONY_SERVICE",
")",
")",
"networkOperatorName",
"=",
"telephonyManager",
".",
"getNetworkOperatorName",
"(",
")",
"return",
"networkOperatorName",
"except",
"Exception",
"as",
"e",
":",
"return",
"None"
] | returns the alphabetic name of current registered operator returns none if an error . | train | false |
19,930 | def color_range(startcolor, goalcolor, steps):
start_tuple = make_color_tuple(startcolor)
goal_tuple = make_color_tuple(goalcolor)
return interpolate_tuple(start_tuple, goal_tuple, steps)
| [
"def",
"color_range",
"(",
"startcolor",
",",
"goalcolor",
",",
"steps",
")",
":",
"start_tuple",
"=",
"make_color_tuple",
"(",
"startcolor",
")",
"goal_tuple",
"=",
"make_color_tuple",
"(",
"goalcolor",
")",
"return",
"interpolate_tuple",
"(",
"start_tuple",
",",
"goal_tuple",
",",
"steps",
")"
] | wrapper for interpolate_tuple that accepts colors as html . | train | true |
19,932 | @public
def assemble_partfrac_list(partial_list):
common = partial_list[0]
polypart = partial_list[1]
pfd = polypart.as_expr()
for (r, nf, df, ex) in partial_list[2]:
if isinstance(r, Poly):
(an, nu) = (nf.variables, nf.expr)
(ad, de) = (df.variables, df.expr)
de = de.subs(ad[0], an[0])
func = Lambda(an, (nu / (de ** ex)))
pfd += RootSum(r, func, auto=False, quadratic=False)
else:
for root in r:
pfd += (nf(root) / (df(root) ** ex))
return (common * pfd)
| [
"@",
"public",
"def",
"assemble_partfrac_list",
"(",
"partial_list",
")",
":",
"common",
"=",
"partial_list",
"[",
"0",
"]",
"polypart",
"=",
"partial_list",
"[",
"1",
"]",
"pfd",
"=",
"polypart",
".",
"as_expr",
"(",
")",
"for",
"(",
"r",
",",
"nf",
",",
"df",
",",
"ex",
")",
"in",
"partial_list",
"[",
"2",
"]",
":",
"if",
"isinstance",
"(",
"r",
",",
"Poly",
")",
":",
"(",
"an",
",",
"nu",
")",
"=",
"(",
"nf",
".",
"variables",
",",
"nf",
".",
"expr",
")",
"(",
"ad",
",",
"de",
")",
"=",
"(",
"df",
".",
"variables",
",",
"df",
".",
"expr",
")",
"de",
"=",
"de",
".",
"subs",
"(",
"ad",
"[",
"0",
"]",
",",
"an",
"[",
"0",
"]",
")",
"func",
"=",
"Lambda",
"(",
"an",
",",
"(",
"nu",
"/",
"(",
"de",
"**",
"ex",
")",
")",
")",
"pfd",
"+=",
"RootSum",
"(",
"r",
",",
"func",
",",
"auto",
"=",
"False",
",",
"quadratic",
"=",
"False",
")",
"else",
":",
"for",
"root",
"in",
"r",
":",
"pfd",
"+=",
"(",
"nf",
"(",
"root",
")",
"/",
"(",
"df",
"(",
"root",
")",
"**",
"ex",
")",
")",
"return",
"(",
"common",
"*",
"pfd",
")"
] | reassemble a full partial fraction decomposition from a structured result obtained by the function apart_list . | train | false |
19,933 | @pytest.fixture(scope='module', params=['gpu', 'cpu'])
def backend_default(request):
be = get_backend(request)
def cleanup():
be = request.getfuncargvalue('backend_default')
del be
request.addfinalizer(cleanup)
return be
| [
"@",
"pytest",
".",
"fixture",
"(",
"scope",
"=",
"'module'",
",",
"params",
"=",
"[",
"'gpu'",
",",
"'cpu'",
"]",
")",
"def",
"backend_default",
"(",
"request",
")",
":",
"be",
"=",
"get_backend",
"(",
"request",
")",
"def",
"cleanup",
"(",
")",
":",
"be",
"=",
"request",
".",
"getfuncargvalue",
"(",
"'backend_default'",
")",
"del",
"be",
"request",
".",
"addfinalizer",
"(",
"cleanup",
")",
"return",
"be"
] | fixture to setup the backend before running a test . | train | false |
19,936 | def _safe_isinstance(obj, module, class_name):
return ((module in sys.modules) and isinstance(obj, getattr(import_module(module), class_name)))
| [
"def",
"_safe_isinstance",
"(",
"obj",
",",
"module",
",",
"class_name",
")",
":",
"return",
"(",
"(",
"module",
"in",
"sys",
".",
"modules",
")",
"and",
"isinstance",
"(",
"obj",
",",
"getattr",
"(",
"import_module",
"(",
"module",
")",
",",
"class_name",
")",
")",
")"
] | checks if obj is an instance of module . | train | false |
19,938 | def islink(p):
return _false
| [
"def",
"islink",
"(",
"p",
")",
":",
"return",
"_false"
] | test whether a path is a symbolic link . | train | false |
19,939 | def to_set(obj, encoder):
return set(obj)
| [
"def",
"to_set",
"(",
"obj",
",",
"encoder",
")",
":",
"return",
"set",
"(",
"obj",
")"
] | converts an arbitrary object c{obj} to a c{set} . | train | false |
19,940 | def is_valid_short_number_for_region(short_number, region_dialing_from):
if isinstance(short_number, PhoneNumber):
short_number = national_significant_number(short_number)
metadata = PhoneMetadata.short_metadata_for_region(region_dialing_from)
if (metadata is None):
return False
general_desc = metadata.general_desc
if ((general_desc.national_number_pattern is None) or (not _is_number_matching_desc(short_number, general_desc))):
return False
short_number_desc = metadata.short_code
if (short_number_desc.national_number_pattern is None):
return False
return _is_number_matching_desc(short_number, short_number_desc)
| [
"def",
"is_valid_short_number_for_region",
"(",
"short_number",
",",
"region_dialing_from",
")",
":",
"if",
"isinstance",
"(",
"short_number",
",",
"PhoneNumber",
")",
":",
"short_number",
"=",
"national_significant_number",
"(",
"short_number",
")",
"metadata",
"=",
"PhoneMetadata",
".",
"short_metadata_for_region",
"(",
"region_dialing_from",
")",
"if",
"(",
"metadata",
"is",
"None",
")",
":",
"return",
"False",
"general_desc",
"=",
"metadata",
".",
"general_desc",
"if",
"(",
"(",
"general_desc",
".",
"national_number_pattern",
"is",
"None",
")",
"or",
"(",
"not",
"_is_number_matching_desc",
"(",
"short_number",
",",
"general_desc",
")",
")",
")",
":",
"return",
"False",
"short_number_desc",
"=",
"metadata",
".",
"short_code",
"if",
"(",
"short_number_desc",
".",
"national_number_pattern",
"is",
"None",
")",
":",
"return",
"False",
"return",
"_is_number_matching_desc",
"(",
"short_number",
",",
"short_number_desc",
")"
] | tests whether a short number matches a valid pattern in a region . | train | false |
19,941 | def paypaltime2datetime(s):
return datetime.datetime(*time.strptime(s, PayPalNVP.TIMESTAMP_FORMAT)[:6])
| [
"def",
"paypaltime2datetime",
"(",
"s",
")",
":",
"return",
"datetime",
".",
"datetime",
"(",
"*",
"time",
".",
"strptime",
"(",
"s",
",",
"PayPalNVP",
".",
"TIMESTAMP_FORMAT",
")",
"[",
":",
"6",
"]",
")"
] | convert a paypal time string to a datetime . | train | false |
19,942 | def _correct_trans(t):
t = np.array(t, np.float64)
t[:3, :3] *= t[3, :3][:, np.newaxis]
t[3, :3] = 0.0
assert (t[(3, 3)] == 1.0)
return t
| [
"def",
"_correct_trans",
"(",
"t",
")",
":",
"t",
"=",
"np",
".",
"array",
"(",
"t",
",",
"np",
".",
"float64",
")",
"t",
"[",
":",
"3",
",",
":",
"3",
"]",
"*=",
"t",
"[",
"3",
",",
":",
"3",
"]",
"[",
":",
",",
"np",
".",
"newaxis",
"]",
"t",
"[",
"3",
",",
":",
"3",
"]",
"=",
"0.0",
"assert",
"(",
"t",
"[",
"(",
"3",
",",
"3",
")",
"]",
"==",
"1.0",
")",
"return",
"t"
] | convert to a transformation matrix . | train | false |
19,943 | def _check_frame(d, frame_str):
if (d['coord_frame'] != _str_to_frame[frame_str]):
raise RuntimeError(('dig point must be in %s coordinate frame, got %s' % (frame_str, _frame_to_str[d['coord_frame']])))
| [
"def",
"_check_frame",
"(",
"d",
",",
"frame_str",
")",
":",
"if",
"(",
"d",
"[",
"'coord_frame'",
"]",
"!=",
"_str_to_frame",
"[",
"frame_str",
"]",
")",
":",
"raise",
"RuntimeError",
"(",
"(",
"'dig point must be in %s coordinate frame, got %s'",
"%",
"(",
"frame_str",
",",
"_frame_to_str",
"[",
"d",
"[",
"'coord_frame'",
"]",
"]",
")",
")",
")"
] | helper to check coordinate frames . | train | false |
19,944 | def getFontReader(fontFamily):
fontLower = fontFamily.lower().replace(' ', '_')
global globalFontReaderDictionary
if (fontLower in globalFontReaderDictionary):
return globalFontReaderDictionary[fontLower]
global globalFontFileNames
if (globalFontFileNames == None):
globalFontFileNames = archive.getPluginFileNamesFromDirectoryPath(getFontsDirectoryPath())
if (fontLower not in globalFontFileNames):
print ('Warning, the %s font was not found in the fonts folder, so Gentium Basic Regular will be substituted.' % fontFamily)
fontLower = 'gentium_basic_regular'
fontReader = FontReader(fontLower)
globalFontReaderDictionary[fontLower] = fontReader
return fontReader
| [
"def",
"getFontReader",
"(",
"fontFamily",
")",
":",
"fontLower",
"=",
"fontFamily",
".",
"lower",
"(",
")",
".",
"replace",
"(",
"' '",
",",
"'_'",
")",
"global",
"globalFontReaderDictionary",
"if",
"(",
"fontLower",
"in",
"globalFontReaderDictionary",
")",
":",
"return",
"globalFontReaderDictionary",
"[",
"fontLower",
"]",
"global",
"globalFontFileNames",
"if",
"(",
"globalFontFileNames",
"==",
"None",
")",
":",
"globalFontFileNames",
"=",
"archive",
".",
"getPluginFileNamesFromDirectoryPath",
"(",
"getFontsDirectoryPath",
"(",
")",
")",
"if",
"(",
"fontLower",
"not",
"in",
"globalFontFileNames",
")",
":",
"print",
"(",
"'Warning, the %s font was not found in the fonts folder, so Gentium Basic Regular will be substituted.'",
"%",
"fontFamily",
")",
"fontLower",
"=",
"'gentium_basic_regular'",
"fontReader",
"=",
"FontReader",
"(",
"fontLower",
")",
"globalFontReaderDictionary",
"[",
"fontLower",
"]",
"=",
"fontReader",
"return",
"fontReader"
] | get the font reader for the fontfamily . | train | false |
19,946 | def _pivot_col(T, tol=1e-12, bland=False):
ma = np.ma.masked_where((T[(-1), :(-1)] >= (- tol)), T[(-1), :(-1)], copy=False)
if (ma.count() == 0):
return (False, np.nan)
if bland:
return (True, np.where((ma.mask == False))[0][0])
return (True, np.ma.where((ma == ma.min()))[0][0])
| [
"def",
"_pivot_col",
"(",
"T",
",",
"tol",
"=",
"1e-12",
",",
"bland",
"=",
"False",
")",
":",
"ma",
"=",
"np",
".",
"ma",
".",
"masked_where",
"(",
"(",
"T",
"[",
"(",
"-",
"1",
")",
",",
":",
"(",
"-",
"1",
")",
"]",
">=",
"(",
"-",
"tol",
")",
")",
",",
"T",
"[",
"(",
"-",
"1",
")",
",",
":",
"(",
"-",
"1",
")",
"]",
",",
"copy",
"=",
"False",
")",
"if",
"(",
"ma",
".",
"count",
"(",
")",
"==",
"0",
")",
":",
"return",
"(",
"False",
",",
"np",
".",
"nan",
")",
"if",
"bland",
":",
"return",
"(",
"True",
",",
"np",
".",
"where",
"(",
"(",
"ma",
".",
"mask",
"==",
"False",
")",
")",
"[",
"0",
"]",
"[",
"0",
"]",
")",
"return",
"(",
"True",
",",
"np",
".",
"ma",
".",
"where",
"(",
"(",
"ma",
"==",
"ma",
".",
"min",
"(",
")",
")",
")",
"[",
"0",
"]",
"[",
"0",
"]",
")"
] | given a linear programming simplex tableau . | train | false |
19,947 | def encompasses_broadcastable(b1, b2):
if (len(b1) < len(b2)):
return False
b1 = b1[(- len(b2)):]
return (not any(((v1 and (not v2)) for (v1, v2) in zip(b1, b2))))
| [
"def",
"encompasses_broadcastable",
"(",
"b1",
",",
"b2",
")",
":",
"if",
"(",
"len",
"(",
"b1",
")",
"<",
"len",
"(",
"b2",
")",
")",
":",
"return",
"False",
"b1",
"=",
"b1",
"[",
"(",
"-",
"len",
"(",
"b2",
")",
")",
":",
"]",
"return",
"(",
"not",
"any",
"(",
"(",
"(",
"v1",
"and",
"(",
"not",
"v2",
")",
")",
"for",
"(",
"v1",
",",
"v2",
")",
"in",
"zip",
"(",
"b1",
",",
"b2",
")",
")",
")",
")"
] | parameters b1 the broadcastable attribute of a tensor type . | train | false |
19,948 | def user_input(prompt=''):
if (PY_MAJOR_VERSION > 2):
return input(prompt)
else:
return raw_input(prompt)
| [
"def",
"user_input",
"(",
"prompt",
"=",
"''",
")",
":",
"if",
"(",
"PY_MAJOR_VERSION",
">",
"2",
")",
":",
"return",
"input",
"(",
"prompt",
")",
"else",
":",
"return",
"raw_input",
"(",
"prompt",
")"
] | for getting raw user input in python 2 and 3 . | train | false |
19,950 | @yield_fixture
def no_patience(app):
with mock.patch.dict(app.tornado_application.settings, {'slow_spawn_timeout': 0, 'slow_stop_timeout': 0}):
(yield)
| [
"@",
"yield_fixture",
"def",
"no_patience",
"(",
"app",
")",
":",
"with",
"mock",
".",
"patch",
".",
"dict",
"(",
"app",
".",
"tornado_application",
".",
"settings",
",",
"{",
"'slow_spawn_timeout'",
":",
"0",
",",
"'slow_stop_timeout'",
":",
"0",
"}",
")",
":",
"(",
"yield",
")"
] | set slow-spawning timeouts to zero . | train | false |
19,952 | @frappe.whitelist()
def make_mapped_doc(method, source_name, selected_children=None):
method = frappe.get_attr(method)
if (method not in frappe.whitelisted):
raise frappe.PermissionError
if selected_children:
selected_children = json.loads(selected_children)
frappe.flags.selected_children = (selected_children or None)
return method(source_name)
| [
"@",
"frappe",
".",
"whitelist",
"(",
")",
"def",
"make_mapped_doc",
"(",
"method",
",",
"source_name",
",",
"selected_children",
"=",
"None",
")",
":",
"method",
"=",
"frappe",
".",
"get_attr",
"(",
"method",
")",
"if",
"(",
"method",
"not",
"in",
"frappe",
".",
"whitelisted",
")",
":",
"raise",
"frappe",
".",
"PermissionError",
"if",
"selected_children",
":",
"selected_children",
"=",
"json",
".",
"loads",
"(",
"selected_children",
")",
"frappe",
".",
"flags",
".",
"selected_children",
"=",
"(",
"selected_children",
"or",
"None",
")",
"return",
"method",
"(",
"source_name",
")"
] | returns the mapped document calling the given mapper method . | train | false |
19,953 | def rename_renewal_config(prev_name, new_name, cli_config):
prev_filename = renewal_filename_for_lineagename(cli_config, prev_name)
new_filename = renewal_filename_for_lineagename(cli_config, new_name)
if os.path.exists(new_filename):
raise errors.ConfigurationError('The new certificate name is already in use.')
try:
os.rename(prev_filename, new_filename)
except OSError:
raise errors.ConfigurationError('Please specify a valid filename for the new certificate name.')
| [
"def",
"rename_renewal_config",
"(",
"prev_name",
",",
"new_name",
",",
"cli_config",
")",
":",
"prev_filename",
"=",
"renewal_filename_for_lineagename",
"(",
"cli_config",
",",
"prev_name",
")",
"new_filename",
"=",
"renewal_filename_for_lineagename",
"(",
"cli_config",
",",
"new_name",
")",
"if",
"os",
".",
"path",
".",
"exists",
"(",
"new_filename",
")",
":",
"raise",
"errors",
".",
"ConfigurationError",
"(",
"'The new certificate name is already in use.'",
")",
"try",
":",
"os",
".",
"rename",
"(",
"prev_filename",
",",
"new_filename",
")",
"except",
"OSError",
":",
"raise",
"errors",
".",
"ConfigurationError",
"(",
"'Please specify a valid filename for the new certificate name.'",
")"
] | renames cli_config . | train | false |
19,954 | def test_user_statement_on_import():
s = u('from datetime import (\n time)')
for pos in [(2, 1), (2, 4)]:
p = ParserWithRecovery(load_grammar(), s)
stmt = p.module.get_statement_for_position(pos)
assert isinstance(stmt, pt.Import)
assert ([str(n) for n in stmt.get_defined_names()] == ['time'])
| [
"def",
"test_user_statement_on_import",
"(",
")",
":",
"s",
"=",
"u",
"(",
"'from datetime import (\\n time)'",
")",
"for",
"pos",
"in",
"[",
"(",
"2",
",",
"1",
")",
",",
"(",
"2",
",",
"4",
")",
"]",
":",
"p",
"=",
"ParserWithRecovery",
"(",
"load_grammar",
"(",
")",
",",
"s",
")",
"stmt",
"=",
"p",
".",
"module",
".",
"get_statement_for_position",
"(",
"pos",
")",
"assert",
"isinstance",
"(",
"stmt",
",",
"pt",
".",
"Import",
")",
"assert",
"(",
"[",
"str",
"(",
"n",
")",
"for",
"n",
"in",
"stmt",
".",
"get_defined_names",
"(",
")",
"]",
"==",
"[",
"'time'",
"]",
")"
] | github #285 . | train | false |
19,955 | def plot_epochs(epochs, picks=None, scalings=None, n_epochs=20, n_channels=20, title=None, events=None, event_colors=None, show=True, block=False):
epochs.drop_bad()
scalings = _compute_scalings(scalings, epochs)
scalings = _handle_default('scalings_plot_raw', scalings)
projs = epochs.info['projs']
params = {'epochs': epochs, 'info': copy.deepcopy(epochs.info), 'bad_color': (0.8, 0.8, 0.8), 't_start': 0, 'histogram': None}
params['label_click_fun'] = partial(_pick_bad_channels, params=params)
_prepare_mne_browse_epochs(params, projs, n_channels, n_epochs, scalings, title, picks, events=events, event_colors=event_colors)
_prepare_projectors(params)
_layout_figure(params)
callback_close = partial(_close_event, params=params)
params['fig'].canvas.mpl_connect('close_event', callback_close)
try:
plt_show(show, block=block)
except TypeError:
plt_show(show)
return params['fig']
| [
"def",
"plot_epochs",
"(",
"epochs",
",",
"picks",
"=",
"None",
",",
"scalings",
"=",
"None",
",",
"n_epochs",
"=",
"20",
",",
"n_channels",
"=",
"20",
",",
"title",
"=",
"None",
",",
"events",
"=",
"None",
",",
"event_colors",
"=",
"None",
",",
"show",
"=",
"True",
",",
"block",
"=",
"False",
")",
":",
"epochs",
".",
"drop_bad",
"(",
")",
"scalings",
"=",
"_compute_scalings",
"(",
"scalings",
",",
"epochs",
")",
"scalings",
"=",
"_handle_default",
"(",
"'scalings_plot_raw'",
",",
"scalings",
")",
"projs",
"=",
"epochs",
".",
"info",
"[",
"'projs'",
"]",
"params",
"=",
"{",
"'epochs'",
":",
"epochs",
",",
"'info'",
":",
"copy",
".",
"deepcopy",
"(",
"epochs",
".",
"info",
")",
",",
"'bad_color'",
":",
"(",
"0.8",
",",
"0.8",
",",
"0.8",
")",
",",
"'t_start'",
":",
"0",
",",
"'histogram'",
":",
"None",
"}",
"params",
"[",
"'label_click_fun'",
"]",
"=",
"partial",
"(",
"_pick_bad_channels",
",",
"params",
"=",
"params",
")",
"_prepare_mne_browse_epochs",
"(",
"params",
",",
"projs",
",",
"n_channels",
",",
"n_epochs",
",",
"scalings",
",",
"title",
",",
"picks",
",",
"events",
"=",
"events",
",",
"event_colors",
"=",
"event_colors",
")",
"_prepare_projectors",
"(",
"params",
")",
"_layout_figure",
"(",
"params",
")",
"callback_close",
"=",
"partial",
"(",
"_close_event",
",",
"params",
"=",
"params",
")",
"params",
"[",
"'fig'",
"]",
".",
"canvas",
".",
"mpl_connect",
"(",
"'close_event'",
",",
"callback_close",
")",
"try",
":",
"plt_show",
"(",
"show",
",",
"block",
"=",
"block",
")",
"except",
"TypeError",
":",
"plt_show",
"(",
"show",
")",
"return",
"params",
"[",
"'fig'",
"]"
] | visualize epochs . | train | false |
19,956 | @intercept_errors(UserAPIInternalError, ignore_errors=[UserAPIRequestError])
def activate_account(activation_key):
try:
registration = Registration.objects.get(activation_key=activation_key)
except Registration.DoesNotExist:
raise UserNotAuthorized
else:
registration.activate()
| [
"@",
"intercept_errors",
"(",
"UserAPIInternalError",
",",
"ignore_errors",
"=",
"[",
"UserAPIRequestError",
"]",
")",
"def",
"activate_account",
"(",
"activation_key",
")",
":",
"try",
":",
"registration",
"=",
"Registration",
".",
"objects",
".",
"get",
"(",
"activation_key",
"=",
"activation_key",
")",
"except",
"Registration",
".",
"DoesNotExist",
":",
"raise",
"UserNotAuthorized",
"else",
":",
"registration",
".",
"activate",
"(",
")"
] | when link in activation e-mail is clicked . | train | false |
19,957 | @gof.local_optimizer([SparseBlockOuter], inplace=True)
def local_inplace_sparse_block_outer(node):
if (isinstance(node.op, SparseBlockOuter) and (not node.op.inplace)):
new_node = sparse_block_outer_inplace(*node.inputs)
copy_stack_trace(node.outputs[0], new_node)
return [new_node]
return False
| [
"@",
"gof",
".",
"local_optimizer",
"(",
"[",
"SparseBlockOuter",
"]",
",",
"inplace",
"=",
"True",
")",
"def",
"local_inplace_sparse_block_outer",
"(",
"node",
")",
":",
"if",
"(",
"isinstance",
"(",
"node",
".",
"op",
",",
"SparseBlockOuter",
")",
"and",
"(",
"not",
"node",
".",
"op",
".",
"inplace",
")",
")",
":",
"new_node",
"=",
"sparse_block_outer_inplace",
"(",
"*",
"node",
".",
"inputs",
")",
"copy_stack_trace",
"(",
"node",
".",
"outputs",
"[",
"0",
"]",
",",
"new_node",
")",
"return",
"[",
"new_node",
"]",
"return",
"False"
] | sparseblockouter -> sparseblockouter . | train | false |
19,959 | def get_metrics_instance():
from stream_framework import settings
metric_cls = get_class_from_string(settings.STREAM_METRIC_CLASS)
return metric_cls(**settings.STREAM_METRICS_OPTIONS)
| [
"def",
"get_metrics_instance",
"(",
")",
":",
"from",
"stream_framework",
"import",
"settings",
"metric_cls",
"=",
"get_class_from_string",
"(",
"settings",
".",
"STREAM_METRIC_CLASS",
")",
"return",
"metric_cls",
"(",
"**",
"settings",
".",
"STREAM_METRICS_OPTIONS",
")"
] | returns an instance of the metric class as defined in stream_framework settings . | train | false |
19,961 | def get_travis_directories(package_list):
if in_travis_pr():
pr_against_branch = travis_branch()
return get_changed_packages('HEAD', pr_against_branch, package_list)
else:
return package_list
| [
"def",
"get_travis_directories",
"(",
"package_list",
")",
":",
"if",
"in_travis_pr",
"(",
")",
":",
"pr_against_branch",
"=",
"travis_branch",
"(",
")",
"return",
"get_changed_packages",
"(",
"'HEAD'",
",",
"pr_against_branch",
",",
"package_list",
")",
"else",
":",
"return",
"package_list"
] | get list of packages that need to be tested on travis ci . | train | false |
19,962 | def test_append_private_mode(hist, config_stub):
hist.handle_private_mode = True
config_stub.data = CONFIG_PRIVATE
hist.append('new item')
assert (hist.history == HISTORY)
| [
"def",
"test_append_private_mode",
"(",
"hist",
",",
"config_stub",
")",
":",
"hist",
".",
"handle_private_mode",
"=",
"True",
"config_stub",
".",
"data",
"=",
"CONFIG_PRIVATE",
"hist",
".",
"append",
"(",
"'new item'",
")",
"assert",
"(",
"hist",
".",
"history",
"==",
"HISTORY",
")"
] | test append in private mode . | train | false |
19,963 | def test_hierarchical():
(raw, events, picks) = _get_data()
event_id = {'a/1': 1, 'a/2': 2, 'b/1': 3, 'b/2': 4}
epochs = Epochs(raw, events, event_id, preload=True)
epochs_a1 = epochs['a/1']
epochs_a2 = epochs['a/2']
epochs_b1 = epochs['b/1']
epochs_b2 = epochs['b/2']
epochs_a = epochs['a']
assert_equal(len(epochs_a), (len(epochs_a1) + len(epochs_a2)))
epochs_b = epochs['b']
assert_equal(len(epochs_b), (len(epochs_b1) + len(epochs_b2)))
epochs_1 = epochs['1']
assert_equal(len(epochs_1), (len(epochs_a1) + len(epochs_b1)))
epochs_2 = epochs['2']
assert_equal(len(epochs_2), (len(epochs_a2) + len(epochs_b2)))
epochs_all = epochs[('1', '2')]
assert_equal(len(epochs), len(epochs_all))
assert_array_equal(epochs.get_data(), epochs_all.get_data())
| [
"def",
"test_hierarchical",
"(",
")",
":",
"(",
"raw",
",",
"events",
",",
"picks",
")",
"=",
"_get_data",
"(",
")",
"event_id",
"=",
"{",
"'a/1'",
":",
"1",
",",
"'a/2'",
":",
"2",
",",
"'b/1'",
":",
"3",
",",
"'b/2'",
":",
"4",
"}",
"epochs",
"=",
"Epochs",
"(",
"raw",
",",
"events",
",",
"event_id",
",",
"preload",
"=",
"True",
")",
"epochs_a1",
"=",
"epochs",
"[",
"'a/1'",
"]",
"epochs_a2",
"=",
"epochs",
"[",
"'a/2'",
"]",
"epochs_b1",
"=",
"epochs",
"[",
"'b/1'",
"]",
"epochs_b2",
"=",
"epochs",
"[",
"'b/2'",
"]",
"epochs_a",
"=",
"epochs",
"[",
"'a'",
"]",
"assert_equal",
"(",
"len",
"(",
"epochs_a",
")",
",",
"(",
"len",
"(",
"epochs_a1",
")",
"+",
"len",
"(",
"epochs_a2",
")",
")",
")",
"epochs_b",
"=",
"epochs",
"[",
"'b'",
"]",
"assert_equal",
"(",
"len",
"(",
"epochs_b",
")",
",",
"(",
"len",
"(",
"epochs_b1",
")",
"+",
"len",
"(",
"epochs_b2",
")",
")",
")",
"epochs_1",
"=",
"epochs",
"[",
"'1'",
"]",
"assert_equal",
"(",
"len",
"(",
"epochs_1",
")",
",",
"(",
"len",
"(",
"epochs_a1",
")",
"+",
"len",
"(",
"epochs_b1",
")",
")",
")",
"epochs_2",
"=",
"epochs",
"[",
"'2'",
"]",
"assert_equal",
"(",
"len",
"(",
"epochs_2",
")",
",",
"(",
"len",
"(",
"epochs_a2",
")",
"+",
"len",
"(",
"epochs_b2",
")",
")",
")",
"epochs_all",
"=",
"epochs",
"[",
"(",
"'1'",
",",
"'2'",
")",
"]",
"assert_equal",
"(",
"len",
"(",
"epochs",
")",
",",
"len",
"(",
"epochs_all",
")",
")",
"assert_array_equal",
"(",
"epochs",
".",
"get_data",
"(",
")",
",",
"epochs_all",
".",
"get_data",
"(",
")",
")"
] | test hierarchical access . | train | false |
19,964 | def hmean(a, axis=0, dtype=None):
if (not isinstance(a, np.ndarray)):
a = np.array(a, dtype=dtype)
if np.all((a > 0)):
if isinstance(a, np.ma.MaskedArray):
size = a.count(axis)
elif (axis is None):
a = a.ravel()
size = a.shape[0]
else:
size = a.shape[axis]
return (size / np.sum((1.0 / a), axis=axis, dtype=dtype))
else:
raise ValueError('Harmonic mean only defined if all elements greater than zero')
| [
"def",
"hmean",
"(",
"a",
",",
"axis",
"=",
"0",
",",
"dtype",
"=",
"None",
")",
":",
"if",
"(",
"not",
"isinstance",
"(",
"a",
",",
"np",
".",
"ndarray",
")",
")",
":",
"a",
"=",
"np",
".",
"array",
"(",
"a",
",",
"dtype",
"=",
"dtype",
")",
"if",
"np",
".",
"all",
"(",
"(",
"a",
">",
"0",
")",
")",
":",
"if",
"isinstance",
"(",
"a",
",",
"np",
".",
"ma",
".",
"MaskedArray",
")",
":",
"size",
"=",
"a",
".",
"count",
"(",
"axis",
")",
"elif",
"(",
"axis",
"is",
"None",
")",
":",
"a",
"=",
"a",
".",
"ravel",
"(",
")",
"size",
"=",
"a",
".",
"shape",
"[",
"0",
"]",
"else",
":",
"size",
"=",
"a",
".",
"shape",
"[",
"axis",
"]",
"return",
"(",
"size",
"/",
"np",
".",
"sum",
"(",
"(",
"1.0",
"/",
"a",
")",
",",
"axis",
"=",
"axis",
",",
"dtype",
"=",
"dtype",
")",
")",
"else",
":",
"raise",
"ValueError",
"(",
"'Harmonic mean only defined if all elements greater than zero'",
")"
] | returns the harmonic mean of the given list of values . | train | false |
19,965 | def _fetch_latest_from_datastore(app_version):
rpc = db.create_rpc(deadline=DATASTORE_DEADLINE, read_policy=db.EVENTUAL_CONSISTENCY)
key = _get_active_config_key(app_version)
config = None
try:
config = Config.get(key, rpc=rpc)
logging.debug('Loaded most recent conf data from datastore.')
except:
logging.warning('Tried but failed to fetch latest conf data from the datastore.')
if config:
memcache.set(app_version, db.model_to_protobuf(config).Encode(), namespace=NAMESPACE)
logging.debug('Wrote most recent conf data into memcache.')
return config
| [
"def",
"_fetch_latest_from_datastore",
"(",
"app_version",
")",
":",
"rpc",
"=",
"db",
".",
"create_rpc",
"(",
"deadline",
"=",
"DATASTORE_DEADLINE",
",",
"read_policy",
"=",
"db",
".",
"EVENTUAL_CONSISTENCY",
")",
"key",
"=",
"_get_active_config_key",
"(",
"app_version",
")",
"config",
"=",
"None",
"try",
":",
"config",
"=",
"Config",
".",
"get",
"(",
"key",
",",
"rpc",
"=",
"rpc",
")",
"logging",
".",
"debug",
"(",
"'Loaded most recent conf data from datastore.'",
")",
"except",
":",
"logging",
".",
"warning",
"(",
"'Tried but failed to fetch latest conf data from the datastore.'",
")",
"if",
"config",
":",
"memcache",
".",
"set",
"(",
"app_version",
",",
"db",
".",
"model_to_protobuf",
"(",
"config",
")",
".",
"Encode",
"(",
")",
",",
"namespace",
"=",
"NAMESPACE",
")",
"logging",
".",
"debug",
"(",
"'Wrote most recent conf data into memcache.'",
")",
"return",
"config"
] | get the latest configuration data for this app-version from the datastore . | train | false |
19,966 | def GroupEnum():
nmembers = 0
resume = 0
while 1:
(data, total, resume) = win32net.NetGroupEnum(server, 1, resume)
for group in data:
verbose(('Found group %(name)s:%(comment)s ' % group))
memberresume = 0
while 1:
(memberdata, total, memberresume) = win32net.NetGroupGetUsers(server, group['name'], 0, resume)
for member in memberdata:
verbose((' Member %(name)s' % member))
nmembers = (nmembers + 1)
if (memberresume == 0):
break
if (not resume):
break
assert nmembers, 'Couldnt find a single member in a single group!'
print 'Enumerated all the groups'
| [
"def",
"GroupEnum",
"(",
")",
":",
"nmembers",
"=",
"0",
"resume",
"=",
"0",
"while",
"1",
":",
"(",
"data",
",",
"total",
",",
"resume",
")",
"=",
"win32net",
".",
"NetGroupEnum",
"(",
"server",
",",
"1",
",",
"resume",
")",
"for",
"group",
"in",
"data",
":",
"verbose",
"(",
"(",
"'Found group %(name)s:%(comment)s '",
"%",
"group",
")",
")",
"memberresume",
"=",
"0",
"while",
"1",
":",
"(",
"memberdata",
",",
"total",
",",
"memberresume",
")",
"=",
"win32net",
".",
"NetGroupGetUsers",
"(",
"server",
",",
"group",
"[",
"'name'",
"]",
",",
"0",
",",
"resume",
")",
"for",
"member",
"in",
"memberdata",
":",
"verbose",
"(",
"(",
"' Member %(name)s'",
"%",
"member",
")",
")",
"nmembers",
"=",
"(",
"nmembers",
"+",
"1",
")",
"if",
"(",
"memberresume",
"==",
"0",
")",
":",
"break",
"if",
"(",
"not",
"resume",
")",
":",
"break",
"assert",
"nmembers",
",",
"'Couldnt find a single member in a single group!'",
"print",
"'Enumerated all the groups'"
] | enumerates all the domain groups . | train | false |
19,967 | def safe_lower(txt):
if txt:
return txt.lower()
else:
return ''
| [
"def",
"safe_lower",
"(",
"txt",
")",
":",
"if",
"txt",
":",
"return",
"txt",
".",
"lower",
"(",
")",
"else",
":",
"return",
"''"
] | return lowercased string . | train | false |
19,968 | def certificate_get_all_by_user(context, user_id):
return IMPL.certificate_get_all_by_user(context, user_id)
| [
"def",
"certificate_get_all_by_user",
"(",
"context",
",",
"user_id",
")",
":",
"return",
"IMPL",
".",
"certificate_get_all_by_user",
"(",
"context",
",",
"user_id",
")"
] | get all certificates for a user . | train | false |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.