id_within_dataset int64 1 55.5k | snippet stringlengths 19 14.2k | tokens listlengths 6 1.63k | nl stringlengths 6 352 | split_within_dataset stringclasses 1 value | is_duplicated bool 2 classes |
|---|---|---|---|---|---|
25,142 | def on_valid(valid_content_type, on_invalid=json):
invalid_kwargs = introspect.generate_accepted_kwargs(on_invalid, 'request', 'response')
invalid_takes_response = introspect.takes_all_arguments(on_invalid, 'response')
def wrapper(function):
valid_kwargs = introspect.generate_accepted_kwargs(function, 'request', 'response')
valid_takes_response = introspect.takes_all_arguments(function, 'response')
@content_type(valid_content_type)
@wraps(function)
def output_content(content, response, **kwargs):
if ((type(content) == dict) and ('errors' in content)):
response.content_type = on_invalid.content_type
if invalid_takes_response:
kwargs['response'] = response
return on_invalid(content, **invalid_kwargs(kwargs))
if valid_takes_response:
kwargs['response'] = response
return function(content, **valid_kwargs(kwargs))
return output_content
return wrapper
| [
"def",
"on_valid",
"(",
"valid_content_type",
",",
"on_invalid",
"=",
"json",
")",
":",
"invalid_kwargs",
"=",
"introspect",
".",
"generate_accepted_kwargs",
"(",
"on_invalid",
",",
"'request'",
",",
"'response'",
")",
"invalid_takes_response",
"=",
"introspect",
".",
"takes_all_arguments",
"(",
"on_invalid",
",",
"'response'",
")",
"def",
"wrapper",
"(",
"function",
")",
":",
"valid_kwargs",
"=",
"introspect",
".",
"generate_accepted_kwargs",
"(",
"function",
",",
"'request'",
",",
"'response'",
")",
"valid_takes_response",
"=",
"introspect",
".",
"takes_all_arguments",
"(",
"function",
",",
"'response'",
")",
"@",
"content_type",
"(",
"valid_content_type",
")",
"@",
"wraps",
"(",
"function",
")",
"def",
"output_content",
"(",
"content",
",",
"response",
",",
"**",
"kwargs",
")",
":",
"if",
"(",
"(",
"type",
"(",
"content",
")",
"==",
"dict",
")",
"and",
"(",
"'errors'",
"in",
"content",
")",
")",
":",
"response",
".",
"content_type",
"=",
"on_invalid",
".",
"content_type",
"if",
"invalid_takes_response",
":",
"kwargs",
"[",
"'response'",
"]",
"=",
"response",
"return",
"on_invalid",
"(",
"content",
",",
"**",
"invalid_kwargs",
"(",
"kwargs",
")",
")",
"if",
"valid_takes_response",
":",
"kwargs",
"[",
"'response'",
"]",
"=",
"response",
"return",
"function",
"(",
"content",
",",
"**",
"valid_kwargs",
"(",
"kwargs",
")",
")",
"return",
"output_content",
"return",
"wrapper"
] | renders as the specified content type only if no errors are found in the provided data object . | train | true |
25,143 | def getManipulatedPaths(close, elementNode, loop, prefix, sideLength):
if (len(loop) < 1):
return [[]]
derivation = BottomDerivation(elementNode, prefix)
targetMatrix = matrix.getBranchMatrixSetElementNode(elementNode)
transformedLoop = matrix.getTransformedVector3s(matrix.getIdentityTetragrid(targetMatrix.tetragrid), loop)
lift = ((derivation.altitude + derivation.getAdditionalPathLift()) - euclidean.getBottomByPath(transformedLoop))
for point in loop:
point.z += lift
return [loop]
| [
"def",
"getManipulatedPaths",
"(",
"close",
",",
"elementNode",
",",
"loop",
",",
"prefix",
",",
"sideLength",
")",
":",
"if",
"(",
"len",
"(",
"loop",
")",
"<",
"1",
")",
":",
"return",
"[",
"[",
"]",
"]",
"derivation",
"=",
"BottomDerivation",
"(",
"elementNode",
",",
"prefix",
")",
"targetMatrix",
"=",
"matrix",
".",
"getBranchMatrixSetElementNode",
"(",
"elementNode",
")",
"transformedLoop",
"=",
"matrix",
".",
"getTransformedVector3s",
"(",
"matrix",
".",
"getIdentityTetragrid",
"(",
"targetMatrix",
".",
"tetragrid",
")",
",",
"loop",
")",
"lift",
"=",
"(",
"(",
"derivation",
".",
"altitude",
"+",
"derivation",
".",
"getAdditionalPathLift",
"(",
")",
")",
"-",
"euclidean",
".",
"getBottomByPath",
"(",
"transformedLoop",
")",
")",
"for",
"point",
"in",
"loop",
":",
"point",
".",
"z",
"+=",
"lift",
"return",
"[",
"loop",
"]"
] | get round loop . | train | false |
25,144 | def create_comm_note(obj, version, author, body, note_type=comm.NO_ACTION, perms=None, attachments=None):
perms = (perms or {})
if (('developer' not in perms) and (note_type in comm.REVIEWER_NOTE_TYPES)):
perms['developer'] = False
create_perms = dict(((('read_permission_%s' % key), has_perm) for (key, has_perm) in perms.iteritems()))
version_param = {}
if (obj.__class__ == Webapp):
version_param['_version'] = version
elif (obj.__class__ == Extension):
version_param['_extension_version'] = version
(thread, created_thread) = obj.threads.safer_get_or_create(defaults=create_perms, **version_param)
note = thread.notes.create(note_type=note_type, body=body, author=author, **create_perms)
if attachments:
create_attachments(note, attachments)
post_create_comm_note(note)
return (thread, note)
| [
"def",
"create_comm_note",
"(",
"obj",
",",
"version",
",",
"author",
",",
"body",
",",
"note_type",
"=",
"comm",
".",
"NO_ACTION",
",",
"perms",
"=",
"None",
",",
"attachments",
"=",
"None",
")",
":",
"perms",
"=",
"(",
"perms",
"or",
"{",
"}",
")",
"if",
"(",
"(",
"'developer'",
"not",
"in",
"perms",
")",
"and",
"(",
"note_type",
"in",
"comm",
".",
"REVIEWER_NOTE_TYPES",
")",
")",
":",
"perms",
"[",
"'developer'",
"]",
"=",
"False",
"create_perms",
"=",
"dict",
"(",
"(",
"(",
"(",
"'read_permission_%s'",
"%",
"key",
")",
",",
"has_perm",
")",
"for",
"(",
"key",
",",
"has_perm",
")",
"in",
"perms",
".",
"iteritems",
"(",
")",
")",
")",
"version_param",
"=",
"{",
"}",
"if",
"(",
"obj",
".",
"__class__",
"==",
"Webapp",
")",
":",
"version_param",
"[",
"'_version'",
"]",
"=",
"version",
"elif",
"(",
"obj",
".",
"__class__",
"==",
"Extension",
")",
":",
"version_param",
"[",
"'_extension_version'",
"]",
"=",
"version",
"(",
"thread",
",",
"created_thread",
")",
"=",
"obj",
".",
"threads",
".",
"safer_get_or_create",
"(",
"defaults",
"=",
"create_perms",
",",
"**",
"version_param",
")",
"note",
"=",
"thread",
".",
"notes",
".",
"create",
"(",
"note_type",
"=",
"note_type",
",",
"body",
"=",
"body",
",",
"author",
"=",
"author",
",",
"**",
"create_perms",
")",
"if",
"attachments",
":",
"create_attachments",
"(",
"note",
",",
"attachments",
")",
"post_create_comm_note",
"(",
"note",
")",
"return",
"(",
"thread",
",",
"note",
")"
] | creates a note on an obj versions thread . | train | false |
25,145 | def parse_partial_identifier(word):
p = sqlparse.parse(word)[0]
n_tok = len(p.tokens)
if ((n_tok == 1) and isinstance(p.tokens[0], Identifier)):
return p.tokens[0]
elif p.token_next_by(m=(Error, '"'))[1]:
return parse_partial_identifier((word + '"'))
else:
return None
| [
"def",
"parse_partial_identifier",
"(",
"word",
")",
":",
"p",
"=",
"sqlparse",
".",
"parse",
"(",
"word",
")",
"[",
"0",
"]",
"n_tok",
"=",
"len",
"(",
"p",
".",
"tokens",
")",
"if",
"(",
"(",
"n_tok",
"==",
"1",
")",
"and",
"isinstance",
"(",
"p",
".",
"tokens",
"[",
"0",
"]",
",",
"Identifier",
")",
")",
":",
"return",
"p",
".",
"tokens",
"[",
"0",
"]",
"elif",
"p",
".",
"token_next_by",
"(",
"m",
"=",
"(",
"Error",
",",
"'\"'",
")",
")",
"[",
"1",
"]",
":",
"return",
"parse_partial_identifier",
"(",
"(",
"word",
"+",
"'\"'",
")",
")",
"else",
":",
"return",
"None"
] | attempt to parse a word as an identifier word may include a schema qualification . | train | false |
25,146 | def stub_out_image_service(test):
image_service = FakeImageService()
test.stub_out('nova.image.glance.get_remote_image_service', (lambda x, y: (image_service, y)))
test.stub_out('nova.image.glance.get_default_image_service', (lambda : image_service))
test.useFixture(nova_fixtures.ConfPatcher(group='glance', api_servers=['http://localhost:9292']))
return image_service
| [
"def",
"stub_out_image_service",
"(",
"test",
")",
":",
"image_service",
"=",
"FakeImageService",
"(",
")",
"test",
".",
"stub_out",
"(",
"'nova.image.glance.get_remote_image_service'",
",",
"(",
"lambda",
"x",
",",
"y",
":",
"(",
"image_service",
",",
"y",
")",
")",
")",
"test",
".",
"stub_out",
"(",
"'nova.image.glance.get_default_image_service'",
",",
"(",
"lambda",
":",
"image_service",
")",
")",
"test",
".",
"useFixture",
"(",
"nova_fixtures",
".",
"ConfPatcher",
"(",
"group",
"=",
"'glance'",
",",
"api_servers",
"=",
"[",
"'http://localhost:9292'",
"]",
")",
")",
"return",
"image_service"
] | stubs out the image service for the test with the fakeimageservice . | train | false |
25,148 | def reorder_title(title, articles=('the',), separators=(',', ', ')):
ltitle = title.lower()
for article in articles:
for separator in separators:
suffix = (separator + article)
if (ltitle[(- len(suffix)):] == suffix):
return ((title[((- len(suffix)) + len(separator)):] + ' ') + title[:(- len(suffix))])
return title
| [
"def",
"reorder_title",
"(",
"title",
",",
"articles",
"=",
"(",
"'the'",
",",
")",
",",
"separators",
"=",
"(",
"','",
",",
"', '",
")",
")",
":",
"ltitle",
"=",
"title",
".",
"lower",
"(",
")",
"for",
"article",
"in",
"articles",
":",
"for",
"separator",
"in",
"separators",
":",
"suffix",
"=",
"(",
"separator",
"+",
"article",
")",
"if",
"(",
"ltitle",
"[",
"(",
"-",
"len",
"(",
"suffix",
")",
")",
":",
"]",
"==",
"suffix",
")",
":",
"return",
"(",
"(",
"title",
"[",
"(",
"(",
"-",
"len",
"(",
"suffix",
")",
")",
"+",
"len",
"(",
"separator",
")",
")",
":",
"]",
"+",
"' '",
")",
"+",
"title",
"[",
":",
"(",
"-",
"len",
"(",
"suffix",
")",
")",
"]",
")",
"return",
"title"
] | reorder the title . | train | false |
25,149 | def get_descrs(arrays, col_name_map):
out_descrs = []
for (out_name, in_names) in six.iteritems(col_name_map):
in_cols = [arr[name] for (arr, name) in zip(arrays, in_names) if (name is not None)]
names = [name for name in in_names if (name is not None)]
try:
dtype = common_dtype(in_cols)
except TableMergeError as tme:
raise TableMergeError(u"The '{0}' columns have incompatible types: {1}".format(names[0], tme._incompat_types))
uniq_shapes = set((col.shape[1:] for col in in_cols))
if (len(uniq_shapes) != 1):
raise TableMergeError(u'Key columns {0!r} have different shape'.format(name))
shape = uniq_shapes.pop()
out_descrs.append((fix_column_name(out_name), dtype, shape))
return out_descrs
| [
"def",
"get_descrs",
"(",
"arrays",
",",
"col_name_map",
")",
":",
"out_descrs",
"=",
"[",
"]",
"for",
"(",
"out_name",
",",
"in_names",
")",
"in",
"six",
".",
"iteritems",
"(",
"col_name_map",
")",
":",
"in_cols",
"=",
"[",
"arr",
"[",
"name",
"]",
"for",
"(",
"arr",
",",
"name",
")",
"in",
"zip",
"(",
"arrays",
",",
"in_names",
")",
"if",
"(",
"name",
"is",
"not",
"None",
")",
"]",
"names",
"=",
"[",
"name",
"for",
"name",
"in",
"in_names",
"if",
"(",
"name",
"is",
"not",
"None",
")",
"]",
"try",
":",
"dtype",
"=",
"common_dtype",
"(",
"in_cols",
")",
"except",
"TableMergeError",
"as",
"tme",
":",
"raise",
"TableMergeError",
"(",
"u\"The '{0}' columns have incompatible types: {1}\"",
".",
"format",
"(",
"names",
"[",
"0",
"]",
",",
"tme",
".",
"_incompat_types",
")",
")",
"uniq_shapes",
"=",
"set",
"(",
"(",
"col",
".",
"shape",
"[",
"1",
":",
"]",
"for",
"col",
"in",
"in_cols",
")",
")",
"if",
"(",
"len",
"(",
"uniq_shapes",
")",
"!=",
"1",
")",
":",
"raise",
"TableMergeError",
"(",
"u'Key columns {0!r} have different shape'",
".",
"format",
"(",
"name",
")",
")",
"shape",
"=",
"uniq_shapes",
".",
"pop",
"(",
")",
"out_descrs",
".",
"append",
"(",
"(",
"fix_column_name",
"(",
"out_name",
")",
",",
"dtype",
",",
"shape",
")",
")",
"return",
"out_descrs"
] | find the dtypes descrs resulting from merging the list of arrays dtypes . | train | false |
25,150 | def generate_config(hass, passwd):
from homeassistant.components.mqtt import PROTOCOL_311
config = {'listeners': {'default': {'max-connections': 50000, 'bind': '0.0.0.0:1883', 'type': 'tcp'}, 'ws-1': {'bind': '0.0.0.0:8080', 'type': 'ws'}}, 'auth': {'allow-anonymous': (hass.config.api.api_password is None)}, 'plugins': ['auth_anonymous']}
if hass.config.api.api_password:
username = 'homeassistant'
password = hass.config.api.api_password
from passlib.apps import custom_app_context
passwd.write('homeassistant:{}\n'.format(custom_app_context.encrypt(hass.config.api.api_password)).encode('utf-8'))
passwd.flush()
config['auth']['password-file'] = passwd.name
config['plugins'].append('auth_file')
else:
username = None
password = None
client_config = ('localhost', 1883, username, password, None, PROTOCOL_311)
return (config, client_config)
| [
"def",
"generate_config",
"(",
"hass",
",",
"passwd",
")",
":",
"from",
"homeassistant",
".",
"components",
".",
"mqtt",
"import",
"PROTOCOL_311",
"config",
"=",
"{",
"'listeners'",
":",
"{",
"'default'",
":",
"{",
"'max-connections'",
":",
"50000",
",",
"'bind'",
":",
"'0.0.0.0:1883'",
",",
"'type'",
":",
"'tcp'",
"}",
",",
"'ws-1'",
":",
"{",
"'bind'",
":",
"'0.0.0.0:8080'",
",",
"'type'",
":",
"'ws'",
"}",
"}",
",",
"'auth'",
":",
"{",
"'allow-anonymous'",
":",
"(",
"hass",
".",
"config",
".",
"api",
".",
"api_password",
"is",
"None",
")",
"}",
",",
"'plugins'",
":",
"[",
"'auth_anonymous'",
"]",
"}",
"if",
"hass",
".",
"config",
".",
"api",
".",
"api_password",
":",
"username",
"=",
"'homeassistant'",
"password",
"=",
"hass",
".",
"config",
".",
"api",
".",
"api_password",
"from",
"passlib",
".",
"apps",
"import",
"custom_app_context",
"passwd",
".",
"write",
"(",
"'homeassistant:{}\\n'",
".",
"format",
"(",
"custom_app_context",
".",
"encrypt",
"(",
"hass",
".",
"config",
".",
"api",
".",
"api_password",
")",
")",
".",
"encode",
"(",
"'utf-8'",
")",
")",
"passwd",
".",
"flush",
"(",
")",
"config",
"[",
"'auth'",
"]",
"[",
"'password-file'",
"]",
"=",
"passwd",
".",
"name",
"config",
"[",
"'plugins'",
"]",
".",
"append",
"(",
"'auth_file'",
")",
"else",
":",
"username",
"=",
"None",
"password",
"=",
"None",
"client_config",
"=",
"(",
"'localhost'",
",",
"1883",
",",
"username",
",",
"password",
",",
"None",
",",
"PROTOCOL_311",
")",
"return",
"(",
"config",
",",
"client_config",
")"
] | generate a configuration based on current home assistant instance . | train | false |
25,151 | def upload_imageattachment(request, obj):
return upload_media(request, ImageAttachmentUploadForm, create_imageattachment, obj=obj)
| [
"def",
"upload_imageattachment",
"(",
"request",
",",
"obj",
")",
":",
"return",
"upload_media",
"(",
"request",
",",
"ImageAttachmentUploadForm",
",",
"create_imageattachment",
",",
"obj",
"=",
"obj",
")"
] | uploads image attachments . | train | false |
25,152 | def simple_rnn(incoming, n_units, activation='sigmoid', dropout=None, bias=True, weights_init=None, return_seq=False, return_state=False, initial_state=None, dynamic=False, trainable=True, restore=True, reuse=False, scope=None, name='SimpleRNN'):
cell = BasicRNNCell(n_units, activation=activation, bias=bias, weights_init=weights_init, trainable=trainable, restore=restore, reuse=reuse)
x = _rnn_template(incoming, cell=cell, dropout=dropout, return_seq=return_seq, return_state=return_state, initial_state=initial_state, dynamic=dynamic, scope=scope, name=name)
return x
| [
"def",
"simple_rnn",
"(",
"incoming",
",",
"n_units",
",",
"activation",
"=",
"'sigmoid'",
",",
"dropout",
"=",
"None",
",",
"bias",
"=",
"True",
",",
"weights_init",
"=",
"None",
",",
"return_seq",
"=",
"False",
",",
"return_state",
"=",
"False",
",",
"initial_state",
"=",
"None",
",",
"dynamic",
"=",
"False",
",",
"trainable",
"=",
"True",
",",
"restore",
"=",
"True",
",",
"reuse",
"=",
"False",
",",
"scope",
"=",
"None",
",",
"name",
"=",
"'SimpleRNN'",
")",
":",
"cell",
"=",
"BasicRNNCell",
"(",
"n_units",
",",
"activation",
"=",
"activation",
",",
"bias",
"=",
"bias",
",",
"weights_init",
"=",
"weights_init",
",",
"trainable",
"=",
"trainable",
",",
"restore",
"=",
"restore",
",",
"reuse",
"=",
"reuse",
")",
"x",
"=",
"_rnn_template",
"(",
"incoming",
",",
"cell",
"=",
"cell",
",",
"dropout",
"=",
"dropout",
",",
"return_seq",
"=",
"return_seq",
",",
"return_state",
"=",
"return_state",
",",
"initial_state",
"=",
"initial_state",
",",
"dynamic",
"=",
"dynamic",
",",
"scope",
"=",
"scope",
",",
"name",
"=",
"name",
")",
"return",
"x"
] | simple rnn . | train | false |
25,157 | def is_git_dir(d):
if osp.isdir(d):
if (osp.isdir(osp.join(d, 'objects')) and osp.isdir(osp.join(d, 'refs'))):
headref = osp.join(d, 'HEAD')
return (osp.isfile(headref) or (osp.islink(headref) and os.readlink(headref).startswith('refs')))
elif (osp.isfile(osp.join(d, 'gitdir')) and osp.isfile(osp.join(d, 'commondir')) and osp.isfile(osp.join(d, 'gitfile'))):
raise WorkTreeRepositoryUnsupported(d)
return False
| [
"def",
"is_git_dir",
"(",
"d",
")",
":",
"if",
"osp",
".",
"isdir",
"(",
"d",
")",
":",
"if",
"(",
"osp",
".",
"isdir",
"(",
"osp",
".",
"join",
"(",
"d",
",",
"'objects'",
")",
")",
"and",
"osp",
".",
"isdir",
"(",
"osp",
".",
"join",
"(",
"d",
",",
"'refs'",
")",
")",
")",
":",
"headref",
"=",
"osp",
".",
"join",
"(",
"d",
",",
"'HEAD'",
")",
"return",
"(",
"osp",
".",
"isfile",
"(",
"headref",
")",
"or",
"(",
"osp",
".",
"islink",
"(",
"headref",
")",
"and",
"os",
".",
"readlink",
"(",
"headref",
")",
".",
"startswith",
"(",
"'refs'",
")",
")",
")",
"elif",
"(",
"osp",
".",
"isfile",
"(",
"osp",
".",
"join",
"(",
"d",
",",
"'gitdir'",
")",
")",
"and",
"osp",
".",
"isfile",
"(",
"osp",
".",
"join",
"(",
"d",
",",
"'commondir'",
")",
")",
"and",
"osp",
".",
"isfile",
"(",
"osp",
".",
"join",
"(",
"d",
",",
"'gitfile'",
")",
")",
")",
":",
"raise",
"WorkTreeRepositoryUnsupported",
"(",
"d",
")",
"return",
"False"
] | from gits setup . | train | true |
25,158 | def advantage(cpu_dt, gpu_dt):
assert ((gpu_dt >= 0) and (cpu_dt >= 0))
if ((gpu_dt == 0) and (cpu_dt == 0)):
return numpy.nan
elif (gpu_dt == 0):
return numpy.inf
else:
return (cpu_dt / gpu_dt)
| [
"def",
"advantage",
"(",
"cpu_dt",
",",
"gpu_dt",
")",
":",
"assert",
"(",
"(",
"gpu_dt",
">=",
"0",
")",
"and",
"(",
"cpu_dt",
">=",
"0",
")",
")",
"if",
"(",
"(",
"gpu_dt",
"==",
"0",
")",
"and",
"(",
"cpu_dt",
"==",
"0",
")",
")",
":",
"return",
"numpy",
".",
"nan",
"elif",
"(",
"gpu_dt",
"==",
"0",
")",
":",
"return",
"numpy",
".",
"inf",
"else",
":",
"return",
"(",
"cpu_dt",
"/",
"gpu_dt",
")"
] | return ratio of cpu_dt / gpu_dt . | train | false |
25,161 | def _validate_connectivity(image_dim, connectivity, offset):
if (connectivity is None):
connectivity = 1
if np.isscalar(connectivity):
c_connectivity = ndi.generate_binary_structure(image_dim, connectivity)
else:
c_connectivity = np.array(connectivity, bool)
if (c_connectivity.ndim != image_dim):
raise ValueError('Connectivity dimension must be same as image')
if (offset is None):
if any([((x % 2) == 0) for x in c_connectivity.shape]):
raise ValueError('Connectivity array must have an unambiguous center')
offset = (np.array(c_connectivity.shape) // 2)
return (c_connectivity, offset)
| [
"def",
"_validate_connectivity",
"(",
"image_dim",
",",
"connectivity",
",",
"offset",
")",
":",
"if",
"(",
"connectivity",
"is",
"None",
")",
":",
"connectivity",
"=",
"1",
"if",
"np",
".",
"isscalar",
"(",
"connectivity",
")",
":",
"c_connectivity",
"=",
"ndi",
".",
"generate_binary_structure",
"(",
"image_dim",
",",
"connectivity",
")",
"else",
":",
"c_connectivity",
"=",
"np",
".",
"array",
"(",
"connectivity",
",",
"bool",
")",
"if",
"(",
"c_connectivity",
".",
"ndim",
"!=",
"image_dim",
")",
":",
"raise",
"ValueError",
"(",
"'Connectivity dimension must be same as image'",
")",
"if",
"(",
"offset",
"is",
"None",
")",
":",
"if",
"any",
"(",
"[",
"(",
"(",
"x",
"%",
"2",
")",
"==",
"0",
")",
"for",
"x",
"in",
"c_connectivity",
".",
"shape",
"]",
")",
":",
"raise",
"ValueError",
"(",
"'Connectivity array must have an unambiguous center'",
")",
"offset",
"=",
"(",
"np",
".",
"array",
"(",
"c_connectivity",
".",
"shape",
")",
"//",
"2",
")",
"return",
"(",
"c_connectivity",
",",
"offset",
")"
] | convert any valid connectivity to a structuring element and offset . | train | false |
25,164 | def unlock_file(filename):
log.trace('Removing lock for {0}'.format(filename))
lock = (filename + '.lock')
try:
os.remove(lock)
except OSError as exc:
log.trace('Unable to remove lock for {0}: {1}'.format(filename, exc))
| [
"def",
"unlock_file",
"(",
"filename",
")",
":",
"log",
".",
"trace",
"(",
"'Removing lock for {0}'",
".",
"format",
"(",
"filename",
")",
")",
"lock",
"=",
"(",
"filename",
"+",
"'.lock'",
")",
"try",
":",
"os",
".",
"remove",
"(",
"lock",
")",
"except",
"OSError",
"as",
"exc",
":",
"log",
".",
"trace",
"(",
"'Unable to remove lock for {0}: {1}'",
".",
"format",
"(",
"filename",
",",
"exc",
")",
")"
] | unlock file . | train | true |
25,165 | def _course_outline_json(request, course_module):
return create_xblock_info(course_module, include_child_info=True, course_outline=True, include_children_predicate=(lambda xblock: (not (xblock.category == 'vertical'))), user=request.user)
| [
"def",
"_course_outline_json",
"(",
"request",
",",
"course_module",
")",
":",
"return",
"create_xblock_info",
"(",
"course_module",
",",
"include_child_info",
"=",
"True",
",",
"course_outline",
"=",
"True",
",",
"include_children_predicate",
"=",
"(",
"lambda",
"xblock",
":",
"(",
"not",
"(",
"xblock",
".",
"category",
"==",
"'vertical'",
")",
")",
")",
",",
"user",
"=",
"request",
".",
"user",
")"
] | returns a json representation of the course module and recursively all of its children . | train | false |
25,166 | def ip_to_int(ip):
ret = 0
for octet in ip.split('.'):
ret = ((ret * 256) + int(octet))
return ret
| [
"def",
"ip_to_int",
"(",
"ip",
")",
":",
"ret",
"=",
"0",
"for",
"octet",
"in",
"ip",
".",
"split",
"(",
"'.'",
")",
":",
"ret",
"=",
"(",
"(",
"ret",
"*",
"256",
")",
"+",
"int",
"(",
"octet",
")",
")",
"return",
"ret"
] | converts an ip address to an integer . | train | true |
25,169 | @testing.requires_testing_data
@requires_mayavi
@requires_mne
@requires_freesurfer
def test_subject_source_with_fsaverage():
from mne.gui._file_traits import MRISubjectSource
tempdir = _TempDir()
mri = MRISubjectSource()
assert_false(mri.can_create_fsaverage)
assert_raises(RuntimeError, mri.create_fsaverage)
mri.subjects_dir = tempdir
assert_true(mri.can_create_fsaverage)
mri.create_fsaverage()
| [
"@",
"testing",
".",
"requires_testing_data",
"@",
"requires_mayavi",
"@",
"requires_mne",
"@",
"requires_freesurfer",
"def",
"test_subject_source_with_fsaverage",
"(",
")",
":",
"from",
"mne",
".",
"gui",
".",
"_file_traits",
"import",
"MRISubjectSource",
"tempdir",
"=",
"_TempDir",
"(",
")",
"mri",
"=",
"MRISubjectSource",
"(",
")",
"assert_false",
"(",
"mri",
".",
"can_create_fsaverage",
")",
"assert_raises",
"(",
"RuntimeError",
",",
"mri",
".",
"create_fsaverage",
")",
"mri",
".",
"subjects_dir",
"=",
"tempdir",
"assert_true",
"(",
"mri",
".",
"can_create_fsaverage",
")",
"mri",
".",
"create_fsaverage",
"(",
")"
] | test subjectselector . | train | false |
25,170 | def build_versions_form(project):
attrs = {'project': project}
versions_qs = project.versions.all()
active = versions_qs.filter(active=True)
if active.exists():
choices = [(version.slug, version.verbose_name) for version in active]
attrs['default-version'] = forms.ChoiceField(label=_('Default Version'), choices=choices, initial=project.get_default_version())
for version in versions_qs:
field_name = ('version-%s' % version.slug)
privacy_name = ('privacy-%s' % version.slug)
if (version.type == TAG):
label = ('%s (%s)' % (version.verbose_name, version.identifier[:8]))
else:
label = version.verbose_name
attrs[field_name] = forms.BooleanField(label=label, widget=DualCheckboxWidget(version), initial=version.active, required=False)
attrs[privacy_name] = forms.ChoiceField(label='privacy', choices=constants.PRIVACY_CHOICES, initial=version.privacy_level)
return type('VersionsForm', (BaseVersionsForm,), attrs)
| [
"def",
"build_versions_form",
"(",
"project",
")",
":",
"attrs",
"=",
"{",
"'project'",
":",
"project",
"}",
"versions_qs",
"=",
"project",
".",
"versions",
".",
"all",
"(",
")",
"active",
"=",
"versions_qs",
".",
"filter",
"(",
"active",
"=",
"True",
")",
"if",
"active",
".",
"exists",
"(",
")",
":",
"choices",
"=",
"[",
"(",
"version",
".",
"slug",
",",
"version",
".",
"verbose_name",
")",
"for",
"version",
"in",
"active",
"]",
"attrs",
"[",
"'default-version'",
"]",
"=",
"forms",
".",
"ChoiceField",
"(",
"label",
"=",
"_",
"(",
"'Default Version'",
")",
",",
"choices",
"=",
"choices",
",",
"initial",
"=",
"project",
".",
"get_default_version",
"(",
")",
")",
"for",
"version",
"in",
"versions_qs",
":",
"field_name",
"=",
"(",
"'version-%s'",
"%",
"version",
".",
"slug",
")",
"privacy_name",
"=",
"(",
"'privacy-%s'",
"%",
"version",
".",
"slug",
")",
"if",
"(",
"version",
".",
"type",
"==",
"TAG",
")",
":",
"label",
"=",
"(",
"'%s (%s)'",
"%",
"(",
"version",
".",
"verbose_name",
",",
"version",
".",
"identifier",
"[",
":",
"8",
"]",
")",
")",
"else",
":",
"label",
"=",
"version",
".",
"verbose_name",
"attrs",
"[",
"field_name",
"]",
"=",
"forms",
".",
"BooleanField",
"(",
"label",
"=",
"label",
",",
"widget",
"=",
"DualCheckboxWidget",
"(",
"version",
")",
",",
"initial",
"=",
"version",
".",
"active",
",",
"required",
"=",
"False",
")",
"attrs",
"[",
"privacy_name",
"]",
"=",
"forms",
".",
"ChoiceField",
"(",
"label",
"=",
"'privacy'",
",",
"choices",
"=",
"constants",
".",
"PRIVACY_CHOICES",
",",
"initial",
"=",
"version",
".",
"privacy_level",
")",
"return",
"type",
"(",
"'VersionsForm'",
",",
"(",
"BaseVersionsForm",
",",
")",
",",
"attrs",
")"
] | versions form with a list of versions and version privacy levels . | train | false |
25,172 | def award():
mode = session.s3.hrm.mode
def prep(r):
if (mode is not None):
auth.permission.fail()
return True
s3.prep = prep
output = s3_rest_controller()
return output
| [
"def",
"award",
"(",
")",
":",
"mode",
"=",
"session",
".",
"s3",
".",
"hrm",
".",
"mode",
"def",
"prep",
"(",
"r",
")",
":",
"if",
"(",
"mode",
"is",
"not",
"None",
")",
":",
"auth",
".",
"permission",
".",
"fail",
"(",
")",
"return",
"True",
"s3",
".",
"prep",
"=",
"prep",
"output",
"=",
"s3_rest_controller",
"(",
")",
"return",
"output"
] | volunteer awards controller . | train | false |
25,173 | def is_iterable(x):
try:
iter(x)
except TypeError:
return False
else:
return True
| [
"def",
"is_iterable",
"(",
"x",
")",
":",
"try",
":",
"iter",
"(",
"x",
")",
"except",
"TypeError",
":",
"return",
"False",
"else",
":",
"return",
"True"
] | are we being asked to look up a list of things . | train | false |
25,174 | def os_like_constants():
os_like = util.get_systemd_os_like()
if os_like:
for os_name in os_like:
if (os_name in CLI_DEFAULTS.keys()):
return CLI_DEFAULTS[os_name]
return {}
| [
"def",
"os_like_constants",
"(",
")",
":",
"os_like",
"=",
"util",
".",
"get_systemd_os_like",
"(",
")",
"if",
"os_like",
":",
"for",
"os_name",
"in",
"os_like",
":",
"if",
"(",
"os_name",
"in",
"CLI_DEFAULTS",
".",
"keys",
"(",
")",
")",
":",
"return",
"CLI_DEFAULTS",
"[",
"os_name",
"]",
"return",
"{",
"}"
] | try to get constants for distribution with similar layout and configuration . | train | false |
25,176 | def is_standard_module(modname, std_path=None):
modname = modname.split('.')[0]
try:
filename = file_from_modpath([modname])
except ImportError:
return False
if (filename is None):
return True
filename = _normalize_path(filename)
if filename.startswith(_cache_normalize_path(EXT_LIB_DIR)):
return False
if (std_path is None):
std_path = STD_LIB_DIRS
for path in std_path:
if filename.startswith(_cache_normalize_path(path)):
return True
return False
| [
"def",
"is_standard_module",
"(",
"modname",
",",
"std_path",
"=",
"None",
")",
":",
"modname",
"=",
"modname",
".",
"split",
"(",
"'.'",
")",
"[",
"0",
"]",
"try",
":",
"filename",
"=",
"file_from_modpath",
"(",
"[",
"modname",
"]",
")",
"except",
"ImportError",
":",
"return",
"False",
"if",
"(",
"filename",
"is",
"None",
")",
":",
"return",
"True",
"filename",
"=",
"_normalize_path",
"(",
"filename",
")",
"if",
"filename",
".",
"startswith",
"(",
"_cache_normalize_path",
"(",
"EXT_LIB_DIR",
")",
")",
":",
"return",
"False",
"if",
"(",
"std_path",
"is",
"None",
")",
":",
"std_path",
"=",
"STD_LIB_DIRS",
"for",
"path",
"in",
"std_path",
":",
"if",
"filename",
".",
"startswith",
"(",
"_cache_normalize_path",
"(",
"path",
")",
")",
":",
"return",
"True",
"return",
"False"
] | try to guess if a module is a standard python module :type modname: str . | train | false |
25,177 | def test_bootstrap_ols():
ols_fit = (lambda X, y: np.dot(np.dot(np.linalg.inv(np.dot(X.T, X)), X.T), y))
X = np.column_stack((rs.randn(50, 4), np.ones(50)))
w = [2, 4, 0, 3, 5]
y_noisy = (np.dot(X, w) + (rs.randn(50) * 20))
y_lownoise = (np.dot(X, w) + rs.randn(50))
n_boot = 500
w_boot_noisy = algo.bootstrap(X, y_noisy, n_boot=n_boot, func=ols_fit)
w_boot_lownoise = algo.bootstrap(X, y_lownoise, n_boot=n_boot, func=ols_fit)
assert_equal(w_boot_noisy.shape, (n_boot, 5))
assert_equal(w_boot_lownoise.shape, (n_boot, 5))
nose.tools.assert_greater(w_boot_noisy.std(), w_boot_lownoise.std())
| [
"def",
"test_bootstrap_ols",
"(",
")",
":",
"ols_fit",
"=",
"(",
"lambda",
"X",
",",
"y",
":",
"np",
".",
"dot",
"(",
"np",
".",
"dot",
"(",
"np",
".",
"linalg",
".",
"inv",
"(",
"np",
".",
"dot",
"(",
"X",
".",
"T",
",",
"X",
")",
")",
",",
"X",
".",
"T",
")",
",",
"y",
")",
")",
"X",
"=",
"np",
".",
"column_stack",
"(",
"(",
"rs",
".",
"randn",
"(",
"50",
",",
"4",
")",
",",
"np",
".",
"ones",
"(",
"50",
")",
")",
")",
"w",
"=",
"[",
"2",
",",
"4",
",",
"0",
",",
"3",
",",
"5",
"]",
"y_noisy",
"=",
"(",
"np",
".",
"dot",
"(",
"X",
",",
"w",
")",
"+",
"(",
"rs",
".",
"randn",
"(",
"50",
")",
"*",
"20",
")",
")",
"y_lownoise",
"=",
"(",
"np",
".",
"dot",
"(",
"X",
",",
"w",
")",
"+",
"rs",
".",
"randn",
"(",
"50",
")",
")",
"n_boot",
"=",
"500",
"w_boot_noisy",
"=",
"algo",
".",
"bootstrap",
"(",
"X",
",",
"y_noisy",
",",
"n_boot",
"=",
"n_boot",
",",
"func",
"=",
"ols_fit",
")",
"w_boot_lownoise",
"=",
"algo",
".",
"bootstrap",
"(",
"X",
",",
"y_lownoise",
",",
"n_boot",
"=",
"n_boot",
",",
"func",
"=",
"ols_fit",
")",
"assert_equal",
"(",
"w_boot_noisy",
".",
"shape",
",",
"(",
"n_boot",
",",
"5",
")",
")",
"assert_equal",
"(",
"w_boot_lownoise",
".",
"shape",
",",
"(",
"n_boot",
",",
"5",
")",
")",
"nose",
".",
"tools",
".",
"assert_greater",
"(",
"w_boot_noisy",
".",
"std",
"(",
")",
",",
"w_boot_lownoise",
".",
"std",
"(",
")",
")"
] | test bootstrap of ols model fit . | train | false |
25,179 | @contextlib.contextmanager
def report_scope(observation):
current = _reporters[(-1)]
old = current.observation
current.observation = observation
(yield)
current.observation = old
| [
"@",
"contextlib",
".",
"contextmanager",
"def",
"report_scope",
"(",
"observation",
")",
":",
"current",
"=",
"_reporters",
"[",
"(",
"-",
"1",
")",
"]",
"old",
"=",
"current",
".",
"observation",
"current",
".",
"observation",
"=",
"observation",
"(",
"yield",
")",
"current",
".",
"observation",
"=",
"old"
] | returns a report scope with the current reporter . | train | false |
25,180 | def test_masked_row_with_object_col():
t = table.Table([[1]], dtype=['O'], masked=True)
if NUMPY_LT_1_8:
with pytest.raises(ValueError):
t['col0'].mask = False
t[0].as_void()
with pytest.raises(ValueError):
t['col0'].mask = True
t[0].as_void()
else:
t['col0'].mask = False
assert (t[0]['col0'] == 1)
t['col0'].mask = True
assert (t[0]['col0'] is np.ma.masked)
| [
"def",
"test_masked_row_with_object_col",
"(",
")",
":",
"t",
"=",
"table",
".",
"Table",
"(",
"[",
"[",
"1",
"]",
"]",
",",
"dtype",
"=",
"[",
"'O'",
"]",
",",
"masked",
"=",
"True",
")",
"if",
"NUMPY_LT_1_8",
":",
"with",
"pytest",
".",
"raises",
"(",
"ValueError",
")",
":",
"t",
"[",
"'col0'",
"]",
".",
"mask",
"=",
"False",
"t",
"[",
"0",
"]",
".",
"as_void",
"(",
")",
"with",
"pytest",
".",
"raises",
"(",
"ValueError",
")",
":",
"t",
"[",
"'col0'",
"]",
".",
"mask",
"=",
"True",
"t",
"[",
"0",
"]",
".",
"as_void",
"(",
")",
"else",
":",
"t",
"[",
"'col0'",
"]",
".",
"mask",
"=",
"False",
"assert",
"(",
"t",
"[",
"0",
"]",
"[",
"'col0'",
"]",
"==",
"1",
")",
"t",
"[",
"'col0'",
"]",
".",
"mask",
"=",
"True",
"assert",
"(",
"t",
"[",
"0",
"]",
"[",
"'col0'",
"]",
"is",
"np",
".",
"ma",
".",
"masked",
")"
] | numpy < 1 . | train | false |
25,182 | def get_organization_courses(organization_id):
if (not organizations_enabled()):
return []
from organizations import api as organizations_api
return organizations_api.get_organization_courses(organization_id)
| [
"def",
"get_organization_courses",
"(",
"organization_id",
")",
":",
"if",
"(",
"not",
"organizations_enabled",
"(",
")",
")",
":",
"return",
"[",
"]",
"from",
"organizations",
"import",
"api",
"as",
"organizations_api",
"return",
"organizations_api",
".",
"get_organization_courses",
"(",
"organization_id",
")"
] | client api operation adapter/wrapper . | train | false |
25,185 | def supportsTeams(endpoint):
return endpoint.usesExtension(ns_uri)
| [
"def",
"supportsTeams",
"(",
"endpoint",
")",
":",
"return",
"endpoint",
".",
"usesExtension",
"(",
"ns_uri",
")"
] | does the given endpoint advertise support for launchpad teams? . | train | false |
25,186 | @pytest.mark.django_db
@pytest.mark.parametrize('view,model,get_object,service_provider_attr', [(PaymentMethodEditView, PaymentMethod, get_default_payment_method, 'payment_processor'), (ShippingMethodEditView, ShippingMethod, get_default_shipping_method, 'carrier')])
def test_behavior_add_save(rf, admin_user, view, model, get_object, service_provider_attr):
get_default_shop()
with override_settings(LANGUAGES=[('en', 'en')]):
object = get_object()
view = view.as_view()
service_provider_attr_field = ('base-%s' % service_provider_attr)
data = get_default_data(object, service_provider_attr, service_provider_attr_field)
components_before = object.behavior_components.count()
assert (not components_before)
request = apply_request_middleware(rf.post('/', data=data, user=admin_user))
view(request, pk=object.pk)
components_after = object.behavior_components.count()
assert (components_after == len(get_default_behavior_settings()))
| [
"@",
"pytest",
".",
"mark",
".",
"django_db",
"@",
"pytest",
".",
"mark",
".",
"parametrize",
"(",
"'view,model,get_object,service_provider_attr'",
",",
"[",
"(",
"PaymentMethodEditView",
",",
"PaymentMethod",
",",
"get_default_payment_method",
",",
"'payment_processor'",
")",
",",
"(",
"ShippingMethodEditView",
",",
"ShippingMethod",
",",
"get_default_shipping_method",
",",
"'carrier'",
")",
"]",
")",
"def",
"test_behavior_add_save",
"(",
"rf",
",",
"admin_user",
",",
"view",
",",
"model",
",",
"get_object",
",",
"service_provider_attr",
")",
":",
"get_default_shop",
"(",
")",
"with",
"override_settings",
"(",
"LANGUAGES",
"=",
"[",
"(",
"'en'",
",",
"'en'",
")",
"]",
")",
":",
"object",
"=",
"get_object",
"(",
")",
"view",
"=",
"view",
".",
"as_view",
"(",
")",
"service_provider_attr_field",
"=",
"(",
"'base-%s'",
"%",
"service_provider_attr",
")",
"data",
"=",
"get_default_data",
"(",
"object",
",",
"service_provider_attr",
",",
"service_provider_attr_field",
")",
"components_before",
"=",
"object",
".",
"behavior_components",
".",
"count",
"(",
")",
"assert",
"(",
"not",
"components_before",
")",
"request",
"=",
"apply_request_middleware",
"(",
"rf",
".",
"post",
"(",
"'/'",
",",
"data",
"=",
"data",
",",
"user",
"=",
"admin_user",
")",
")",
"view",
"(",
"request",
",",
"pk",
"=",
"object",
".",
"pk",
")",
"components_after",
"=",
"object",
".",
"behavior_components",
".",
"count",
"(",
")",
"assert",
"(",
"components_after",
"==",
"len",
"(",
"get_default_behavior_settings",
"(",
")",
")",
")"
] | to make things little bit more simple lets use only english as a language . | train | false |
25,187 | def preprocess_file(filename, cpp_path='cpp', cpp_args=''):
path_list = [cpp_path]
if isinstance(cpp_args, list):
path_list += cpp_args
elif (cpp_args != ''):
path_list += [cpp_args]
path_list += [filename]
try:
pipe = Popen(path_list, stdout=PIPE, universal_newlines=True)
text = pipe.communicate()[0]
except OSError as e:
raise RuntimeError((("Unable to invoke 'cpp'. " + 'Make sure its path was passed correctly\n') + ('Original error: %s' % e)))
return text
| [
"def",
"preprocess_file",
"(",
"filename",
",",
"cpp_path",
"=",
"'cpp'",
",",
"cpp_args",
"=",
"''",
")",
":",
"path_list",
"=",
"[",
"cpp_path",
"]",
"if",
"isinstance",
"(",
"cpp_args",
",",
"list",
")",
":",
"path_list",
"+=",
"cpp_args",
"elif",
"(",
"cpp_args",
"!=",
"''",
")",
":",
"path_list",
"+=",
"[",
"cpp_args",
"]",
"path_list",
"+=",
"[",
"filename",
"]",
"try",
":",
"pipe",
"=",
"Popen",
"(",
"path_list",
",",
"stdout",
"=",
"PIPE",
",",
"universal_newlines",
"=",
"True",
")",
"text",
"=",
"pipe",
".",
"communicate",
"(",
")",
"[",
"0",
"]",
"except",
"OSError",
"as",
"e",
":",
"raise",
"RuntimeError",
"(",
"(",
"(",
"\"Unable to invoke 'cpp'. \"",
"+",
"'Make sure its path was passed correctly\\n'",
")",
"+",
"(",
"'Original error: %s'",
"%",
"e",
")",
")",
")",
"return",
"text"
] | preprocess a file using cpp . | train | false |
25,188 | @cli.command()
@click.option('--username', prompt=True, help="The developer's shown username.")
@click.option('--email', prompt='E-Mail', help="The developer's email address")
@click.password_option(help='The login password.')
@pass_repo
def setuser(repo, username, email, password):
repo.set_config('username', username)
repo.set_config('email', email)
repo.set_config('password', ('*' * len(password)))
click.echo('Changed credentials.')
| [
"@",
"cli",
".",
"command",
"(",
")",
"@",
"click",
".",
"option",
"(",
"'--username'",
",",
"prompt",
"=",
"True",
",",
"help",
"=",
"\"The developer's shown username.\"",
")",
"@",
"click",
".",
"option",
"(",
"'--email'",
",",
"prompt",
"=",
"'E-Mail'",
",",
"help",
"=",
"\"The developer's email address\"",
")",
"@",
"click",
".",
"password_option",
"(",
"help",
"=",
"'The login password.'",
")",
"@",
"pass_repo",
"def",
"setuser",
"(",
"repo",
",",
"username",
",",
"email",
",",
"password",
")",
":",
"repo",
".",
"set_config",
"(",
"'username'",
",",
"username",
")",
"repo",
".",
"set_config",
"(",
"'email'",
",",
"email",
")",
"repo",
".",
"set_config",
"(",
"'password'",
",",
"(",
"'*'",
"*",
"len",
"(",
"password",
")",
")",
")",
"click",
".",
"echo",
"(",
"'Changed credentials.'",
")"
] | sets the user credentials . | train | false |
25,189 | def load_pytest_conf(path, parser):
namespace = {}
exec open(path, 'rb').read() in namespace
if ('update_parser' in namespace):
namespace['update_parser'](parser)
return namespace.get('CustomPyTester', PyTester)
| [
"def",
"load_pytest_conf",
"(",
"path",
",",
"parser",
")",
":",
"namespace",
"=",
"{",
"}",
"exec",
"open",
"(",
"path",
",",
"'rb'",
")",
".",
"read",
"(",
")",
"in",
"namespace",
"if",
"(",
"'update_parser'",
"in",
"namespace",
")",
":",
"namespace",
"[",
"'update_parser'",
"]",
"(",
"parser",
")",
"return",
"namespace",
".",
"get",
"(",
"'CustomPyTester'",
",",
"PyTester",
")"
] | loads a pytestconf . | train | false |
25,190 | def gen_lib_options(compiler, library_dirs, runtime_library_dirs, libraries):
lib_opts = []
for dir in library_dirs:
lib_opts.append(compiler.library_dir_option(dir))
for dir in runtime_library_dirs:
opt = compiler.runtime_library_dir_option(dir)
if isinstance(opt, list):
lib_opts = (lib_opts + opt)
else:
lib_opts.append(opt)
for lib in libraries:
(lib_dir, lib_name) = os.path.split(lib)
if lib_dir:
lib_file = compiler.find_library_file([lib_dir], lib_name)
if lib_file:
lib_opts.append(lib_file)
else:
compiler.warn(("no library file corresponding to '%s' found (skipping)" % lib))
else:
lib_opts.append(compiler.library_option(lib))
return lib_opts
| [
"def",
"gen_lib_options",
"(",
"compiler",
",",
"library_dirs",
",",
"runtime_library_dirs",
",",
"libraries",
")",
":",
"lib_opts",
"=",
"[",
"]",
"for",
"dir",
"in",
"library_dirs",
":",
"lib_opts",
".",
"append",
"(",
"compiler",
".",
"library_dir_option",
"(",
"dir",
")",
")",
"for",
"dir",
"in",
"runtime_library_dirs",
":",
"opt",
"=",
"compiler",
".",
"runtime_library_dir_option",
"(",
"dir",
")",
"if",
"isinstance",
"(",
"opt",
",",
"list",
")",
":",
"lib_opts",
"=",
"(",
"lib_opts",
"+",
"opt",
")",
"else",
":",
"lib_opts",
".",
"append",
"(",
"opt",
")",
"for",
"lib",
"in",
"libraries",
":",
"(",
"lib_dir",
",",
"lib_name",
")",
"=",
"os",
".",
"path",
".",
"split",
"(",
"lib",
")",
"if",
"lib_dir",
":",
"lib_file",
"=",
"compiler",
".",
"find_library_file",
"(",
"[",
"lib_dir",
"]",
",",
"lib_name",
")",
"if",
"lib_file",
":",
"lib_opts",
".",
"append",
"(",
"lib_file",
")",
"else",
":",
"compiler",
".",
"warn",
"(",
"(",
"\"no library file corresponding to '%s' found (skipping)\"",
"%",
"lib",
")",
")",
"else",
":",
"lib_opts",
".",
"append",
"(",
"compiler",
".",
"library_option",
"(",
"lib",
")",
")",
"return",
"lib_opts"
] | generate linker options for searching library directories and linking with specific libraries . | train | false |
25,191 | def py_encode(symb2freq):
heap = [[wt, [sym, u'']] for (sym, wt) in symb2freq.items()]
heapify(heap)
while (len(heap) > 1):
lo = heappop(heap)
hi = heappop(heap)
for pair in lo[1:]:
pair[1] = (u'0' + pair[1])
for pair in hi[1:]:
pair[1] = (u'1' + pair[1])
heappush(heap, (([(lo[0] + hi[0])] + lo[1:]) + hi[1:]))
return dict(heappop(heap)[1:])
| [
"def",
"py_encode",
"(",
"symb2freq",
")",
":",
"heap",
"=",
"[",
"[",
"wt",
",",
"[",
"sym",
",",
"u''",
"]",
"]",
"for",
"(",
"sym",
",",
"wt",
")",
"in",
"symb2freq",
".",
"items",
"(",
")",
"]",
"heapify",
"(",
"heap",
")",
"while",
"(",
"len",
"(",
"heap",
")",
">",
"1",
")",
":",
"lo",
"=",
"heappop",
"(",
"heap",
")",
"hi",
"=",
"heappop",
"(",
"heap",
")",
"for",
"pair",
"in",
"lo",
"[",
"1",
":",
"]",
":",
"pair",
"[",
"1",
"]",
"=",
"(",
"u'0'",
"+",
"pair",
"[",
"1",
"]",
")",
"for",
"pair",
"in",
"hi",
"[",
"1",
":",
"]",
":",
"pair",
"[",
"1",
"]",
"=",
"(",
"u'1'",
"+",
"pair",
"[",
"1",
"]",
")",
"heappush",
"(",
"heap",
",",
"(",
"(",
"[",
"(",
"lo",
"[",
"0",
"]",
"+",
"hi",
"[",
"0",
"]",
")",
"]",
"+",
"lo",
"[",
"1",
":",
"]",
")",
"+",
"hi",
"[",
"1",
":",
"]",
")",
")",
"return",
"dict",
"(",
"heappop",
"(",
"heap",
")",
"[",
"1",
":",
"]",
")"
] | huffman encode the given dict mapping symbols to weights from rosetta code . | train | false |
25,193 | def sign_certificate_request(keypair, dn, request, serial, validity_period, digest, start=None, additional_extensions=()):
if (start is None):
start = datetime.datetime.utcnow()
expire = (start + datetime.timedelta(seconds=validity_period))
start = start.strftime('%Y%m%d%H%M%SZ')
expire = expire.strftime('%Y%m%d%H%M%SZ')
req = request.original
cert = crypto.X509()
cert.set_issuer(dn)
cert.set_subject(req.get_subject())
cert.set_pubkey(req.get_pubkey())
cert.set_notBefore(start)
cert.set_notAfter(expire)
cert.set_serial_number(serial)
cert.add_extensions(additional_extensions)
cert.sign(keypair.original, digest)
return Certificate(cert)
| [
"def",
"sign_certificate_request",
"(",
"keypair",
",",
"dn",
",",
"request",
",",
"serial",
",",
"validity_period",
",",
"digest",
",",
"start",
"=",
"None",
",",
"additional_extensions",
"=",
"(",
")",
")",
":",
"if",
"(",
"start",
"is",
"None",
")",
":",
"start",
"=",
"datetime",
".",
"datetime",
".",
"utcnow",
"(",
")",
"expire",
"=",
"(",
"start",
"+",
"datetime",
".",
"timedelta",
"(",
"seconds",
"=",
"validity_period",
")",
")",
"start",
"=",
"start",
".",
"strftime",
"(",
"'%Y%m%d%H%M%SZ'",
")",
"expire",
"=",
"expire",
".",
"strftime",
"(",
"'%Y%m%d%H%M%SZ'",
")",
"req",
"=",
"request",
".",
"original",
"cert",
"=",
"crypto",
".",
"X509",
"(",
")",
"cert",
".",
"set_issuer",
"(",
"dn",
")",
"cert",
".",
"set_subject",
"(",
"req",
".",
"get_subject",
"(",
")",
")",
"cert",
".",
"set_pubkey",
"(",
"req",
".",
"get_pubkey",
"(",
")",
")",
"cert",
".",
"set_notBefore",
"(",
"start",
")",
"cert",
".",
"set_notAfter",
"(",
"expire",
")",
"cert",
".",
"set_serial_number",
"(",
"serial",
")",
"cert",
".",
"add_extensions",
"(",
"additional_extensions",
")",
"cert",
".",
"sign",
"(",
"keypair",
".",
"original",
",",
"digest",
")",
"return",
"Certificate",
"(",
"cert",
")"
] | sign a certificaterequest and return a certificate . | train | false |
25,194 | def getNewRepository():
return ExportRepository()
| [
"def",
"getNewRepository",
"(",
")",
":",
"return",
"ExportRepository",
"(",
")"
] | get the repository constructor . | train | false |
25,195 | def remove_profile_images(profile_image_names):
storage = get_profile_image_storage()
for name in profile_image_names.values():
storage.delete(name)
| [
"def",
"remove_profile_images",
"(",
"profile_image_names",
")",
":",
"storage",
"=",
"get_profile_image_storage",
"(",
")",
"for",
"name",
"in",
"profile_image_names",
".",
"values",
"(",
")",
":",
"storage",
".",
"delete",
"(",
"name",
")"
] | physically remove the image files specified in profile_image_names . | train | false |
25,196 | def profile_start(name):
_profiles_running[name] = time.time()
_profiles_stack.append(name)
| [
"def",
"profile_start",
"(",
"name",
")",
":",
"_profiles_running",
"[",
"name",
"]",
"=",
"time",
".",
"time",
"(",
")",
"_profiles_stack",
".",
"append",
"(",
"name",
")"
] | starts a profiling interval with specific name profiling data is sent to the client with next data batch . | train | false |
25,197 | def getLargestCenterOutsetLoopFromLoop(loop, radius, thresholdRatio=0.9):
if (radius == 0.0):
return loop
radius = abs(radius)
points = getPointsFromLoop(loop, radius, thresholdRatio)
centers = getCentersFromPoints(points, (globalIntercircleMultiplier * radius))
largestCenterOutset = None
largestOutsetArea = (-987654321.0)
for center in centers:
outset = getSimplifiedInsetFromClockwiseLoop(center, radius)
if isLargeSameDirection(outset, center, radius):
if (euclidean.isPathInsideLoop(loop, outset) != euclidean.isWiddershins(loop)):
centerOutset = CenterOutset(center, outset)
outsetArea = abs(euclidean.getAreaLoop(outset))
if (outsetArea > largestOutsetArea):
largestOutsetArea = outsetArea
largestCenterOutset = centerOutset
if (largestCenterOutset == None):
return None
largestCenterOutset.center = euclidean.getSimplifiedLoop(largestCenterOutset.center, radius)
return largestCenterOutset
| [
"def",
"getLargestCenterOutsetLoopFromLoop",
"(",
"loop",
",",
"radius",
",",
"thresholdRatio",
"=",
"0.9",
")",
":",
"if",
"(",
"radius",
"==",
"0.0",
")",
":",
"return",
"loop",
"radius",
"=",
"abs",
"(",
"radius",
")",
"points",
"=",
"getPointsFromLoop",
"(",
"loop",
",",
"radius",
",",
"thresholdRatio",
")",
"centers",
"=",
"getCentersFromPoints",
"(",
"points",
",",
"(",
"globalIntercircleMultiplier",
"*",
"radius",
")",
")",
"largestCenterOutset",
"=",
"None",
"largestOutsetArea",
"=",
"(",
"-",
"987654321.0",
")",
"for",
"center",
"in",
"centers",
":",
"outset",
"=",
"getSimplifiedInsetFromClockwiseLoop",
"(",
"center",
",",
"radius",
")",
"if",
"isLargeSameDirection",
"(",
"outset",
",",
"center",
",",
"radius",
")",
":",
"if",
"(",
"euclidean",
".",
"isPathInsideLoop",
"(",
"loop",
",",
"outset",
")",
"!=",
"euclidean",
".",
"isWiddershins",
"(",
"loop",
")",
")",
":",
"centerOutset",
"=",
"CenterOutset",
"(",
"center",
",",
"outset",
")",
"outsetArea",
"=",
"abs",
"(",
"euclidean",
".",
"getAreaLoop",
"(",
"outset",
")",
")",
"if",
"(",
"outsetArea",
">",
"largestOutsetArea",
")",
":",
"largestOutsetArea",
"=",
"outsetArea",
"largestCenterOutset",
"=",
"centerOutset",
"if",
"(",
"largestCenterOutset",
"==",
"None",
")",
":",
"return",
"None",
"largestCenterOutset",
".",
"center",
"=",
"euclidean",
".",
"getSimplifiedLoop",
"(",
"largestCenterOutset",
".",
"center",
",",
"radius",
")",
"return",
"largestCenterOutset"
] | get the largest circle outset loop from the loop . | train | false |
25,198 | def copy_data(data, dest, header=None, use_put=None):
ret = None
if use_put:
udata = data
else:
udata = urllib.urlencode(data)
if utils.is_url(dest):
ret = copy_remote(udata, dest, use_put)
if header:
return ret[header]
else:
if header:
ret = (dest + str(time.time()))
dest = (ret + '/_task_result')
copy_local(udata, dest, use_put)
return ret
| [
"def",
"copy_data",
"(",
"data",
",",
"dest",
",",
"header",
"=",
"None",
",",
"use_put",
"=",
"None",
")",
":",
"ret",
"=",
"None",
"if",
"use_put",
":",
"udata",
"=",
"data",
"else",
":",
"udata",
"=",
"urllib",
".",
"urlencode",
"(",
"data",
")",
"if",
"utils",
".",
"is_url",
"(",
"dest",
")",
":",
"ret",
"=",
"copy_remote",
"(",
"udata",
",",
"dest",
",",
"use_put",
")",
"if",
"header",
":",
"return",
"ret",
"[",
"header",
"]",
"else",
":",
"if",
"header",
":",
"ret",
"=",
"(",
"dest",
"+",
"str",
"(",
"time",
".",
"time",
"(",
")",
")",
")",
"dest",
"=",
"(",
"ret",
"+",
"'/_task_result'",
")",
"copy_local",
"(",
"udata",
",",
"dest",
",",
"use_put",
")",
"return",
"ret"
] | copy data to a destination to aid in debugging . | train | false |
25,199 | def _any_pandas_objects(terms):
return any((isinstance(term.value, pd.core.generic.PandasObject) for term in terms))
| [
"def",
"_any_pandas_objects",
"(",
"terms",
")",
":",
"return",
"any",
"(",
"(",
"isinstance",
"(",
"term",
".",
"value",
",",
"pd",
".",
"core",
".",
"generic",
".",
"PandasObject",
")",
"for",
"term",
"in",
"terms",
")",
")"
] | check a sequence of terms for instances of pandasobject . | train | true |
25,200 | def make_dataset(X, y, sample_weight, random_state=None):
rng = check_random_state(random_state)
seed = rng.randint(1, np.iinfo(np.int32).max)
if sp.issparse(X):
dataset = CSRDataset(X.data, X.indptr, X.indices, y, sample_weight, seed=seed)
intercept_decay = SPARSE_INTERCEPT_DECAY
else:
dataset = ArrayDataset(X, y, sample_weight, seed=seed)
intercept_decay = 1.0
return (dataset, intercept_decay)
| [
"def",
"make_dataset",
"(",
"X",
",",
"y",
",",
"sample_weight",
",",
"random_state",
"=",
"None",
")",
":",
"rng",
"=",
"check_random_state",
"(",
"random_state",
")",
"seed",
"=",
"rng",
".",
"randint",
"(",
"1",
",",
"np",
".",
"iinfo",
"(",
"np",
".",
"int32",
")",
".",
"max",
")",
"if",
"sp",
".",
"issparse",
"(",
"X",
")",
":",
"dataset",
"=",
"CSRDataset",
"(",
"X",
".",
"data",
",",
"X",
".",
"indptr",
",",
"X",
".",
"indices",
",",
"y",
",",
"sample_weight",
",",
"seed",
"=",
"seed",
")",
"intercept_decay",
"=",
"SPARSE_INTERCEPT_DECAY",
"else",
":",
"dataset",
"=",
"ArrayDataset",
"(",
"X",
",",
"y",
",",
"sample_weight",
",",
"seed",
"=",
"seed",
")",
"intercept_decay",
"=",
"1.0",
"return",
"(",
"dataset",
",",
"intercept_decay",
")"
] | create dataset abstraction for sparse and dense inputs . | train | false |
25,201 | def askUser(text, parent=None, help='', defaultno=False, msgfunc=None, title='Anki'):
if (not parent):
parent = aqt.mw.app.activeWindow()
if (not msgfunc):
msgfunc = QMessageBox.question
sb = (QMessageBox.Yes | QMessageBox.No)
if help:
sb |= QMessageBox.Help
while 1:
if defaultno:
default = QMessageBox.No
else:
default = QMessageBox.Yes
r = msgfunc(parent, title, text, sb, default)
if (r == QMessageBox.Help):
openHelp(help)
else:
break
return (r == QMessageBox.Yes)
| [
"def",
"askUser",
"(",
"text",
",",
"parent",
"=",
"None",
",",
"help",
"=",
"''",
",",
"defaultno",
"=",
"False",
",",
"msgfunc",
"=",
"None",
",",
"title",
"=",
"'Anki'",
")",
":",
"if",
"(",
"not",
"parent",
")",
":",
"parent",
"=",
"aqt",
".",
"mw",
".",
"app",
".",
"activeWindow",
"(",
")",
"if",
"(",
"not",
"msgfunc",
")",
":",
"msgfunc",
"=",
"QMessageBox",
".",
"question",
"sb",
"=",
"(",
"QMessageBox",
".",
"Yes",
"|",
"QMessageBox",
".",
"No",
")",
"if",
"help",
":",
"sb",
"|=",
"QMessageBox",
".",
"Help",
"while",
"1",
":",
"if",
"defaultno",
":",
"default",
"=",
"QMessageBox",
".",
"No",
"else",
":",
"default",
"=",
"QMessageBox",
".",
"Yes",
"r",
"=",
"msgfunc",
"(",
"parent",
",",
"title",
",",
"text",
",",
"sb",
",",
"default",
")",
"if",
"(",
"r",
"==",
"QMessageBox",
".",
"Help",
")",
":",
"openHelp",
"(",
"help",
")",
"else",
":",
"break",
"return",
"(",
"r",
"==",
"QMessageBox",
".",
"Yes",
")"
] | show a yes/no question . | train | false |
25,203 | def broadcast_to_sharejs(action, sharejs_uuid, node=None, wiki_name='home', data=None):
url = 'http://{host}:{port}/{action}/{id}/'.format(host=wiki_settings.SHAREJS_HOST, port=wiki_settings.SHAREJS_PORT, action=action, id=sharejs_uuid)
if ((action == 'redirect') or (action == 'delete')):
redirect_url = urllib.quote(node.web_url_for('project_wiki_view', wname=wiki_name, _guid=True), safe='')
url = os.path.join(url, redirect_url)
try:
requests.post(url, json=data)
except requests.ConnectionError:
pass
| [
"def",
"broadcast_to_sharejs",
"(",
"action",
",",
"sharejs_uuid",
",",
"node",
"=",
"None",
",",
"wiki_name",
"=",
"'home'",
",",
"data",
"=",
"None",
")",
":",
"url",
"=",
"'http://{host}:{port}/{action}/{id}/'",
".",
"format",
"(",
"host",
"=",
"wiki_settings",
".",
"SHAREJS_HOST",
",",
"port",
"=",
"wiki_settings",
".",
"SHAREJS_PORT",
",",
"action",
"=",
"action",
",",
"id",
"=",
"sharejs_uuid",
")",
"if",
"(",
"(",
"action",
"==",
"'redirect'",
")",
"or",
"(",
"action",
"==",
"'delete'",
")",
")",
":",
"redirect_url",
"=",
"urllib",
".",
"quote",
"(",
"node",
".",
"web_url_for",
"(",
"'project_wiki_view'",
",",
"wname",
"=",
"wiki_name",
",",
"_guid",
"=",
"True",
")",
",",
"safe",
"=",
"''",
")",
"url",
"=",
"os",
".",
"path",
".",
"join",
"(",
"url",
",",
"redirect_url",
")",
"try",
":",
"requests",
".",
"post",
"(",
"url",
",",
"json",
"=",
"data",
")",
"except",
"requests",
".",
"ConnectionError",
":",
"pass"
] | broadcast an action to all documents connected to a wiki . | train | false |
25,204 | def dict_formatter(view, value):
return json.dumps(value, ensure_ascii=False)
| [
"def",
"dict_formatter",
"(",
"view",
",",
"value",
")",
":",
"return",
"json",
".",
"dumps",
"(",
"value",
",",
"ensure_ascii",
"=",
"False",
")"
] | removes unicode entities when displaying dict as string . | train | false |
25,205 | def test_eppstein_matching():
G = nx.Graph()
G.add_nodes_from(['a', 2, 3, 4], bipartite=0)
G.add_nodes_from([1, 'b', 'c'], bipartite=1)
G.add_edges_from([('a', 1), ('a', 'b'), (2, 'b'), (2, 'c'), (3, 'c'), (4, 1)])
matching = eppstein_matching(G)
assert_true((len(matching) == len(maximum_matching(G))))
assert all(((x in set(matching.keys())) for x in set(matching.values())))
| [
"def",
"test_eppstein_matching",
"(",
")",
":",
"G",
"=",
"nx",
".",
"Graph",
"(",
")",
"G",
".",
"add_nodes_from",
"(",
"[",
"'a'",
",",
"2",
",",
"3",
",",
"4",
"]",
",",
"bipartite",
"=",
"0",
")",
"G",
".",
"add_nodes_from",
"(",
"[",
"1",
",",
"'b'",
",",
"'c'",
"]",
",",
"bipartite",
"=",
"1",
")",
"G",
".",
"add_edges_from",
"(",
"[",
"(",
"'a'",
",",
"1",
")",
",",
"(",
"'a'",
",",
"'b'",
")",
",",
"(",
"2",
",",
"'b'",
")",
",",
"(",
"2",
",",
"'c'",
")",
",",
"(",
"3",
",",
"'c'",
")",
",",
"(",
"4",
",",
"1",
")",
"]",
")",
"matching",
"=",
"eppstein_matching",
"(",
"G",
")",
"assert_true",
"(",
"(",
"len",
"(",
"matching",
")",
"==",
"len",
"(",
"maximum_matching",
"(",
"G",
")",
")",
")",
")",
"assert",
"all",
"(",
"(",
"(",
"x",
"in",
"set",
"(",
"matching",
".",
"keys",
"(",
")",
")",
")",
"for",
"x",
"in",
"set",
"(",
"matching",
".",
"values",
"(",
")",
")",
")",
")"
] | test in accordance to issue #1927 . | train | false |
25,206 | def clashing():
ns = {}
exec_('from sympy import *', ns)
clash1 = {}
clash2 = {}
while ns:
(k, _) = ns.popitem()
if (k in _greek):
clash2[k] = Symbol(k)
_greek.remove(k)
elif (k in _latin):
clash1[k] = Symbol(k)
_latin.remove(k)
clash = {}
clash.update(clash1)
clash.update(clash2)
return (clash1, clash2, clash)
| [
"def",
"clashing",
"(",
")",
":",
"ns",
"=",
"{",
"}",
"exec_",
"(",
"'from sympy import *'",
",",
"ns",
")",
"clash1",
"=",
"{",
"}",
"clash2",
"=",
"{",
"}",
"while",
"ns",
":",
"(",
"k",
",",
"_",
")",
"=",
"ns",
".",
"popitem",
"(",
")",
"if",
"(",
"k",
"in",
"_greek",
")",
":",
"clash2",
"[",
"k",
"]",
"=",
"Symbol",
"(",
"k",
")",
"_greek",
".",
"remove",
"(",
"k",
")",
"elif",
"(",
"k",
"in",
"_latin",
")",
":",
"clash1",
"[",
"k",
"]",
"=",
"Symbol",
"(",
"k",
")",
"_latin",
".",
"remove",
"(",
"k",
")",
"clash",
"=",
"{",
"}",
"clash",
".",
"update",
"(",
"clash1",
")",
"clash",
".",
"update",
"(",
"clash2",
")",
"return",
"(",
"clash1",
",",
"clash2",
",",
"clash",
")"
] | return the clashing-symbols dictionaries . | train | false |
25,207 | @handle_response_format
@treeio_login_required
def item_view(request, folderPath, itemPath, response_format='html'):
try:
item = KnowledgeItem.by_path(folderPath, itemPath)
except KnowledgeItem.DoesNotExist:
raise Http404
if (not item):
raise Http404
items = Object.filter_permitted(manager=KnowledgeItem.objects, user=request.user.profile, mode='r')
if (not request.user.profile.has_permission(item)):
return user_denied(request, message="You don't have access to this Knowledge Item")
context = _get_default_context(request)
context.update({'items': items, 'item': item})
return render_to_response('knowledge/item_view', context, context_instance=RequestContext(request), response_format=response_format)
| [
"@",
"handle_response_format",
"@",
"treeio_login_required",
"def",
"item_view",
"(",
"request",
",",
"folderPath",
",",
"itemPath",
",",
"response_format",
"=",
"'html'",
")",
":",
"try",
":",
"item",
"=",
"KnowledgeItem",
".",
"by_path",
"(",
"folderPath",
",",
"itemPath",
")",
"except",
"KnowledgeItem",
".",
"DoesNotExist",
":",
"raise",
"Http404",
"if",
"(",
"not",
"item",
")",
":",
"raise",
"Http404",
"items",
"=",
"Object",
".",
"filter_permitted",
"(",
"manager",
"=",
"KnowledgeItem",
".",
"objects",
",",
"user",
"=",
"request",
".",
"user",
".",
"profile",
",",
"mode",
"=",
"'r'",
")",
"if",
"(",
"not",
"request",
".",
"user",
".",
"profile",
".",
"has_permission",
"(",
"item",
")",
")",
":",
"return",
"user_denied",
"(",
"request",
",",
"message",
"=",
"\"You don't have access to this Knowledge Item\"",
")",
"context",
"=",
"_get_default_context",
"(",
"request",
")",
"context",
".",
"update",
"(",
"{",
"'items'",
":",
"items",
",",
"'item'",
":",
"item",
"}",
")",
"return",
"render_to_response",
"(",
"'knowledge/item_view'",
",",
"context",
",",
"context_instance",
"=",
"RequestContext",
"(",
"request",
")",
",",
"response_format",
"=",
"response_format",
")"
] | item view . | train | false |
25,208 | @before.each_scenario
def process_requires_tags(scenario):
tag_re = re.compile('requires_stub_(?P<server>[^_]+)')
for tag in scenario.tags:
requires = tag_re.match(tag)
if requires:
if (requires.group('server') == 'youtube'):
if (not is_youtube_available(YOUTUBE_API_URLS)):
scenario.steps = []
return
start_stub(requires.group('server'))
| [
"@",
"before",
".",
"each_scenario",
"def",
"process_requires_tags",
"(",
"scenario",
")",
":",
"tag_re",
"=",
"re",
".",
"compile",
"(",
"'requires_stub_(?P<server>[^_]+)'",
")",
"for",
"tag",
"in",
"scenario",
".",
"tags",
":",
"requires",
"=",
"tag_re",
".",
"match",
"(",
"tag",
")",
"if",
"requires",
":",
"if",
"(",
"requires",
".",
"group",
"(",
"'server'",
")",
"==",
"'youtube'",
")",
":",
"if",
"(",
"not",
"is_youtube_available",
"(",
"YOUTUBE_API_URLS",
")",
")",
":",
"scenario",
".",
"steps",
"=",
"[",
"]",
"return",
"start_stub",
"(",
"requires",
".",
"group",
"(",
"'server'",
")",
")"
] | process the scenario tags to make sure that any requirements are met prior to that scenario being executed . | train | false |
25,209 | def api_prefix(url_prefix=None, api_version=None):
if (url_prefix is None):
url_prefix = config.URL_PREFIX
if (api_version is None):
api_version = config.API_VERSION
prefix = (('/%s' % url_prefix) if url_prefix else '')
version = (('/%s' % api_version) if api_version else '')
return (prefix + version)
| [
"def",
"api_prefix",
"(",
"url_prefix",
"=",
"None",
",",
"api_version",
"=",
"None",
")",
":",
"if",
"(",
"url_prefix",
"is",
"None",
")",
":",
"url_prefix",
"=",
"config",
".",
"URL_PREFIX",
"if",
"(",
"api_version",
"is",
"None",
")",
":",
"api_version",
"=",
"config",
".",
"API_VERSION",
"prefix",
"=",
"(",
"(",
"'/%s'",
"%",
"url_prefix",
")",
"if",
"url_prefix",
"else",
"''",
")",
"version",
"=",
"(",
"(",
"'/%s'",
"%",
"api_version",
")",
"if",
"api_version",
"else",
"''",
")",
"return",
"(",
"prefix",
"+",
"version",
")"
] | returns the prefix to api endpoints . | train | false |
25,212 | def test_human_readable_custom():
f = formatters.HumanReadable()
assert (f(None) == u('\xe2\x88\x85'))
f = formatters.HumanReadable(none_char='/')
assert (f(None) == '/')
| [
"def",
"test_human_readable_custom",
"(",
")",
":",
"f",
"=",
"formatters",
".",
"HumanReadable",
"(",
")",
"assert",
"(",
"f",
"(",
"None",
")",
"==",
"u",
"(",
"'\\xe2\\x88\\x85'",
")",
")",
"f",
"=",
"formatters",
".",
"HumanReadable",
"(",
"none_char",
"=",
"'/'",
")",
"assert",
"(",
"f",
"(",
"None",
")",
"==",
"'/'",
")"
] | test human_readable formatter option . | train | false |
25,213 | def gather_bootstrap_script(bootstrap=None):
if (not HAS_CLOUD):
return (False, 'config.gather_bootstrap_script is unavailable')
ret = salt.utils.cloud.update_bootstrap(__opts__, url=bootstrap)
if (('Success' in ret) and (len(ret['Success']['Files updated']) > 0)):
return ret['Success']['Files updated'][0]
| [
"def",
"gather_bootstrap_script",
"(",
"bootstrap",
"=",
"None",
")",
":",
"if",
"(",
"not",
"HAS_CLOUD",
")",
":",
"return",
"(",
"False",
",",
"'config.gather_bootstrap_script is unavailable'",
")",
"ret",
"=",
"salt",
".",
"utils",
".",
"cloud",
".",
"update_bootstrap",
"(",
"__opts__",
",",
"url",
"=",
"bootstrap",
")",
"if",
"(",
"(",
"'Success'",
"in",
"ret",
")",
"and",
"(",
"len",
"(",
"ret",
"[",
"'Success'",
"]",
"[",
"'Files updated'",
"]",
")",
">",
"0",
")",
")",
":",
"return",
"ret",
"[",
"'Success'",
"]",
"[",
"'Files updated'",
"]",
"[",
"0",
"]"
] | download the salt-bootstrap script . | train | true |
25,214 | def float_to_byte(value, mantissabits=5, zeroexp=2):
fzero = ((63 - zeroexp) << mantissabits)
bits = unpack('i', pack('f', value))[0]
smallfloat = (bits >> (24 - mantissabits))
if (smallfloat < fzero):
if (bits <= 0):
result = chr(0)
else:
result = chr(1)
elif (smallfloat >= (fzero + 256)):
result = chr(255)
else:
result = chr((smallfloat - fzero))
return b(result)
| [
"def",
"float_to_byte",
"(",
"value",
",",
"mantissabits",
"=",
"5",
",",
"zeroexp",
"=",
"2",
")",
":",
"fzero",
"=",
"(",
"(",
"63",
"-",
"zeroexp",
")",
"<<",
"mantissabits",
")",
"bits",
"=",
"unpack",
"(",
"'i'",
",",
"pack",
"(",
"'f'",
",",
"value",
")",
")",
"[",
"0",
"]",
"smallfloat",
"=",
"(",
"bits",
">>",
"(",
"24",
"-",
"mantissabits",
")",
")",
"if",
"(",
"smallfloat",
"<",
"fzero",
")",
":",
"if",
"(",
"bits",
"<=",
"0",
")",
":",
"result",
"=",
"chr",
"(",
"0",
")",
"else",
":",
"result",
"=",
"chr",
"(",
"1",
")",
"elif",
"(",
"smallfloat",
">=",
"(",
"fzero",
"+",
"256",
")",
")",
":",
"result",
"=",
"chr",
"(",
"255",
")",
"else",
":",
"result",
"=",
"chr",
"(",
"(",
"smallfloat",
"-",
"fzero",
")",
")",
"return",
"b",
"(",
"result",
")"
] | encodes a floating point number in a single byte . | train | false |
25,216 | def test_sigmoid():
def ref_sigmoid(x):
if (x >= 0):
return (1 / (1 + np.exp((- x))))
else:
z = np.exp(x)
return (z / (1 + z))
sigmoid = np.vectorize(ref_sigmoid)
x = K.placeholder(ndim=2)
f = K.function([x], [activations.sigmoid(x)])
test_values = get_standard_values()
result = f([test_values])[0]
expected = sigmoid(test_values)
assert_allclose(result, expected, rtol=1e-05)
| [
"def",
"test_sigmoid",
"(",
")",
":",
"def",
"ref_sigmoid",
"(",
"x",
")",
":",
"if",
"(",
"x",
">=",
"0",
")",
":",
"return",
"(",
"1",
"/",
"(",
"1",
"+",
"np",
".",
"exp",
"(",
"(",
"-",
"x",
")",
")",
")",
")",
"else",
":",
"z",
"=",
"np",
".",
"exp",
"(",
"x",
")",
"return",
"(",
"z",
"/",
"(",
"1",
"+",
"z",
")",
")",
"sigmoid",
"=",
"np",
".",
"vectorize",
"(",
"ref_sigmoid",
")",
"x",
"=",
"K",
".",
"placeholder",
"(",
"ndim",
"=",
"2",
")",
"f",
"=",
"K",
".",
"function",
"(",
"[",
"x",
"]",
",",
"[",
"activations",
".",
"sigmoid",
"(",
"x",
")",
"]",
")",
"test_values",
"=",
"get_standard_values",
"(",
")",
"result",
"=",
"f",
"(",
"[",
"test_values",
"]",
")",
"[",
"0",
"]",
"expected",
"=",
"sigmoid",
"(",
"test_values",
")",
"assert_allclose",
"(",
"result",
",",
"expected",
",",
"rtol",
"=",
"1e-05",
")"
] | test using a numerically stable reference sigmoid implementation . | train | false |
25,217 | def generate_tasks(**kwargs):
tasks = _local_tasks()
for (fun, args) in tasks:
fun(*args)
tasks.clear()
| [
"def",
"generate_tasks",
"(",
"**",
"kwargs",
")",
":",
"tasks",
"=",
"_local_tasks",
"(",
")",
"for",
"(",
"fun",
",",
"args",
")",
"in",
"tasks",
":",
"fun",
"(",
"*",
"args",
")",
"tasks",
".",
"clear",
"(",
")"
] | goes through thread local index update tasks set and generates celery tasks for all tasks in the set . | train | false |
25,220 | def get_cursor_position(fd=1):
csbi = get_console_screen_buffer_info(fd=fd)
coord = csbi.dwCursorPosition
return (coord.X, coord.Y)
| [
"def",
"get_cursor_position",
"(",
"fd",
"=",
"1",
")",
":",
"csbi",
"=",
"get_console_screen_buffer_info",
"(",
"fd",
"=",
"fd",
")",
"coord",
"=",
"csbi",
".",
"dwCursorPosition",
"return",
"(",
"coord",
".",
"X",
",",
"coord",
".",
"Y",
")"
] | gets the current cursor position as an tuple . | train | false |
25,222 | def get_local_facts_from_file(filename):
local_facts = dict()
try:
ini_facts = configparser.SafeConfigParser()
ini_facts.read(filename)
for section in ini_facts.sections():
local_facts[section] = dict()
for (key, value) in ini_facts.items(section):
local_facts[section][key] = value
except (configparser.MissingSectionHeaderError, configparser.ParsingError):
try:
with open(filename, 'r') as facts_file:
local_facts = json.load(facts_file)
except (ValueError, IOError):
pass
return local_facts
| [
"def",
"get_local_facts_from_file",
"(",
"filename",
")",
":",
"local_facts",
"=",
"dict",
"(",
")",
"try",
":",
"ini_facts",
"=",
"configparser",
".",
"SafeConfigParser",
"(",
")",
"ini_facts",
".",
"read",
"(",
"filename",
")",
"for",
"section",
"in",
"ini_facts",
".",
"sections",
"(",
")",
":",
"local_facts",
"[",
"section",
"]",
"=",
"dict",
"(",
")",
"for",
"(",
"key",
",",
"value",
")",
"in",
"ini_facts",
".",
"items",
"(",
"section",
")",
":",
"local_facts",
"[",
"section",
"]",
"[",
"key",
"]",
"=",
"value",
"except",
"(",
"configparser",
".",
"MissingSectionHeaderError",
",",
"configparser",
".",
"ParsingError",
")",
":",
"try",
":",
"with",
"open",
"(",
"filename",
",",
"'r'",
")",
"as",
"facts_file",
":",
"local_facts",
"=",
"json",
".",
"load",
"(",
"facts_file",
")",
"except",
"(",
"ValueError",
",",
"IOError",
")",
":",
"pass",
"return",
"local_facts"
] | retrieve local facts from fact file args: filename : local facts file returns: dict: the retrieved facts . | train | false |
25,223 | def miniEditImages():
return {'Select': BitmapImage(file='/usr/include/X11/bitmaps/left_ptr'), 'Switch': PhotoImage(data='\nR0lGODlhLgAgAPcAAB2ZxGq61imex4zH3RWWwmK41tzd3vn9/jCiyfX7/Q6SwFay0gBlmtnZ2snJ\nyr+2tAuMu6rY6D6kyfHx8XO/2Uqszjmly6DU5uXz+JLN4uz3+kSrzlKx0ZeZm2K21BuYw67a6QB9\nr+Xl5rW2uHW61On1+UGpzbrf6xiXwny9166vsMLCwgBdlAmHt8TFxgBwpNTs9C2hyO7t7ZnR5L/B\nw0yv0NXV1gBimKGjpABtoQBuoqKkpiaUvqWmqHbB2/j4+Pf39729vgB/sN7w9obH3hSMugCAsonJ\n4M/q8wBglgB6rCCaxLO0tX7C2wBqniGMuABzpuPl5f3+/v39/fr6+r7i7vP6/ABonV621LLc6zWk\nyrq6uq6wskGlyUaszp6gohmYw8HDxKaoqn3E3LGztWGuzcnLzKmrrOnp6gB1qCaex1q001ewz+Dg\n4QB3qrCxstHS09LR0dHR0s7Oz8zNzsfIyQaJuQB0pozL4YzI3re4uAGFtYDG3hOUwb+/wQB5rOvr\n6wB2qdju9TWfxgBpniOcxeLj48vn8dvc3VKuzwB2qp6fos/Q0aXV6D+jxwB7rsXHyLu8vb27vCSc\nxSGZwxyZxH3A2RuUv0+uzz+ozCedxgCDtABnnABroKutr/7+/n2/2LTd6wBvo9bX2OLo6lGv0C6d\nxS6avjmmzLTR2uzr6m651RuXw4jF3CqfxySaxSadyAuRv9bd4cPExRiMuDKjyUWevNPS0sXl8BeY\nxKytr8G/wABypXvC23vD3O73+3vE3cvU2PH5+7S1t7q7vCGVwO/v8JfM3zymyyyZwrWys+Hy90Ki\nxK6qqg+TwBKXxMvMzaWtsK7U4jemzLXEygBxpW++2aCho97Z18bP0/T09fX29vb19ViuzdDR0crf\n51qd01y00ujo6Onq6hCDs2Gpw3i71CqWv3S71nO92M/h52m207bJ0AN6rPPz9Nrh5Nvo7K/b6oTI\n37Td7ABqneHi4yScxo/M4RiWwRqVwcro8n3B2lGoylStzszMzAAAACH5BAEAAP8ALAAAAAAuACAA\nBwj/AP8JHEjw3wEkEY74WOjrQhUNBSNKnCjRSoYKCOwJcKWpEAACBFBRGEKxZMkDjRAg2OBlQyYL\nWhDEcOWxDwofv0zqHIhhDYIFC2p4MYFMS62ZaiYVWlJJAYIqO00KMlEjABYOQokaRbp0CYBKffpE\niDpxSKYC1gqswToUmYVaCFyp6QrgwwcCscaSJZhgQYBeAdRyqFBhgwWkGyct8WoXRZ8Ph/YOxMOB\nCIUAHsBxwGQBAII1YwpMI5Brcd0PKFA4Q2ZFMgYteZqkwxyu1KQNJzQc+CdFCrxypyqdRoEPX6x7\nki/n2TfbAxtNRHYTVCWpWTRbuRoX7yMgZ9QSFQa0/7LU/BXygjIWXVOBTR2sxp7BxGpENgKbY+PR\nreqyIOKnOh0M445AjTjDCgrPSBNFKt9w8wMVU5g0Bg8kDAAKOutQAkNEQNBwDRAEeVEcAV6w84Ay\nKowQSRhmzNGAASIAYow2IP6DySPk8ANKCv1wINE2cpjxCUEgOIOPAKicQMMbKnhyhhg97HDNF4vs\nIEYkNkzwjwSP/PHIE2VIgIdEnxjAiBwNGIKGDKS8I0sw2VAzApNOQimGLlyMAIkDw2yhZTF/KKGE\nlxCEMtEPBtDhACQurLDCLkFIsoUeZLyRpx8OmEGHN3AEcU0HkFAhUDFulDroJvOU5M44iDjgDTQO\n1P/hzRw2IFJPGw3AAY0LI/SAwxc7jEKQI2mkEUipRoxp0g821AMIGlG0McockMzihx5c1LkDDmSg\nUVAiafACRbGPVKDTFG3MYUYdLoThRxDE6DEMGUww8eQONGwTER9piFINFOPasaFJVIjTwC1xzOGP\nA3HUKoIMDTwJR4QRgdBOJzq8UM0Lj5QihU5ZdGMOCSSYUwYzAwwkDhNtUKTBOZ10koMOoohihDwm\nHZKPEDwb4fMe9An0g5Yl+SDKFTHnkMMLLQAjXUTxUCLEIyH0bIQAwuxVQhEMcEIIIUmHUEsWGCQg\nxQEaIFGAHV0+QnUIIWwyg2T/3MPLDQwwcAUhTjiswYsQl1SAxQKmbBJCIMe6ISjVmXwsWQKJEJJE\n3l1/TY8O4wZyh8ZQ3IF4qX9cggTdAmEwCAMs3IB311fsDfbMGv97BxSBQBAP6QMN0QUhLCSRhOp5\ne923zDpk/EIaRdyO+0C/eHBHEiz0vjrrfMfciSKD4LJ8RBEk88IN0ff+O/CEVEPLGK1tH1ECM7Dx\nRDWdcMLJFTpUQ44jfCyjvlShZNDE/0QAgT6ypr6AAAA7\n '), 'LegacySwitch': PhotoImage(data='\nR0lGODlhMgAYAPcAAAEBAXmDjbe4uAE5cjF7xwFWq2Sa0S9biSlrrdTW1k2Ly02a5xUvSQFHjmep\n6bfI2Q5SlQIYLwFfvj6M3Jaan8fHyDuFzwFp0Vah60uU3AEiRhFgrgFRogFr10N9uTFrpytHYQFM\nmGWt9wIwX+bm5kaT4gtFgR1cnJPF9yt80CF0yAIMGHmp2c/P0AEoUb/P4Fei7qK4zgpLjgFkyQlf\nt1mf5jKD1WWJrQ86ZwFAgBhYmVOa4MPV52uv8y+A0iR3ywFbtUyX5ECI0Q1UmwIcOUGQ3RBXoQI0\naRJbpr3BxVeJvQUJDafH5wIlS2aq7xBmv52lr7fH12el5Wml3097ph1ru7vM3HCz91Ke6lid40KQ\n4GSQvgQGClFnfwVJjszMzVCX3hljrdPT1AFLlBRnutPf6yd5zjeI2QE9eRBdrBNVl+3v70mV4ydf\nlwMVKwErVlul8AFChTGB1QE3bsTFxQImTVmAp0FjiUSM1k+b6QQvWQ1SlxMgLgFixEqU3xJhsgFT\npn2Xs5OluZ+1yz1Xb6HN+Td9wy1zuYClykV5r0x2oeDh4qmvt8LDwxhuxRlLfyRioo2124mft9bi\n71mDr7fT79nl8Z2hpQs9b7vN4QMQIOPj5XOPrU2Jx32z6xtvwzeBywFFikFnjwcPFa29yxJjuFmP\nxQFv3qGxwRc/Z8vb6wsRGBNqwqmpqTdvqQIbNQFPngMzZAEfP0mQ13mHlQFYsAFnznOXu2mPtQxj\nvQ1Vn4Ot1+/x8my0/CJgnxNNh8DT5CdJaWyx+AELFWmt8QxPkxBZpwMFB015pgFduGCNuyx7zdnZ\n2WKm6h1xyOPp8aW70QtPkUmM0LrCyr/FyztljwFPm0OJzwFny7/L1xFjswE/e12i50iR2VR8o2Gf\n3xszS2eTvz2BxSlloQdJiwMHDzF3u7bJ3T2I1WCp8+Xt80FokQFJklef6mORw2ap7SJ1y77Q47nN\n3wFfu1Kb5cXJyxdhrdDR0wlNkTSF11Oa4yp4yQEuW0WQ3QIDBQI7dSH5BAEAAAAALAAAAAAyABgA\nBwj/AAEIHDjKF6SDvhImPMHwhA6HOiLqUENRDYSLEIplxBcNHz4Z5GTI8BLKS5OBA1Ply2fDhxwf\nPlLITGFmmRkzP+DlVKHCmU9nnz45csSqKKsn9gileZKrVC4aRFACOGZu5UobNuRohRkzhc2b+36o\nqCaqrFmzZEV1ERBg3BOmMl5JZTBhwhm7ZyycYZnvJdeuNl21qkCHTiPDhxspTtKoQgUKCJ6wehMV\n5QctWupeo6TkjOd8e1lmdQkTGbTTMaDFiDGINeskX6YhEicUiQa5A/kUKaFFwQ0oXzjZ8Tbcm3Hj\nirwpMtTSgg9QMJf5WEZ9375AiED19ImpSQSUB4Kw/8HFSMyiRWJaqG/xhf2X91+oCbmq1e/MFD/2\nEcApVkWVJhp8J9AqsywQxDfAbLJJPAy+kMkL8shjxTkUnhOJZ5+JVp8cKfhwxwdf4fQLgG4MFAwW\nKOZRAxM81EAPPQvoE0QQfrDhx4399OMBMjz2yCMVivCoCAWXKLKMTPvoUYcsKwi0RCcwYCAlFjU0\nA6OBM4pXAhsl8FYELYWFWZhiZCbRQgIC2AGTLy408coxAoEDx5wwtGPALTVg0E4NKC7gp4FsBKoA\nKi8U+oIVmVih6DnZPMBMAlGwIARWOLiggSYC+ZNIOulwY4AkSZCyxaikbqHMqaeaIp4+rAaxQxBg\n2P+IozuRzvLZIS4syYVAfMAhwhSC1EPCGoskIIYY9yS7Hny75OFnEIAGyiVvWkjjRxF11fXIG3WU\nKNA6wghDTCW88PKMJZOkm24Z7LarSjPtoIjFn1lKyyVmmBVhwRtvaDDMgFL0Eu4VhaiDwhXCXNFD\nD8QQw7ATEDsBw8RSxotFHs7CKJ60XWrRBj91EOGPQCA48c7J7zTjSTPctOzynjVkkYU+O9S8Axg4\nZ6BzBt30003Ps+AhNB5C4PCGC5gKJMMTZJBRytOl/CH1HxvQkMbVVxujtdZGGKGL17rsEfYQe+xR\nzNnFcGQCv7LsKlAtp8R9Sgd0032BLXjPoPcMffTd3YcEgAMOxOBA1GJ4AYgXAMjiHDTgggveCgRI\n3RfcnffefgcOeDKEG3444osDwgEspMNiTQhx5FoOShxcrrfff0uQjOycD+554qFzMHrpp4cwBju/\n5+CmVNbArnntndeCO+O689777+w0IH0o1P/TRJMohRA4EJwn47nyiocOSOmkn/57COxE3wD11Mfh\nfg45zCGyVF4Ufvvyze8ewv5jQK9++6FwXxzglwM0GPAfR8AeSo4gwAHCbxsQNCAa/kHBAVhwAHPI\n4BE2eIRYeHAEIBwBP0Y4Qn41YWRSCQgAOw==\n '), 'LegacyRouter': PhotoImage(data='\nR0lGODlhMgAYAPcAAAEBAXZ8gQNAgL29vQNctjl/xVSa4j1dfCF+3QFq1DmL3wJMmAMzZZW11dnZ\n2SFrtyNdmTSO6gIZMUKa8gJVqEOHzR9Pf5W74wFjxgFx4jltn+np6Eyi+DuT6qKiohdtwwUPGWiq\n6ymF4LHH3Rh11CV81kKT5AMoUA9dq1ap/mV0gxdXlytRdR1ptRNPjTt9vwNgvwJZsX+69gsXJQFH\njTtjizF0tvHx8VOm9z2V736Dhz2N3QM2acPZ70qe8gFo0HS19wVRnTiR6hMpP0eP1i6J5iNlqAtg\ntktjfQFu3TNxryx4xAMTIzOE1XqAh1uf5SWC4AcfNy1XgQJny93n8a2trRh312Gt+VGm/AQIDTmB\nyAF37QJasydzvxM/ayF3zhdLf8zLywFdu4i56gFlyi2J4yV/1w8wUo2/8j+X8D2Q5Eee9jeR7Uia\n7DpeggFt2QNPm97e3jRong9bpziH2DuT7aipqQoVICmG45vI9R5720eT4Q1hs1er/yVVhwJJktPh\n70tfdbHP7Xev5xs5V7W1sz9jhz11rUVZcQ9WoCVVhQk7cRdtwWuw9QYOFyFHbSBnr0dznxtWkS18\nzKfP9wwcLAMHCwFFiS5UeqGtuRNNiwMfPS1hlQMtWRE5XzGM5yhxusLCwCljnwMdOFWh7cve8pG/\n7Tlxp+Tr8g9bpXF3f0lheStrrYu13QEXLS1ppTV3uUuR1RMjNTF3vU2X4TZupwRSolNne4nB+T+L\n2YGz4zJ/zYe99YGHjRdDcT95sx09XQldsgMLEwMrVc/X3yN3yQ1JhTRbggsdMQNfu9HPz6WlpW2t\n7RctQ0GFyeHh4dvl8SBZklCb5kOO2kWR3Vmt/zdjkQIQHi90uvPz8wIVKBp42SV5zbfT7wtXpStV\nfwFWrBVvyTt3swFz5kGBv2+1/QlbrVFjdQM7d1+j54i67UmX51qn9i1vsy+D2TuR5zddhQsjOR1t\nu0GV6ghbsDVZf4+76RRisent8Xd9hQFBgwFNmwJLlcPDwwFr1z2T5yH5BAEAAAAALAAAAAAyABgA\nBwj/AAEIHEiQYJY7Qwg9UsTplRIbENuxEiXJgpcz8e5YKsixY8Essh7JcbbOBwcOa1JOmJAmTY4c\nHeoIabJrCShI0XyB8YRso0eOjoAdWpciBZajJ1GuWcnSZY46Ed5N8hPATqEBoRB9gVJsxRlhPwHI\n0kDkVywcRpGe9LF0adOnMpt8CxDnxg1o9lphKoEACoIvmlxxvHOKVg0n/Tzku2WoVoU2J1P6WNkS\nrtwADuxCG/MOjwgRUEIjGG3FhaOBzaThiDSCil27G8Isc3LLjZwXsA6YYJmDjhTMmseoKQIFDx7R\noxHo2abnwygAlUj1mV6tWjlelEpRwfd6gzI7VeJQ/2vZoVaDUqigqftXpH0R46H9Kl++zUo4JnKq\n9dGvv09RHFhcIUMe0NiFDyql0OJUHWywMc87TXRhhCRGiHAccvNZUR8JxpDTH38p9HEUFhxgMSAv\njbBjQge8PSXEC6uo0IsHA6gAAShmgCbffNtsQwIJifhRHX/TpUUiSijlUk8AqgQixSwdNBjCa7CF\noVggmEgCyRf01WcFCYvYUgB104k4YlK5HONEXXfpokYdMrXRAzMhmNINNNzB9p0T57AgyZckpKKP\nGFNgw06ZWKR10jTw6MAmFWj4AJcQQkQQwSefvFeGCemMIQggeaJywSQ/wgHOAmJskQEfWqBlFBEH\n1P/QaGY3QOpDZXA2+A6m7hl3IRQKGDCIAj6iwE8yGKC6xbJv8IHNHgACQQybN2QiTi5NwdlBpZdi\nisd7vyanByOJ7CMGGRhgwE+qyy47DhnBPLDLEzLIAEQjBtChRmVPNWgpr+Be+Nc9icARww9TkIEu\nDAsQ0O7DzGIQzD2QdDEJHTsIAROc3F7qWQncyHPPHN5QQAAG/vjzw8oKp8sPPxDH3O44/kwBQzLB\nxBCMOTzzHEMMBMBARgJvZJBBEm/4k0ACKydMBgwYoKNNEjJXbTXE42Q9jtFIp8z0Dy1jQMA1AGzi\nz9VoW7310V0znYDTGMQgwUDXLDBO2nhvoTXbbyRk/XXL+pxWkAT8UJ331WsbnbTSK8MggDZhCTOM\nLQkcjvXeSPedAAw0nABWWARZIgEDfyTzxt15Z53BG1PEcEknrvgEelhZMDHKCTwI8EcQFHBBAAFc\ngGPLHwLwcMIo12Qxu0ABAQA7\n '), 'Controller': PhotoImage(data='\n R0lGODlhMAAwAPcAAAEBAWfNAYWFhcfHx+3t6/f390lJUaWlpfPz8/Hx72lpaZGRke/v77m5uc0B\n AeHh4e/v7WNjY3t7e5eXlyMjI4mJidPT0+3t7f///09PT7Ozs/X19fHx8ZWTk8HBwX9/fwAAAAAA\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAACH5BAEAAAAALAAAAAAwADAA\n Bwj/AAEIHEiwoMGDCBMqXMiwocOHECNKnEixosWLGAEIeMCxo8ePHwVkBGABg8mTKFOmtDByAIYN\n MGPCRCCzQIENNzEMGOkBAwIKQIMKpYCgKAIHCDB4GNkAA4OnUJ9++CDhQ1QGFzA0GKkBA4GvYMOK\n BYtBA1cNaNOqXcuWq8q3b81m7Cqdbk2bMMu6/Tl0qFEEAZLKxdj1KlSqVA3rnet1rOOwiwmznUzZ\n LdzLJgdfpIv3pmebN2Pm1GyRbocNp1PLNMDaAM3Im1/alQk4gO28pCt2RdCBt+/eRg8IP1AUdmmf\n f5MrL56bYlcOvaP7Xo6Ag3HdGDho3869u/YE1507t+3AgLz58ujPMwg/sTBUCAzgy49PH0LW5u0x\n XFiwvz////5dcJ9bjxVIAHsSdUXAAgs2yOCDDn6FYEQaFGDgYxNCpEFfHHKIX4IDhCjiiCSS+CGF\n FlCmogYpcnVABTDGKGOMAlRQYwUHnKjhAjX2aOOPN8LImgAL6PiQBhLMqCSNAThQgQRGOqRBBD1W\n aaOVAggnQARRNqRBBxmEKeaYZIrZQZcMKbDiigqM5OabcMYp55x01ilnQAA7\n '), 'Host': PhotoImage(data='\n R0lGODlhIAAYAPcAMf//////zP//mf//Zv//M///AP/M///MzP/M\n mf/MZv/MM//MAP+Z//+ZzP+Zmf+ZZv+ZM/+ZAP9m//9mzP9mmf9m\n Zv9mM/9mAP8z//8zzP8zmf8zZv8zM/8zAP8A//8AzP8Amf8AZv8A\n M/8AAMz//8z/zMz/mcz/Zsz/M8z/AMzM/8zMzMzMmczMZszMM8zM\n AMyZ/8yZzMyZmcyZZsyZM8yZAMxm/8xmzMxmmcxmZsxmM8xmAMwz\n /8wzzMwzmcwzZswzM8wzAMwA/8wAzMwAmcwAZswAM8wAAJn//5n/\n zJn/mZn/Zpn/M5n/AJnM/5nMzJnMmZnMZpnMM5nMAJmZ/5mZzJmZ\n mZmZZpmZM5mZAJlm/5lmzJlmmZlmZplmM5lmAJkz/5kzzJkzmZkz\n ZpkzM5kzAJkA/5kAzJkAmZkAZpkAM5kAAGb//2b/zGb/mWb/Zmb/\n M2b/AGbM/2bMzGbMmWbMZmbMM2bMAGaZ/2aZzGaZmWaZZmaZM2aZ\n AGZm/2ZmzGZmmWZmZmZmM2ZmAGYz/2YzzGYzmWYzZmYzM2YzAGYA\n /2YAzGYAmWYAZmYAM2YAADP//zP/zDP/mTP/ZjP/MzP/ADPM/zPM\n zDPMmTPMZjPMMzPMADOZ/zOZzDOZmTOZZjOZMzOZADNm/zNmzDNm\n mTNmZjNmMzNmADMz/zMzzDMzmTMzZjMzMzMzADMA/zMAzDMAmTMA\n ZjMAMzMAAAD//wD/zAD/mQD/ZgD/MwD/AADM/wDMzADMmQDMZgDM\n MwDMAACZ/wCZzACZmQCZZgCZMwCZAABm/wBmzABmmQBmZgBmMwBm\n AAAz/wAzzAAzmQAzZgAzMwAzAAAA/wAAzAAAmQAAZgAAM+4AAN0A\n ALsAAKoAAIgAAHcAAFUAAEQAACIAABEAAADuAADdAAC7AACqAACI\n AAB3AABVAABEAAAiAAARAAAA7gAA3QAAuwAAqgAAiAAAdwAAVQAA\n RAAAIgAAEe7u7t3d3bu7u6qqqoiIiHd3d1VVVURERCIiIhEREQAA\n ACH5BAEAAAAALAAAAAAgABgAAAiNAAH8G0iwoMGDCAcKTMiw4UBw\n BPXVm0ixosWLFvVBHFjPoUeC9Tb+6/jRY0iQ/8iVbHiS40CVKxG2\n HEkQZsyCM0mmvGkw50uePUV2tEnOZkyfQA8iTYpTKNOgKJ+C3AhO\n p9SWVaVOfWj1KdauTL9q5UgVbFKsEjGqXVtP40NwcBnCjXtw7tx/\n C8cSBBAQADs=\n '), 'OldSwitch': PhotoImage(data='\n R0lGODlhIAAYAPcAMf//////zP//mf//Zv//M///AP/M///MzP/M\n mf/MZv/MM//MAP+Z//+ZzP+Zmf+ZZv+ZM/+ZAP9m//9mzP9mmf9m\n Zv9mM/9mAP8z//8zzP8zmf8zZv8zM/8zAP8A//8AzP8Amf8AZv8A\n M/8AAMz//8z/zMz/mcz/Zsz/M8z/AMzM/8zMzMzMmczMZszMM8zM\n AMyZ/8yZzMyZmcyZZsyZM8yZAMxm/8xmzMxmmcxmZsxmM8xmAMwz\n /8wzzMwzmcwzZswzM8wzAMwA/8wAzMwAmcwAZswAM8wAAJn//5n/\n zJn/mZn/Zpn/M5n/AJnM/5nMzJnMmZnMZpnMM5nMAJmZ/5mZzJmZ\n mZmZZpmZM5mZAJlm/5lmzJlmmZlmZplmM5lmAJkz/5kzzJkzmZkz\n ZpkzM5kzAJkA/5kAzJkAmZkAZpkAM5kAAGb//2b/zGb/mWb/Zmb/\n M2b/AGbM/2bMzGbMmWbMZmbMM2bMAGaZ/2aZzGaZmWaZZmaZM2aZ\n AGZm/2ZmzGZmmWZmZmZmM2ZmAGYz/2YzzGYzmWYzZmYzM2YzAGYA\n /2YAzGYAmWYAZmYAM2YAADP//zP/zDP/mTP/ZjP/MzP/ADPM/zPM\n zDPMmTPMZjPMMzPMADOZ/zOZzDOZmTOZZjOZMzOZADNm/zNmzDNm\n mTNmZjNmMzNmADMz/zMzzDMzmTMzZjMzMzMzADMA/zMAzDMAmTMA\n ZjMAMzMAAAD//wD/zAD/mQD/ZgD/MwD/AADM/wDMzADMmQDMZgDM\n MwDMAACZ/wCZzACZmQCZZgCZMwCZAABm/wBmzABmmQBmZgBmMwBm\n AAAz/wAzzAAzmQAzZgAzMwAzAAAA/wAAzAAAmQAAZgAAM+4AAN0A\n ALsAAKoAAIgAAHcAAFUAAEQAACIAABEAAADuAADdAAC7AACqAACI\n AAB3AABVAABEAAAiAAARAAAA7gAA3QAAuwAAqgAAiAAAdwAAVQAA\n RAAAIgAAEe7u7t3d3bu7u6qqqoiIiHd3d1VVVURERCIiIhEREQAA\n ACH5BAEAAAAALAAAAAAgABgAAAhwAAEIHEiwoMGDCBMqXMiwocOH\n ECNKnEixosWB3zJq3Mixo0eNAL7xG0mypMmTKPl9Cznyn8uWL/m5\n /AeTpsyYI1eKlBnO5r+eLYHy9Ck0J8ubPmPOrMmUpM6UUKMa/Ui1\n 6saLWLNq3cq1q9evYB0GBAA7\n '), 'NetLink': PhotoImage(data='\n R0lGODlhFgAWAPcAMf//////zP//mf//Zv//M///AP/M///MzP/M\n mf/MZv/MM//MAP+Z//+ZzP+Zmf+ZZv+ZM/+ZAP9m//9mzP9mmf9m\n Zv9mM/9mAP8z//8zzP8zmf8zZv8zM/8zAP8A//8AzP8Amf8AZv8A\n M/8AAMz//8z/zMz/mcz/Zsz/M8z/AMzM/8zMzMzMmczMZszMM8zM\n AMyZ/8yZzMyZmcyZZsyZM8yZAMxm/8xmzMxmmcxmZsxmM8xmAMwz\n /8wzzMwzmcwzZswzM8wzAMwA/8wAzMwAmcwAZswAM8wAAJn//5n/\n zJn/mZn/Zpn/M5n/AJnM/5nMzJnMmZnMZpnMM5nMAJmZ/5mZzJmZ\n mZmZZpmZM5mZAJlm/5lmzJlmmZlmZplmM5lmAJkz/5kzzJkzmZkz\n ZpkzM5kzAJkA/5kAzJkAmZkAZpkAM5kAAGb//2b/zGb/mWb/Zmb/\n M2b/AGbM/2bMzGbMmWbMZmbMM2bMAGaZ/2aZzGaZmWaZZmaZM2aZ\n AGZm/2ZmzGZmmWZmZmZmM2ZmAGYz/2YzzGYzmWYzZmYzM2YzAGYA\n /2YAzGYAmWYAZmYAM2YAADP//zP/zDP/mTP/ZjP/MzP/ADPM/zPM\n zDPMmTPMZjPMMzPMADOZ/zOZzDOZmTOZZjOZMzOZADNm/zNmzDNm\n mTNmZjNmMzNmADMz/zMzzDMzmTMzZjMzMzMzADMA/zMAzDMAmTMA\n ZjMAMzMAAAD//wD/zAD/mQD/ZgD/MwD/AADM/wDMzADMmQDMZgDM\n MwDMAACZ/wCZzACZmQCZZgCZMwCZAABm/wBmzABmmQBmZgBmMwBm\n AAAz/wAzzAAzmQAzZgAzMwAzAAAA/wAAzAAAmQAAZgAAM+4AAN0A\n ALsAAKoAAIgAAHcAAFUAAEQAACIAABEAAADuAADdAAC7AACqAACI\n AAB3AABVAABEAAAiAAARAAAA7gAA3QAAuwAAqgAAiAAAdwAAVQAA\n RAAAIgAAEe7u7t3d3bu7u6qqqoiIiHd3d1VVVURERCIiIhEREQAA\n ACH5BAEAAAAALAAAAAAWABYAAAhIAAEIHEiwoEGBrhIeXEgwoUKG\n Cx0+hGhQoiuKBy1irChxY0GNHgeCDAlgZEiTHlFuVImRJUWXEGEy\n lBmxI8mSNknm1Dnx5sCAADs=\n ')}
| [
"def",
"miniEditImages",
"(",
")",
":",
"return",
"{",
"'Select'",
":",
"BitmapImage",
"(",
"file",
"=",
"'/usr/include/X11/bitmaps/left_ptr'",
")",
",",
"'Switch'",
":",
"PhotoImage",
"(",
"data",
"=",
"'\\nR0lGODlhLgAgAPcAAB2ZxGq61imex4zH3RWWwmK41tzd3vn9/jCiyfX7/Q6SwFay0gBlmtnZ2snJ\\nyr+2tAuMu6rY6D6kyfHx8XO/2Uqszjmly6DU5uXz+JLN4uz3+kSrzlKx0ZeZm2K21BuYw67a6QB9\\nr+Xl5rW2uHW61On1+UGpzbrf6xiXwny9166vsMLCwgBdlAmHt8TFxgBwpNTs9C2hyO7t7ZnR5L/B\\nw0yv0NXV1gBimKGjpABtoQBuoqKkpiaUvqWmqHbB2/j4+Pf39729vgB/sN7w9obH3hSMugCAsonJ\\n4M/q8wBglgB6rCCaxLO0tX7C2wBqniGMuABzpuPl5f3+/v39/fr6+r7i7vP6/ABonV621LLc6zWk\\nyrq6uq6wskGlyUaszp6gohmYw8HDxKaoqn3E3LGztWGuzcnLzKmrrOnp6gB1qCaex1q001ewz+Dg\\n4QB3qrCxstHS09LR0dHR0s7Oz8zNzsfIyQaJuQB0pozL4YzI3re4uAGFtYDG3hOUwb+/wQB5rOvr\\n6wB2qdju9TWfxgBpniOcxeLj48vn8dvc3VKuzwB2qp6fos/Q0aXV6D+jxwB7rsXHyLu8vb27vCSc\\nxSGZwxyZxH3A2RuUv0+uzz+ozCedxgCDtABnnABroKutr/7+/n2/2LTd6wBvo9bX2OLo6lGv0C6d\\nxS6avjmmzLTR2uzr6m651RuXw4jF3CqfxySaxSadyAuRv9bd4cPExRiMuDKjyUWevNPS0sXl8BeY\\nxKytr8G/wABypXvC23vD3O73+3vE3cvU2PH5+7S1t7q7vCGVwO/v8JfM3zymyyyZwrWys+Hy90Ki\\nxK6qqg+TwBKXxMvMzaWtsK7U4jemzLXEygBxpW++2aCho97Z18bP0/T09fX29vb19ViuzdDR0crf\\n51qd01y00ujo6Onq6hCDs2Gpw3i71CqWv3S71nO92M/h52m207bJ0AN6rPPz9Nrh5Nvo7K/b6oTI\\n37Td7ABqneHi4yScxo/M4RiWwRqVwcro8n3B2lGoylStzszMzAAAACH5BAEAAP8ALAAAAAAuACAA\\nBwj/AP8JHEjw3wEkEY74WOjrQhUNBSNKnCjRSoYKCOwJcKWpEAACBFBRGEKxZMkDjRAg2OBlQyYL\\nWhDEcOWxDwofv0zqHIhhDYIFC2p4MYFMS62ZaiYVWlJJAYIqO00KMlEjABYOQokaRbp0CYBKffpE\\niDpxSKYC1gqswToUmYVaCFyp6QrgwwcCscaSJZhgQYBeAdRyqFBhgwWkGyct8WoXRZ8Ph/YOxMOB\\nCIUAHsBxwGQBAII1YwpMI5Brcd0PKFA4Q2ZFMgYteZqkwxyu1KQNJzQc+CdFCrxypyqdRoEPX6x7\\nki/n2TfbAxtNRHYTVCWpWTRbuRoX7yMgZ9QSFQa0/7LU/BXygjIWXVOBTR2sxp7BxGpENgKbY+PR\\nreqyIOKnOh0M445AjTjDCgrPSBNFKt9w8wMVU5g0Bg8kDAAKOutQAkNEQNBwDRAEeVEcAV6w84Ay\\nKowQSRhmzNGAASIAYow2IP6DySPk8ANKCv1wINE2cpjxCUEgOIOPAKicQMMbKnhyhhg97HDNF4vs\\nIEYkNkzwjwSP/PHIE2VIgIdEnxjAiBwNGIKGDKS8I0sw2VAzApNOQimGLlyMAIkDw2yhZTF/KKGE\\nlxCEMtEPBtDhACQurLDCLkFIsoUeZLyRpx8OmEGHN3AEcU0HkFAhUDFulDroJvOU5M44iDjgDTQO\\n1P/hzRw2IFJPGw3AAY0LI/SAwxc7jEKQI2mkEUipRoxp0g821AMIGlG0McockMzihx5c1LkDDmSg\\nUVAiafACRbGPVKDTFG3MYUYdLoThRxDE6DEMGUww8eQONGwTER9piFINFOPasaFJVIjTwC1xzOGP\\nA3HUKoIMDTwJR4QRgdBOJzq8UM0Lj5QihU5ZdGMOCSSYUwYzAwwkDhNtUKTBOZ10koMOoohihDwm\\nHZKPEDwb4fMe9An0g5Yl+SDKFTHnkMMLLQAjXUTxUCLEIyH0bIQAwuxVQhEMcEIIIUmHUEsWGCQg\\nxQEaIFGAHV0+QnUIIWwyg2T/3MPLDQwwcAUhTjiswYsQl1SAxQKmbBJCIMe6ISjVmXwsWQKJEJJE\\n3l1/TY8O4wZyh8ZQ3IF4qX9cggTdAmEwCAMs3IB311fsDfbMGv97BxSBQBAP6QMN0QUhLCSRhOp5\\ne923zDpk/EIaRdyO+0C/eHBHEiz0vjrrfMfciSKD4LJ8RBEk88IN0ff+O/CEVEPLGK1tH1ECM7Dx\\nRDWdcMLJFTpUQ44jfCyjvlShZNDE/0QAgT6ypr6AAAA7\\n '",
")",
",",
"'LegacySwitch'",
":",
"PhotoImage",
"(",
"data",
"=",
"'\\nR0lGODlhMgAYAPcAAAEBAXmDjbe4uAE5cjF7xwFWq2Sa0S9biSlrrdTW1k2Ly02a5xUvSQFHjmep\\n6bfI2Q5SlQIYLwFfvj6M3Jaan8fHyDuFzwFp0Vah60uU3AEiRhFgrgFRogFr10N9uTFrpytHYQFM\\nmGWt9wIwX+bm5kaT4gtFgR1cnJPF9yt80CF0yAIMGHmp2c/P0AEoUb/P4Fei7qK4zgpLjgFkyQlf\\nt1mf5jKD1WWJrQ86ZwFAgBhYmVOa4MPV52uv8y+A0iR3ywFbtUyX5ECI0Q1UmwIcOUGQ3RBXoQI0\\naRJbpr3BxVeJvQUJDafH5wIlS2aq7xBmv52lr7fH12el5Wml3097ph1ru7vM3HCz91Ke6lid40KQ\\n4GSQvgQGClFnfwVJjszMzVCX3hljrdPT1AFLlBRnutPf6yd5zjeI2QE9eRBdrBNVl+3v70mV4ydf\\nlwMVKwErVlul8AFChTGB1QE3bsTFxQImTVmAp0FjiUSM1k+b6QQvWQ1SlxMgLgFixEqU3xJhsgFT\\npn2Xs5OluZ+1yz1Xb6HN+Td9wy1zuYClykV5r0x2oeDh4qmvt8LDwxhuxRlLfyRioo2124mft9bi\\n71mDr7fT79nl8Z2hpQs9b7vN4QMQIOPj5XOPrU2Jx32z6xtvwzeBywFFikFnjwcPFa29yxJjuFmP\\nxQFv3qGxwRc/Z8vb6wsRGBNqwqmpqTdvqQIbNQFPngMzZAEfP0mQ13mHlQFYsAFnznOXu2mPtQxj\\nvQ1Vn4Ot1+/x8my0/CJgnxNNh8DT5CdJaWyx+AELFWmt8QxPkxBZpwMFB015pgFduGCNuyx7zdnZ\\n2WKm6h1xyOPp8aW70QtPkUmM0LrCyr/FyztljwFPm0OJzwFny7/L1xFjswE/e12i50iR2VR8o2Gf\\n3xszS2eTvz2BxSlloQdJiwMHDzF3u7bJ3T2I1WCp8+Xt80FokQFJklef6mORw2ap7SJ1y77Q47nN\\n3wFfu1Kb5cXJyxdhrdDR0wlNkTSF11Oa4yp4yQEuW0WQ3QIDBQI7dSH5BAEAAAAALAAAAAAyABgA\\nBwj/AAEIHDjKF6SDvhImPMHwhA6HOiLqUENRDYSLEIplxBcNHz4Z5GTI8BLKS5OBA1Ply2fDhxwf\\nPlLITGFmmRkzP+DlVKHCmU9nnz45csSqKKsn9gileZKrVC4aRFACOGZu5UobNuRohRkzhc2b+36o\\nqCaqrFmzZEV1ERBg3BOmMl5JZTBhwhm7ZyycYZnvJdeuNl21qkCHTiPDhxspTtKoQgUKCJ6wehMV\\n5QctWupeo6TkjOd8e1lmdQkTGbTTMaDFiDGINeskX6YhEicUiQa5A/kUKaFFwQ0oXzjZ8Tbcm3Hj\\nirwpMtTSgg9QMJf5WEZ9375AiED19ImpSQSUB4Kw/8HFSMyiRWJaqG/xhf2X91+oCbmq1e/MFD/2\\nEcApVkWVJhp8J9AqsywQxDfAbLJJPAy+kMkL8shjxTkUnhOJZ5+JVp8cKfhwxwdf4fQLgG4MFAwW\\nKOZRAxM81EAPPQvoE0QQfrDhx4399OMBMjz2yCMVivCoCAWXKLKMTPvoUYcsKwi0RCcwYCAlFjU0\\nA6OBM4pXAhsl8FYELYWFWZhiZCbRQgIC2AGTLy408coxAoEDx5wwtGPALTVg0E4NKC7gp4FsBKoA\\nKi8U+oIVmVih6DnZPMBMAlGwIARWOLiggSYC+ZNIOulwY4AkSZCyxaikbqHMqaeaIp4+rAaxQxBg\\n2P+IozuRzvLZIS4syYVAfMAhwhSC1EPCGoskIIYY9yS7Hny75OFnEIAGyiVvWkjjRxF11fXIG3WU\\nKNA6wghDTCW88PKMJZOkm24Z7LarSjPtoIjFn1lKyyVmmBVhwRtvaDDMgFL0Eu4VhaiDwhXCXNFD\\nD8QQw7ATEDsBw8RSxotFHs7CKJ60XWrRBj91EOGPQCA48c7J7zTjSTPctOzynjVkkYU+O9S8Axg4\\nZ6BzBt30003Ps+AhNB5C4PCGC5gKJMMTZJBRytOl/CH1HxvQkMbVVxujtdZGGKGL17rsEfYQe+xR\\nzNnFcGQCv7LsKlAtp8R9Sgd0032BLXjPoPcMffTd3YcEgAMOxOBA1GJ4AYgXAMjiHDTgggveCgRI\\n3RfcnffefgcOeDKEG3444osDwgEspMNiTQhx5FoOShxcrrfff0uQjOycD+554qFzMHrpp4cwBju/\\n5+CmVNbArnntndeCO+O689777+w0IH0o1P/TRJMohRA4EJwn47nyiocOSOmkn/57COxE3wD11Mfh\\nfg45zCGyVF4Ufvvyze8ewv5jQK9++6FwXxzglwM0GPAfR8AeSo4gwAHCbxsQNCAa/kHBAVhwAHPI\\n4BE2eIRYeHAEIBwBP0Y4Qn41YWRSCQgAOw==\\n '",
")",
",",
"'LegacyRouter'",
":",
"PhotoImage",
"(",
"data",
"=",
"'\\nR0lGODlhMgAYAPcAAAEBAXZ8gQNAgL29vQNctjl/xVSa4j1dfCF+3QFq1DmL3wJMmAMzZZW11dnZ\\n2SFrtyNdmTSO6gIZMUKa8gJVqEOHzR9Pf5W74wFjxgFx4jltn+np6Eyi+DuT6qKiohdtwwUPGWiq\\n6ymF4LHH3Rh11CV81kKT5AMoUA9dq1ap/mV0gxdXlytRdR1ptRNPjTt9vwNgvwJZsX+69gsXJQFH\\njTtjizF0tvHx8VOm9z2V736Dhz2N3QM2acPZ70qe8gFo0HS19wVRnTiR6hMpP0eP1i6J5iNlqAtg\\ntktjfQFu3TNxryx4xAMTIzOE1XqAh1uf5SWC4AcfNy1XgQJny93n8a2trRh312Gt+VGm/AQIDTmB\\nyAF37QJasydzvxM/ayF3zhdLf8zLywFdu4i56gFlyi2J4yV/1w8wUo2/8j+X8D2Q5Eee9jeR7Uia\\n7DpeggFt2QNPm97e3jRong9bpziH2DuT7aipqQoVICmG45vI9R5720eT4Q1hs1er/yVVhwJJktPh\\n70tfdbHP7Xev5xs5V7W1sz9jhz11rUVZcQ9WoCVVhQk7cRdtwWuw9QYOFyFHbSBnr0dznxtWkS18\\nzKfP9wwcLAMHCwFFiS5UeqGtuRNNiwMfPS1hlQMtWRE5XzGM5yhxusLCwCljnwMdOFWh7cve8pG/\\n7Tlxp+Tr8g9bpXF3f0lheStrrYu13QEXLS1ppTV3uUuR1RMjNTF3vU2X4TZupwRSolNne4nB+T+L\\n2YGz4zJ/zYe99YGHjRdDcT95sx09XQldsgMLEwMrVc/X3yN3yQ1JhTRbggsdMQNfu9HPz6WlpW2t\\n7RctQ0GFyeHh4dvl8SBZklCb5kOO2kWR3Vmt/zdjkQIQHi90uvPz8wIVKBp42SV5zbfT7wtXpStV\\nfwFWrBVvyTt3swFz5kGBv2+1/QlbrVFjdQM7d1+j54i67UmX51qn9i1vsy+D2TuR5zddhQsjOR1t\\nu0GV6ghbsDVZf4+76RRisent8Xd9hQFBgwFNmwJLlcPDwwFr1z2T5yH5BAEAAAAALAAAAAAyABgA\\nBwj/AAEIHEiQYJY7Qwg9UsTplRIbENuxEiXJgpcz8e5YKsixY8Essh7JcbbOBwcOa1JOmJAmTY4c\\nHeoIabJrCShI0XyB8YRso0eOjoAdWpciBZajJ1GuWcnSZY46Ed5N8hPATqEBoRB9gVJsxRlhPwHI\\n0kDkVywcRpGe9LF0adOnMpt8CxDnxg1o9lphKoEACoIvmlxxvHOKVg0n/Tzku2WoVoU2J1P6WNkS\\nrtwADuxCG/MOjwgRUEIjGG3FhaOBzaThiDSCil27G8Isc3LLjZwXsA6YYJmDjhTMmseoKQIFDx7R\\noxHo2abnwygAlUj1mV6tWjlelEpRwfd6gzI7VeJQ/2vZoVaDUqigqftXpH0R46H9Kl++zUo4JnKq\\n9dGvv09RHFhcIUMe0NiFDyql0OJUHWywMc87TXRhhCRGiHAccvNZUR8JxpDTH38p9HEUFhxgMSAv\\njbBjQge8PSXEC6uo0IsHA6gAAShmgCbffNtsQwIJifhRHX/TpUUiSijlUk8AqgQixSwdNBjCa7CF\\noVggmEgCyRf01WcFCYvYUgB104k4YlK5HONEXXfpokYdMrXRAzMhmNINNNzB9p0T57AgyZckpKKP\\nGFNgw06ZWKR10jTw6MAmFWj4AJcQQkQQwSefvFeGCemMIQggeaJywSQ/wgHOAmJskQEfWqBlFBEH\\n1P/QaGY3QOpDZXA2+A6m7hl3IRQKGDCIAj6iwE8yGKC6xbJv8IHNHgACQQybN2QiTi5NwdlBpZdi\\nisd7vyanByOJ7CMGGRhgwE+qyy47DhnBPLDLEzLIAEQjBtChRmVPNWgpr+Be+Nc9icARww9TkIEu\\nDAsQ0O7DzGIQzD2QdDEJHTsIAROc3F7qWQncyHPPHN5QQAAG/vjzw8oKp8sPPxDH3O44/kwBQzLB\\nxBCMOTzzHEMMBMBARgJvZJBBEm/4k0ACKydMBgwYoKNNEjJXbTXE42Q9jtFIp8z0Dy1jQMA1AGzi\\nz9VoW7310V0znYDTGMQgwUDXLDBO2nhvoTXbbyRk/XXL+pxWkAT8UJ331WsbnbTSK8MggDZhCTOM\\nLQkcjvXeSPedAAw0nABWWARZIgEDfyTzxt15Z53BG1PEcEknrvgEelhZMDHKCTwI8EcQFHBBAAFc\\ngGPLHwLwcMIo12Qxu0ABAQA7\\n '",
")",
",",
"'Controller'",
":",
"PhotoImage",
"(",
"data",
"=",
"'\\n R0lGODlhMAAwAPcAAAEBAWfNAYWFhcfHx+3t6/f390lJUaWlpfPz8/Hx72lpaZGRke/v77m5uc0B\\n AeHh4e/v7WNjY3t7e5eXlyMjI4mJidPT0+3t7f///09PT7Ozs/X19fHx8ZWTk8HBwX9/fwAAAAAA\\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAACH5BAEAAAAALAAAAAAwADAA\\n Bwj/AAEIHEiwoMGDCBMqXMiwocOHECNKnEixosWLGAEIeMCxo8ePHwVkBGABg8mTKFOmtDByAIYN\\n MGPCRCCzQIENNzEMGOkBAwIKQIMKpYCgKAIHCDB4GNkAA4OnUJ9++CDhQ1QGFzA0GKkBA4GvYMOK\\n BYtBA1cNaNOqXcuWq8q3b81m7Cqdbk2bMMu6/Tl0qFEEAZLKxdj1KlSqVA3rnet1rOOwiwmznUzZ\\n LdzLJgdfpIv3pmebN2Pm1GyRbocNp1PLNMDaAM3Im1/alQk4gO28pCt2RdCBt+/eRg8IP1AUdmmf\\n f5MrL56bYlcOvaP7Xo6Ag3HdGDho3869u/YE1507t+3AgLz58ujPMwg/sTBUCAzgy49PH0LW5u0x\\n XFiwvz////5dcJ9bjxVIAHsSdUXAAgs2yOCDDn6FYEQaFGDgYxNCpEFfHHKIX4IDhCjiiCSS+CGF\\n FlCmogYpcnVABTDGKGOMAlRQYwUHnKjhAjX2aOOPN8LImgAL6PiQBhLMqCSNAThQgQRGOqRBBD1W\\n aaOVAggnQARRNqRBBxmEKeaYZIrZQZcMKbDiigqM5OabcMYp55x01ilnQAA7\\n '",
")",
",",
"'Host'",
":",
"PhotoImage",
"(",
"data",
"=",
"'\\n R0lGODlhIAAYAPcAMf//////zP//mf//Zv//M///AP/M///MzP/M\\n mf/MZv/MM//MAP+Z//+ZzP+Zmf+ZZv+ZM/+ZAP9m//9mzP9mmf9m\\n Zv9mM/9mAP8z//8zzP8zmf8zZv8zM/8zAP8A//8AzP8Amf8AZv8A\\n M/8AAMz//8z/zMz/mcz/Zsz/M8z/AMzM/8zMzMzMmczMZszMM8zM\\n AMyZ/8yZzMyZmcyZZsyZM8yZAMxm/8xmzMxmmcxmZsxmM8xmAMwz\\n /8wzzMwzmcwzZswzM8wzAMwA/8wAzMwAmcwAZswAM8wAAJn//5n/\\n zJn/mZn/Zpn/M5n/AJnM/5nMzJnMmZnMZpnMM5nMAJmZ/5mZzJmZ\\n mZmZZpmZM5mZAJlm/5lmzJlmmZlmZplmM5lmAJkz/5kzzJkzmZkz\\n ZpkzM5kzAJkA/5kAzJkAmZkAZpkAM5kAAGb//2b/zGb/mWb/Zmb/\\n M2b/AGbM/2bMzGbMmWbMZmbMM2bMAGaZ/2aZzGaZmWaZZmaZM2aZ\\n AGZm/2ZmzGZmmWZmZmZmM2ZmAGYz/2YzzGYzmWYzZmYzM2YzAGYA\\n /2YAzGYAmWYAZmYAM2YAADP//zP/zDP/mTP/ZjP/MzP/ADPM/zPM\\n zDPMmTPMZjPMMzPMADOZ/zOZzDOZmTOZZjOZMzOZADNm/zNmzDNm\\n mTNmZjNmMzNmADMz/zMzzDMzmTMzZjMzMzMzADMA/zMAzDMAmTMA\\n ZjMAMzMAAAD//wD/zAD/mQD/ZgD/MwD/AADM/wDMzADMmQDMZgDM\\n MwDMAACZ/wCZzACZmQCZZgCZMwCZAABm/wBmzABmmQBmZgBmMwBm\\n AAAz/wAzzAAzmQAzZgAzMwAzAAAA/wAAzAAAmQAAZgAAM+4AAN0A\\n ALsAAKoAAIgAAHcAAFUAAEQAACIAABEAAADuAADdAAC7AACqAACI\\n AAB3AABVAABEAAAiAAARAAAA7gAA3QAAuwAAqgAAiAAAdwAAVQAA\\n RAAAIgAAEe7u7t3d3bu7u6qqqoiIiHd3d1VVVURERCIiIhEREQAA\\n ACH5BAEAAAAALAAAAAAgABgAAAiNAAH8G0iwoMGDCAcKTMiw4UBw\\n BPXVm0ixosWLFvVBHFjPoUeC9Tb+6/jRY0iQ/8iVbHiS40CVKxG2\\n HEkQZsyCM0mmvGkw50uePUV2tEnOZkyfQA8iTYpTKNOgKJ+C3AhO\\n p9SWVaVOfWj1KdauTL9q5UgVbFKsEjGqXVtP40NwcBnCjXtw7tx/\\n C8cSBBAQADs=\\n '",
")",
",",
"'OldSwitch'",
":",
"PhotoImage",
"(",
"data",
"=",
"'\\n R0lGODlhIAAYAPcAMf//////zP//mf//Zv//M///AP/M///MzP/M\\n mf/MZv/MM//MAP+Z//+ZzP+Zmf+ZZv+ZM/+ZAP9m//9mzP9mmf9m\\n Zv9mM/9mAP8z//8zzP8zmf8zZv8zM/8zAP8A//8AzP8Amf8AZv8A\\n M/8AAMz//8z/zMz/mcz/Zsz/M8z/AMzM/8zMzMzMmczMZszMM8zM\\n AMyZ/8yZzMyZmcyZZsyZM8yZAMxm/8xmzMxmmcxmZsxmM8xmAMwz\\n /8wzzMwzmcwzZswzM8wzAMwA/8wAzMwAmcwAZswAM8wAAJn//5n/\\n zJn/mZn/Zpn/M5n/AJnM/5nMzJnMmZnMZpnMM5nMAJmZ/5mZzJmZ\\n mZmZZpmZM5mZAJlm/5lmzJlmmZlmZplmM5lmAJkz/5kzzJkzmZkz\\n ZpkzM5kzAJkA/5kAzJkAmZkAZpkAM5kAAGb//2b/zGb/mWb/Zmb/\\n M2b/AGbM/2bMzGbMmWbMZmbMM2bMAGaZ/2aZzGaZmWaZZmaZM2aZ\\n AGZm/2ZmzGZmmWZmZmZmM2ZmAGYz/2YzzGYzmWYzZmYzM2YzAGYA\\n /2YAzGYAmWYAZmYAM2YAADP//zP/zDP/mTP/ZjP/MzP/ADPM/zPM\\n zDPMmTPMZjPMMzPMADOZ/zOZzDOZmTOZZjOZMzOZADNm/zNmzDNm\\n mTNmZjNmMzNmADMz/zMzzDMzmTMzZjMzMzMzADMA/zMAzDMAmTMA\\n ZjMAMzMAAAD//wD/zAD/mQD/ZgD/MwD/AADM/wDMzADMmQDMZgDM\\n MwDMAACZ/wCZzACZmQCZZgCZMwCZAABm/wBmzABmmQBmZgBmMwBm\\n AAAz/wAzzAAzmQAzZgAzMwAzAAAA/wAAzAAAmQAAZgAAM+4AAN0A\\n ALsAAKoAAIgAAHcAAFUAAEQAACIAABEAAADuAADdAAC7AACqAACI\\n AAB3AABVAABEAAAiAAARAAAA7gAA3QAAuwAAqgAAiAAAdwAAVQAA\\n RAAAIgAAEe7u7t3d3bu7u6qqqoiIiHd3d1VVVURERCIiIhEREQAA\\n ACH5BAEAAAAALAAAAAAgABgAAAhwAAEIHEiwoMGDCBMqXMiwocOH\\n ECNKnEixosWB3zJq3Mixo0eNAL7xG0mypMmTKPl9Cznyn8uWL/m5\\n /AeTpsyYI1eKlBnO5r+eLYHy9Ck0J8ubPmPOrMmUpM6UUKMa/Ui1\\n 6saLWLNq3cq1q9evYB0GBAA7\\n '",
")",
",",
"'NetLink'",
":",
"PhotoImage",
"(",
"data",
"=",
"'\\n R0lGODlhFgAWAPcAMf//////zP//mf//Zv//M///AP/M///MzP/M\\n mf/MZv/MM//MAP+Z//+ZzP+Zmf+ZZv+ZM/+ZAP9m//9mzP9mmf9m\\n Zv9mM/9mAP8z//8zzP8zmf8zZv8zM/8zAP8A//8AzP8Amf8AZv8A\\n M/8AAMz//8z/zMz/mcz/Zsz/M8z/AMzM/8zMzMzMmczMZszMM8zM\\n AMyZ/8yZzMyZmcyZZsyZM8yZAMxm/8xmzMxmmcxmZsxmM8xmAMwz\\n /8wzzMwzmcwzZswzM8wzAMwA/8wAzMwAmcwAZswAM8wAAJn//5n/\\n zJn/mZn/Zpn/M5n/AJnM/5nMzJnMmZnMZpnMM5nMAJmZ/5mZzJmZ\\n mZmZZpmZM5mZAJlm/5lmzJlmmZlmZplmM5lmAJkz/5kzzJkzmZkz\\n ZpkzM5kzAJkA/5kAzJkAmZkAZpkAM5kAAGb//2b/zGb/mWb/Zmb/\\n M2b/AGbM/2bMzGbMmWbMZmbMM2bMAGaZ/2aZzGaZmWaZZmaZM2aZ\\n AGZm/2ZmzGZmmWZmZmZmM2ZmAGYz/2YzzGYzmWYzZmYzM2YzAGYA\\n /2YAzGYAmWYAZmYAM2YAADP//zP/zDP/mTP/ZjP/MzP/ADPM/zPM\\n zDPMmTPMZjPMMzPMADOZ/zOZzDOZmTOZZjOZMzOZADNm/zNmzDNm\\n mTNmZjNmMzNmADMz/zMzzDMzmTMzZjMzMzMzADMA/zMAzDMAmTMA\\n ZjMAMzMAAAD//wD/zAD/mQD/ZgD/MwD/AADM/wDMzADMmQDMZgDM\\n MwDMAACZ/wCZzACZmQCZZgCZMwCZAABm/wBmzABmmQBmZgBmMwBm\\n AAAz/wAzzAAzmQAzZgAzMwAzAAAA/wAAzAAAmQAAZgAAM+4AAN0A\\n ALsAAKoAAIgAAHcAAFUAAEQAACIAABEAAADuAADdAAC7AACqAACI\\n AAB3AABVAABEAAAiAAARAAAA7gAA3QAAuwAAqgAAiAAAdwAAVQAA\\n RAAAIgAAEe7u7t3d3bu7u6qqqoiIiHd3d1VVVURERCIiIhEREQAA\\n ACH5BAEAAAAALAAAAAAWABYAAAhIAAEIHEiwoEGBrhIeXEgwoUKG\\n Cx0+hGhQoiuKBy1irChxY0GNHgeCDAlgZEiTHlFuVImRJUWXEGEy\\n lBmxI8mSNknm1Dnx5sCAADs=\\n '",
")",
"}"
] | create and return images for miniedit . | train | false |
25,224 | def unpack_ext(ext_path):
modcache = os.path.join(OPTIONS.saltdir, 'running_data', 'var', 'cache', 'salt', 'minion', 'extmods')
tfile = tarfile.TarFile.gzopen(ext_path)
old_umask = os.umask(63)
tfile.extractall(path=modcache)
tfile.close()
os.umask(old_umask)
os.unlink(ext_path)
ver_path = os.path.join(modcache, 'ext_version')
ver_dst = os.path.join(OPTIONS.saltdir, 'ext_version')
shutil.move(ver_path, ver_dst)
| [
"def",
"unpack_ext",
"(",
"ext_path",
")",
":",
"modcache",
"=",
"os",
".",
"path",
".",
"join",
"(",
"OPTIONS",
".",
"saltdir",
",",
"'running_data'",
",",
"'var'",
",",
"'cache'",
",",
"'salt'",
",",
"'minion'",
",",
"'extmods'",
")",
"tfile",
"=",
"tarfile",
".",
"TarFile",
".",
"gzopen",
"(",
"ext_path",
")",
"old_umask",
"=",
"os",
".",
"umask",
"(",
"63",
")",
"tfile",
".",
"extractall",
"(",
"path",
"=",
"modcache",
")",
"tfile",
".",
"close",
"(",
")",
"os",
".",
"umask",
"(",
"old_umask",
")",
"os",
".",
"unlink",
"(",
"ext_path",
")",
"ver_path",
"=",
"os",
".",
"path",
".",
"join",
"(",
"modcache",
",",
"'ext_version'",
")",
"ver_dst",
"=",
"os",
".",
"path",
".",
"join",
"(",
"OPTIONS",
".",
"saltdir",
",",
"'ext_version'",
")",
"shutil",
".",
"move",
"(",
"ver_path",
",",
"ver_dst",
")"
] | unpack the external modules . | train | true |
25,225 | def _stringsToObjects(strings, arglist, proto):
objects = {}
myStrings = strings.copy()
for (argname, argparser) in arglist:
argparser.fromBox(argname, myStrings, objects, proto)
return objects
| [
"def",
"_stringsToObjects",
"(",
"strings",
",",
"arglist",
",",
"proto",
")",
":",
"objects",
"=",
"{",
"}",
"myStrings",
"=",
"strings",
".",
"copy",
"(",
")",
"for",
"(",
"argname",
",",
"argparser",
")",
"in",
"arglist",
":",
"argparser",
".",
"fromBox",
"(",
"argname",
",",
"myStrings",
",",
"objects",
",",
"proto",
")",
"return",
"objects"
] | convert an ampbox to a dictionary of python objects . | train | true |
25,226 | def _find_closest_point_on_path(lc, point):
ds = np.sum(((lc - point[None, :]) ** 2), 1)
imin = np.argmin(ds)
dmin = np.inf
xcmin = None
legmin = (None, None)
closed = mlab.is_closed_polygon(lc)
legs = []
if ((imin > 0) or closed):
legs.append((((imin - 1) % len(lc)), imin))
if ((imin < (len(lc) - 1)) or closed):
legs.append((imin, ((imin + 1) % len(lc))))
for leg in legs:
(d, xc) = _find_closest_point_on_leg(lc[leg[0]], lc[leg[1]], point)
if (d < dmin):
dmin = d
xcmin = xc
legmin = leg
return (dmin, xcmin, legmin)
| [
"def",
"_find_closest_point_on_path",
"(",
"lc",
",",
"point",
")",
":",
"ds",
"=",
"np",
".",
"sum",
"(",
"(",
"(",
"lc",
"-",
"point",
"[",
"None",
",",
":",
"]",
")",
"**",
"2",
")",
",",
"1",
")",
"imin",
"=",
"np",
".",
"argmin",
"(",
"ds",
")",
"dmin",
"=",
"np",
".",
"inf",
"xcmin",
"=",
"None",
"legmin",
"=",
"(",
"None",
",",
"None",
")",
"closed",
"=",
"mlab",
".",
"is_closed_polygon",
"(",
"lc",
")",
"legs",
"=",
"[",
"]",
"if",
"(",
"(",
"imin",
">",
"0",
")",
"or",
"closed",
")",
":",
"legs",
".",
"append",
"(",
"(",
"(",
"(",
"imin",
"-",
"1",
")",
"%",
"len",
"(",
"lc",
")",
")",
",",
"imin",
")",
")",
"if",
"(",
"(",
"imin",
"<",
"(",
"len",
"(",
"lc",
")",
"-",
"1",
")",
")",
"or",
"closed",
")",
":",
"legs",
".",
"append",
"(",
"(",
"imin",
",",
"(",
"(",
"imin",
"+",
"1",
")",
"%",
"len",
"(",
"lc",
")",
")",
")",
")",
"for",
"leg",
"in",
"legs",
":",
"(",
"d",
",",
"xc",
")",
"=",
"_find_closest_point_on_leg",
"(",
"lc",
"[",
"leg",
"[",
"0",
"]",
"]",
",",
"lc",
"[",
"leg",
"[",
"1",
"]",
"]",
",",
"point",
")",
"if",
"(",
"d",
"<",
"dmin",
")",
":",
"dmin",
"=",
"d",
"xcmin",
"=",
"xc",
"legmin",
"=",
"leg",
"return",
"(",
"dmin",
",",
"xcmin",
",",
"legmin",
")"
] | lc: coordinates of vertices point: coordinates of test point . | train | false |
25,227 | def flushdb(host=None, port=None, db=None, password=None):
server = _connect(host, port, db, password)
return server.flushdb()
| [
"def",
"flushdb",
"(",
"host",
"=",
"None",
",",
"port",
"=",
"None",
",",
"db",
"=",
"None",
",",
"password",
"=",
"None",
")",
":",
"server",
"=",
"_connect",
"(",
"host",
",",
"port",
",",
"db",
",",
"password",
")",
"return",
"server",
".",
"flushdb",
"(",
")"
] | remove all keys from the selected database cli example: . | train | true |
25,228 | def _addHeaderToRequest(request, header):
requestHeaders = request.requestHeaders
(name, value) = header
values = requestHeaders.getRawHeaders(name)
if (values is not None):
values.append(value)
else:
requestHeaders.setRawHeaders(name, [value])
if (name == 'content-length'):
request.gotLength(int(value))
return True
return False
| [
"def",
"_addHeaderToRequest",
"(",
"request",
",",
"header",
")",
":",
"requestHeaders",
"=",
"request",
".",
"requestHeaders",
"(",
"name",
",",
"value",
")",
"=",
"header",
"values",
"=",
"requestHeaders",
".",
"getRawHeaders",
"(",
"name",
")",
"if",
"(",
"values",
"is",
"not",
"None",
")",
":",
"values",
".",
"append",
"(",
"value",
")",
"else",
":",
"requestHeaders",
".",
"setRawHeaders",
"(",
"name",
",",
"[",
"value",
"]",
")",
"if",
"(",
"name",
"==",
"'content-length'",
")",
":",
"request",
".",
"gotLength",
"(",
"int",
"(",
"value",
")",
")",
"return",
"True",
"return",
"False"
] | add a header tuple to a request header object . | train | false |
25,229 | def project_update(project_id=None, name=None, description=None, enabled=None, profile=None, **connection_args):
auth(profile, **connection_args)
if (_OS_IDENTITY_API_VERSION > 2):
return tenant_update(tenant_id=project_id, name=name, description=description, enabled=enabled, profile=profile, **connection_args)
else:
return False
| [
"def",
"project_update",
"(",
"project_id",
"=",
"None",
",",
"name",
"=",
"None",
",",
"description",
"=",
"None",
",",
"enabled",
"=",
"None",
",",
"profile",
"=",
"None",
",",
"**",
"connection_args",
")",
":",
"auth",
"(",
"profile",
",",
"**",
"connection_args",
")",
"if",
"(",
"_OS_IDENTITY_API_VERSION",
">",
"2",
")",
":",
"return",
"tenant_update",
"(",
"tenant_id",
"=",
"project_id",
",",
"name",
"=",
"name",
",",
"description",
"=",
"description",
",",
"enabled",
"=",
"enabled",
",",
"profile",
"=",
"profile",
",",
"**",
"connection_args",
")",
"else",
":",
"return",
"False"
] | update a tenants information the following fields may be updated: name . | train | true |
25,230 | def v7(method_v7):
frame = currentframe().f_back
return frame.f_locals.get(method_v7.__name__, method_v7)
| [
"def",
"v7",
"(",
"method_v7",
")",
":",
"frame",
"=",
"currentframe",
"(",
")",
".",
"f_back",
"return",
"frame",
".",
"f_locals",
".",
"get",
"(",
"method_v7",
".",
"__name__",
",",
"method_v7",
")"
] | decorate a method that supports the old-style api only . | train | false |
25,232 | def format_jnlp_file_lines(web_flag, url, tep_fp):
lines = [jnlp_top_block]
if web_flag:
lines += ['http://topiaryexplorer.sourceforge.net/app/']
else:
topiaryexplorer_project_dir = load_qiime_config()['topiaryexplorer_project_dir']
if topiaryexplorer_project_dir:
lines += [('file:' + topiaryexplorer_project_dir)]
else:
print "WARNING: Couldn't create jnlp file - topiaryexplorer_project_dir is not defined in your qiime_config. tep file was created sucessfully."
lines += [jnlp_middle_block]
if url:
lines += [url]
else:
lines += [abspath(tep_fp)]
lines += [jnlp_bottom_block]
return lines
| [
"def",
"format_jnlp_file_lines",
"(",
"web_flag",
",",
"url",
",",
"tep_fp",
")",
":",
"lines",
"=",
"[",
"jnlp_top_block",
"]",
"if",
"web_flag",
":",
"lines",
"+=",
"[",
"'http://topiaryexplorer.sourceforge.net/app/'",
"]",
"else",
":",
"topiaryexplorer_project_dir",
"=",
"load_qiime_config",
"(",
")",
"[",
"'topiaryexplorer_project_dir'",
"]",
"if",
"topiaryexplorer_project_dir",
":",
"lines",
"+=",
"[",
"(",
"'file:'",
"+",
"topiaryexplorer_project_dir",
")",
"]",
"else",
":",
"print",
"\"WARNING: Couldn't create jnlp file - topiaryexplorer_project_dir is not defined in your qiime_config. tep file was created sucessfully.\"",
"lines",
"+=",
"[",
"jnlp_middle_block",
"]",
"if",
"url",
":",
"lines",
"+=",
"[",
"url",
"]",
"else",
":",
"lines",
"+=",
"[",
"abspath",
"(",
"tep_fp",
")",
"]",
"lines",
"+=",
"[",
"jnlp_bottom_block",
"]",
"return",
"lines"
] | format the jnlp file for topiaryexplorer . | train | false |
25,233 | @core_helper
def sorted_extras(package_extras, auto_clean=False, subs=None, exclude=None):
if (not exclude):
exclude = config.get('package_hide_extras', [])
output = []
for extra in sorted(package_extras, key=(lambda x: x['key'])):
if (extra.get('state') == 'deleted'):
continue
(k, v) = (extra['key'], extra['value'])
if (k in exclude):
continue
if (subs and (k in subs)):
k = subs[k]
elif auto_clean:
k = k.replace('_', ' ').replace('-', ' ').title()
if isinstance(v, (list, tuple)):
v = ', '.join(map(unicode, v))
output.append((k, v))
return output
| [
"@",
"core_helper",
"def",
"sorted_extras",
"(",
"package_extras",
",",
"auto_clean",
"=",
"False",
",",
"subs",
"=",
"None",
",",
"exclude",
"=",
"None",
")",
":",
"if",
"(",
"not",
"exclude",
")",
":",
"exclude",
"=",
"config",
".",
"get",
"(",
"'package_hide_extras'",
",",
"[",
"]",
")",
"output",
"=",
"[",
"]",
"for",
"extra",
"in",
"sorted",
"(",
"package_extras",
",",
"key",
"=",
"(",
"lambda",
"x",
":",
"x",
"[",
"'key'",
"]",
")",
")",
":",
"if",
"(",
"extra",
".",
"get",
"(",
"'state'",
")",
"==",
"'deleted'",
")",
":",
"continue",
"(",
"k",
",",
"v",
")",
"=",
"(",
"extra",
"[",
"'key'",
"]",
",",
"extra",
"[",
"'value'",
"]",
")",
"if",
"(",
"k",
"in",
"exclude",
")",
":",
"continue",
"if",
"(",
"subs",
"and",
"(",
"k",
"in",
"subs",
")",
")",
":",
"k",
"=",
"subs",
"[",
"k",
"]",
"elif",
"auto_clean",
":",
"k",
"=",
"k",
".",
"replace",
"(",
"'_'",
",",
"' '",
")",
".",
"replace",
"(",
"'-'",
",",
"' '",
")",
".",
"title",
"(",
")",
"if",
"isinstance",
"(",
"v",
",",
"(",
"list",
",",
"tuple",
")",
")",
":",
"v",
"=",
"', '",
".",
"join",
"(",
"map",
"(",
"unicode",
",",
"v",
")",
")",
"output",
".",
"append",
"(",
"(",
"k",
",",
"v",
")",
")",
"return",
"output"
] | used for outputting package extras . | train | false |
25,236 | def getitem(a, b):
return a[b]
| [
"def",
"getitem",
"(",
"a",
",",
"b",
")",
":",
"return",
"a",
"[",
"b",
"]"
] | get item from an iterable at index . | train | false |
25,237 | def normalize_spaces(s):
return ' '.join(s.split())
| [
"def",
"normalize_spaces",
"(",
"s",
")",
":",
"return",
"' '",
".",
"join",
"(",
"s",
".",
"split",
"(",
")",
")"
] | replace any sequence of whitespace characters with a single space . | train | false |
25,238 | def test_solve_rational():
assert (solve(((x - (y ** 3)) / ((y ** 2) * sqrt((1 - (y ** 2))))), x) == [(y ** 3)])
| [
"def",
"test_solve_rational",
"(",
")",
":",
"assert",
"(",
"solve",
"(",
"(",
"(",
"x",
"-",
"(",
"y",
"**",
"3",
")",
")",
"/",
"(",
"(",
"y",
"**",
"2",
")",
"*",
"sqrt",
"(",
"(",
"1",
"-",
"(",
"y",
"**",
"2",
")",
")",
")",
")",
")",
",",
"x",
")",
"==",
"[",
"(",
"y",
"**",
"3",
")",
"]",
")"
] | test solve for rational functions . | train | false |
25,239 | def backup_create(context, values):
return IMPL.backup_create(context, values)
| [
"def",
"backup_create",
"(",
"context",
",",
"values",
")",
":",
"return",
"IMPL",
".",
"backup_create",
"(",
"context",
",",
"values",
")"
] | create a backup from the values dictionary . | train | false |
25,240 | def _delete_asset(course_key, asset_key_string):
if asset_key_string:
try:
asset_key = AssetKey.from_string(asset_key_string)
except InvalidKeyError:
if ('/' == asset_key_string[0]):
asset_key_string = asset_key_string[1:]
try:
asset_key = AssetKey.from_string(asset_key_string)
except InvalidKeyError:
LOGGER.info('In course %r, unable to parse asset key %r, not attempting to delete signatory.', course_key, asset_key_string)
return
else:
LOGGER.info('In course %r, unable to parse asset key %r, not attempting to delete signatory.', course_key, asset_key_string)
return
try:
delete_asset(course_key, asset_key)
except AssetNotFoundException:
pass
| [
"def",
"_delete_asset",
"(",
"course_key",
",",
"asset_key_string",
")",
":",
"if",
"asset_key_string",
":",
"try",
":",
"asset_key",
"=",
"AssetKey",
".",
"from_string",
"(",
"asset_key_string",
")",
"except",
"InvalidKeyError",
":",
"if",
"(",
"'/'",
"==",
"asset_key_string",
"[",
"0",
"]",
")",
":",
"asset_key_string",
"=",
"asset_key_string",
"[",
"1",
":",
"]",
"try",
":",
"asset_key",
"=",
"AssetKey",
".",
"from_string",
"(",
"asset_key_string",
")",
"except",
"InvalidKeyError",
":",
"LOGGER",
".",
"info",
"(",
"'In course %r, unable to parse asset key %r, not attempting to delete signatory.'",
",",
"course_key",
",",
"asset_key_string",
")",
"return",
"else",
":",
"LOGGER",
".",
"info",
"(",
"'In course %r, unable to parse asset key %r, not attempting to delete signatory.'",
",",
"course_key",
",",
"asset_key_string",
")",
"return",
"try",
":",
"delete_asset",
"(",
"course_key",
",",
"asset_key",
")",
"except",
"AssetNotFoundException",
":",
"pass"
] | internal method used to create asset key from string and remove asset by calling delete_asset method of assets module . | train | false |
25,241 | def no_os_popen(logical_line):
if ('os.popen(' in logical_line):
(yield (0, 'N348 Deprecated library function os.popen(). Replace it using subprocess module. '))
| [
"def",
"no_os_popen",
"(",
"logical_line",
")",
":",
"if",
"(",
"'os.popen('",
"in",
"logical_line",
")",
":",
"(",
"yield",
"(",
"0",
",",
"'N348 Deprecated library function os.popen(). Replace it using subprocess module. '",
")",
")"
] | disallow os . | train | false |
25,243 | def instance_metadata_update(context, instance_uuid, metadata, delete):
return IMPL.instance_metadata_update(context, instance_uuid, metadata, delete)
| [
"def",
"instance_metadata_update",
"(",
"context",
",",
"instance_uuid",
",",
"metadata",
",",
"delete",
")",
":",
"return",
"IMPL",
".",
"instance_metadata_update",
"(",
"context",
",",
"instance_uuid",
",",
"metadata",
",",
"delete",
")"
] | update metadata if it exists . | train | false |
25,244 | @utils.arg('fixed_ip', metavar='<fixed_ip>', help=_('Fixed IP Address.'))
@deprecated_network
def do_fixed_ip_reserve(cs, args):
cs.fixed_ips.reserve(args.fixed_ip)
| [
"@",
"utils",
".",
"arg",
"(",
"'fixed_ip'",
",",
"metavar",
"=",
"'<fixed_ip>'",
",",
"help",
"=",
"_",
"(",
"'Fixed IP Address.'",
")",
")",
"@",
"deprecated_network",
"def",
"do_fixed_ip_reserve",
"(",
"cs",
",",
"args",
")",
":",
"cs",
".",
"fixed_ips",
".",
"reserve",
"(",
"args",
".",
"fixed_ip",
")"
] | reserve a fixed ip . | train | false |
25,245 | def getWrappedInteger(integer, modulo):
if (integer >= modulo):
return modulo
if (integer >= 0):
return integer
return (integer % modulo)
| [
"def",
"getWrappedInteger",
"(",
"integer",
",",
"modulo",
")",
":",
"if",
"(",
"integer",
">=",
"modulo",
")",
":",
"return",
"modulo",
"if",
"(",
"integer",
">=",
"0",
")",
":",
"return",
"integer",
"return",
"(",
"integer",
"%",
"modulo",
")"
] | get wrapped integer . | train | false |
25,246 | def Std(xs, mu=None, ddof=0):
var = Var(xs, mu, ddof)
return math.sqrt(var)
| [
"def",
"Std",
"(",
"xs",
",",
"mu",
"=",
"None",
",",
"ddof",
"=",
"0",
")",
":",
"var",
"=",
"Var",
"(",
"xs",
",",
"mu",
",",
"ddof",
")",
"return",
"math",
".",
"sqrt",
"(",
"var",
")"
] | computes standard deviation . | train | false |
25,247 | @hook.command()
def shorten(text):
args = text.split()
url = args[0]
custom = (args[1] if (len(args) > 1) else None)
try:
return web.shorten(url, custom=custom)
except web.ServiceError as e:
return e.message
| [
"@",
"hook",
".",
"command",
"(",
")",
"def",
"shorten",
"(",
"text",
")",
":",
"args",
"=",
"text",
".",
"split",
"(",
")",
"url",
"=",
"args",
"[",
"0",
"]",
"custom",
"=",
"(",
"args",
"[",
"1",
"]",
"if",
"(",
"len",
"(",
"args",
")",
">",
"1",
")",
"else",
"None",
")",
"try",
":",
"return",
"web",
".",
"shorten",
"(",
"url",
",",
"custom",
"=",
"custom",
")",
"except",
"web",
".",
"ServiceError",
"as",
"e",
":",
"return",
"e",
".",
"message"
] | smartly shorten a given string . | train | false |
25,248 | def clientresp(data, **kwargs):
creds = config.get_credentials()
cfg = config.get_config()
dumps_kwargs = {'sort_keys': True, 'cls': utils.PlotlyJSONEncoder}
payload = {'platform': 'python', 'version': version.__version__, 'args': _json.dumps(data, **dumps_kwargs), 'un': creds['username'], 'key': creds['api_key'], 'origin': 'plot', 'kwargs': _json.dumps(kwargs, **dumps_kwargs)}
url = '{plotly_domain}/clientresp'.format(**cfg)
response = request('post', url, data=payload)
parsed_content = response.json()
if parsed_content.get('warning'):
warnings.warn(parsed_content['warning'])
if parsed_content.get('message'):
print parsed_content['message']
return response
| [
"def",
"clientresp",
"(",
"data",
",",
"**",
"kwargs",
")",
":",
"creds",
"=",
"config",
".",
"get_credentials",
"(",
")",
"cfg",
"=",
"config",
".",
"get_config",
"(",
")",
"dumps_kwargs",
"=",
"{",
"'sort_keys'",
":",
"True",
",",
"'cls'",
":",
"utils",
".",
"PlotlyJSONEncoder",
"}",
"payload",
"=",
"{",
"'platform'",
":",
"'python'",
",",
"'version'",
":",
"version",
".",
"__version__",
",",
"'args'",
":",
"_json",
".",
"dumps",
"(",
"data",
",",
"**",
"dumps_kwargs",
")",
",",
"'un'",
":",
"creds",
"[",
"'username'",
"]",
",",
"'key'",
":",
"creds",
"[",
"'api_key'",
"]",
",",
"'origin'",
":",
"'plot'",
",",
"'kwargs'",
":",
"_json",
".",
"dumps",
"(",
"kwargs",
",",
"**",
"dumps_kwargs",
")",
"}",
"url",
"=",
"'{plotly_domain}/clientresp'",
".",
"format",
"(",
"**",
"cfg",
")",
"response",
"=",
"request",
"(",
"'post'",
",",
"url",
",",
"data",
"=",
"payload",
")",
"parsed_content",
"=",
"response",
".",
"json",
"(",
")",
"if",
"parsed_content",
".",
"get",
"(",
"'warning'",
")",
":",
"warnings",
".",
"warn",
"(",
"parsed_content",
"[",
"'warning'",
"]",
")",
"if",
"parsed_content",
".",
"get",
"(",
"'message'",
")",
":",
"print",
"parsed_content",
"[",
"'message'",
"]",
"return",
"response"
] | deprecated endpoint . | train | false |
25,250 | def _check_header_magic_bytes(file_obj):
file_obj.seek(0, 0)
magic = file_obj.read(4)
return (magic == 'PAR1')
| [
"def",
"_check_header_magic_bytes",
"(",
"file_obj",
")",
":",
"file_obj",
".",
"seek",
"(",
"0",
",",
"0",
")",
"magic",
"=",
"file_obj",
".",
"read",
"(",
"4",
")",
"return",
"(",
"magic",
"==",
"'PAR1'",
")"
] | check if the file-like obj has the par1 magic bytes at the header . | train | false |
25,252 | def cliques_containing_node(G, nodes=None, cliques=None):
if (cliques is None):
cliques = list(find_cliques(G))
if (nodes is None):
nodes = list(G.nodes())
if (not isinstance(nodes, list)):
v = nodes
vcliques = [c for c in cliques if (v in c)]
else:
vcliques = {}
for v in nodes:
vcliques[v] = [c for c in cliques if (v in c)]
return vcliques
| [
"def",
"cliques_containing_node",
"(",
"G",
",",
"nodes",
"=",
"None",
",",
"cliques",
"=",
"None",
")",
":",
"if",
"(",
"cliques",
"is",
"None",
")",
":",
"cliques",
"=",
"list",
"(",
"find_cliques",
"(",
"G",
")",
")",
"if",
"(",
"nodes",
"is",
"None",
")",
":",
"nodes",
"=",
"list",
"(",
"G",
".",
"nodes",
"(",
")",
")",
"if",
"(",
"not",
"isinstance",
"(",
"nodes",
",",
"list",
")",
")",
":",
"v",
"=",
"nodes",
"vcliques",
"=",
"[",
"c",
"for",
"c",
"in",
"cliques",
"if",
"(",
"v",
"in",
"c",
")",
"]",
"else",
":",
"vcliques",
"=",
"{",
"}",
"for",
"v",
"in",
"nodes",
":",
"vcliques",
"[",
"v",
"]",
"=",
"[",
"c",
"for",
"c",
"in",
"cliques",
"if",
"(",
"v",
"in",
"c",
")",
"]",
"return",
"vcliques"
] | returns a list of cliques containing the given node . | train | false |
25,253 | def _add_retry_host(filter_properties, host, node):
retry = filter_properties.get('retry', None)
if (not retry):
return
hosts = retry['hosts']
hosts.append([host, node])
| [
"def",
"_add_retry_host",
"(",
"filter_properties",
",",
"host",
",",
"node",
")",
":",
"retry",
"=",
"filter_properties",
".",
"get",
"(",
"'retry'",
",",
"None",
")",
"if",
"(",
"not",
"retry",
")",
":",
"return",
"hosts",
"=",
"retry",
"[",
"'hosts'",
"]",
"hosts",
".",
"append",
"(",
"[",
"host",
",",
"node",
"]",
")"
] | add a retry entry for the selected compute node . | train | false |
25,254 | def mirror_y(clip, apply_to='mask'):
return clip.fl_image((lambda f: f[::(-1)]), apply_to=apply_to)
| [
"def",
"mirror_y",
"(",
"clip",
",",
"apply_to",
"=",
"'mask'",
")",
":",
"return",
"clip",
".",
"fl_image",
"(",
"(",
"lambda",
"f",
":",
"f",
"[",
":",
":",
"(",
"-",
"1",
")",
"]",
")",
",",
"apply_to",
"=",
"apply_to",
")"
] | flips the clip vertically . | train | false |
25,256 | def make_derivative_operator(M, z):
def doit(C):
r = ((z * C.diff(z)) + (C * M))
r = r.applyfunc(make_simp(z))
return r
return doit
| [
"def",
"make_derivative_operator",
"(",
"M",
",",
"z",
")",
":",
"def",
"doit",
"(",
"C",
")",
":",
"r",
"=",
"(",
"(",
"z",
"*",
"C",
".",
"diff",
"(",
"z",
")",
")",
"+",
"(",
"C",
"*",
"M",
")",
")",
"r",
"=",
"r",
".",
"applyfunc",
"(",
"make_simp",
"(",
"z",
")",
")",
"return",
"r",
"return",
"doit"
] | create a derivative operator . | train | false |
25,257 | def hmc_move(s_rng, positions, energy_fn, stepsize, n_steps):
initial_vel = s_rng.normal(size=positions.shape)
(final_pos, final_vel) = simulate_dynamics(initial_pos=positions, initial_vel=initial_vel, stepsize=stepsize, n_steps=n_steps, energy_fn=energy_fn)
accept = metropolis_hastings_accept(energy_prev=hamiltonian(positions, initial_vel, energy_fn), energy_next=hamiltonian(final_pos, final_vel, energy_fn), s_rng=s_rng)
return (accept, final_pos)
| [
"def",
"hmc_move",
"(",
"s_rng",
",",
"positions",
",",
"energy_fn",
",",
"stepsize",
",",
"n_steps",
")",
":",
"initial_vel",
"=",
"s_rng",
".",
"normal",
"(",
"size",
"=",
"positions",
".",
"shape",
")",
"(",
"final_pos",
",",
"final_vel",
")",
"=",
"simulate_dynamics",
"(",
"initial_pos",
"=",
"positions",
",",
"initial_vel",
"=",
"initial_vel",
",",
"stepsize",
"=",
"stepsize",
",",
"n_steps",
"=",
"n_steps",
",",
"energy_fn",
"=",
"energy_fn",
")",
"accept",
"=",
"metropolis_hastings_accept",
"(",
"energy_prev",
"=",
"hamiltonian",
"(",
"positions",
",",
"initial_vel",
",",
"energy_fn",
")",
",",
"energy_next",
"=",
"hamiltonian",
"(",
"final_pos",
",",
"final_vel",
",",
"energy_fn",
")",
",",
"s_rng",
"=",
"s_rng",
")",
"return",
"(",
"accept",
",",
"final_pos",
")"
] | this function performs one-step of hybrid monte-carlo sampling . | train | false |
25,258 | def _zpklp2lp(z, p, k, wo=1):
degree = _relative_degree(z, p)
z_lp = [(wo * z1) for z1 in z]
p_lp = [(wo * p1) for p1 in p]
k_lp = (k * (wo ** degree))
return (z_lp, p_lp, k_lp)
| [
"def",
"_zpklp2lp",
"(",
"z",
",",
"p",
",",
"k",
",",
"wo",
"=",
"1",
")",
":",
"degree",
"=",
"_relative_degree",
"(",
"z",
",",
"p",
")",
"z_lp",
"=",
"[",
"(",
"wo",
"*",
"z1",
")",
"for",
"z1",
"in",
"z",
"]",
"p_lp",
"=",
"[",
"(",
"wo",
"*",
"p1",
")",
"for",
"p1",
"in",
"p",
"]",
"k_lp",
"=",
"(",
"k",
"*",
"(",
"wo",
"**",
"degree",
")",
")",
"return",
"(",
"z_lp",
",",
"p_lp",
",",
"k_lp",
")"
] | transform a lowpass filter prototype to a different frequency . | train | false |
25,260 | def __get_tags(vm_):
t = config.get_cloud_config_value('tags', vm_, __opts__, default='[]', search_global=False)
try:
tags = literal_eval(t)
except Exception:
tags = None
if ((not tags) or (not isinstance(tags, list))):
tags = None
return tags
| [
"def",
"__get_tags",
"(",
"vm_",
")",
":",
"t",
"=",
"config",
".",
"get_cloud_config_value",
"(",
"'tags'",
",",
"vm_",
",",
"__opts__",
",",
"default",
"=",
"'[]'",
",",
"search_global",
"=",
"False",
")",
"try",
":",
"tags",
"=",
"literal_eval",
"(",
"t",
")",
"except",
"Exception",
":",
"tags",
"=",
"None",
"if",
"(",
"(",
"not",
"tags",
")",
"or",
"(",
"not",
"isinstance",
"(",
"tags",
",",
"list",
")",
")",
")",
":",
"tags",
"=",
"None",
"return",
"tags"
] | get configured tags . | train | true |
25,261 | def reset():
_runtime.reset()
| [
"def",
"reset",
"(",
")",
":",
"_runtime",
".",
"reset",
"(",
")"
] | interface to git-reset(1)_ . | train | false |
25,262 | def test_allknn_sk_estimator():
check_estimator(AllKNN)
| [
"def",
"test_allknn_sk_estimator",
"(",
")",
":",
"check_estimator",
"(",
"AllKNN",
")"
] | test the sklearn estimator compatibility . | train | false |
25,263 | def compareSegmentLength(endpoint, otherEndpoint):
if (endpoint.segmentLength > otherEndpoint.segmentLength):
return 1
if (endpoint.segmentLength < otherEndpoint.segmentLength):
return (-1)
return 0
| [
"def",
"compareSegmentLength",
"(",
"endpoint",
",",
"otherEndpoint",
")",
":",
"if",
"(",
"endpoint",
".",
"segmentLength",
">",
"otherEndpoint",
".",
"segmentLength",
")",
":",
"return",
"1",
"if",
"(",
"endpoint",
".",
"segmentLength",
"<",
"otherEndpoint",
".",
"segmentLength",
")",
":",
"return",
"(",
"-",
"1",
")",
"return",
"0"
] | get comparison in order to sort endpoints in ascending order of segment length . | train | false |
25,265 | def _stdin_ready_other():
return True
| [
"def",
"_stdin_ready_other",
"(",
")",
":",
"return",
"True"
] | return true . | train | false |
25,267 | def get_bind_data(zone_id, profile):
conn = _get_driver(profile=profile)
zone = conn.get_zone(zone_id)
return conn.export_zone_to_bind_format(zone)
| [
"def",
"get_bind_data",
"(",
"zone_id",
",",
"profile",
")",
":",
"conn",
"=",
"_get_driver",
"(",
"profile",
"=",
"profile",
")",
"zone",
"=",
"conn",
".",
"get_zone",
"(",
"zone_id",
")",
"return",
"conn",
".",
"export_zone_to_bind_format",
"(",
"zone",
")"
] | export zone to the bind compatible format . | train | true |
25,269 | def manage_entries(options, selection, approved):
approved_text = (u'approved' if approved else u'pending')
with Session() as session:
if (selection == u'all'):
entries = list_pending_entries(session=session, approved=(not approved))
else:
try:
entry = get_entry_by_id(session, selection)
if (entry.approved is approved):
console((colorize(u'red', u'ERROR: ') + (u'Entry with ID %s is already %s' % (entry.id, approved_text))))
sys.exit(1)
except NoResultFound:
console((u'Pending entry with ID %s does not exist' % selection))
sys.exit(1)
else:
entries = [entry]
if (not entries):
console((u'All entries are already %s' % approved_text))
return
for entry in entries:
if (entry.approved is not approved):
console((u'Setting pending entry with ID %s status to %s' % (entry.id, approved_text)))
entry.approved = approved
| [
"def",
"manage_entries",
"(",
"options",
",",
"selection",
",",
"approved",
")",
":",
"approved_text",
"=",
"(",
"u'approved'",
"if",
"approved",
"else",
"u'pending'",
")",
"with",
"Session",
"(",
")",
"as",
"session",
":",
"if",
"(",
"selection",
"==",
"u'all'",
")",
":",
"entries",
"=",
"list_pending_entries",
"(",
"session",
"=",
"session",
",",
"approved",
"=",
"(",
"not",
"approved",
")",
")",
"else",
":",
"try",
":",
"entry",
"=",
"get_entry_by_id",
"(",
"session",
",",
"selection",
")",
"if",
"(",
"entry",
".",
"approved",
"is",
"approved",
")",
":",
"console",
"(",
"(",
"colorize",
"(",
"u'red'",
",",
"u'ERROR: '",
")",
"+",
"(",
"u'Entry with ID %s is already %s'",
"%",
"(",
"entry",
".",
"id",
",",
"approved_text",
")",
")",
")",
")",
"sys",
".",
"exit",
"(",
"1",
")",
"except",
"NoResultFound",
":",
"console",
"(",
"(",
"u'Pending entry with ID %s does not exist'",
"%",
"selection",
")",
")",
"sys",
".",
"exit",
"(",
"1",
")",
"else",
":",
"entries",
"=",
"[",
"entry",
"]",
"if",
"(",
"not",
"entries",
")",
":",
"console",
"(",
"(",
"u'All entries are already %s'",
"%",
"approved_text",
")",
")",
"return",
"for",
"entry",
"in",
"entries",
":",
"if",
"(",
"entry",
".",
"approved",
"is",
"not",
"approved",
")",
":",
"console",
"(",
"(",
"u'Setting pending entry with ID %s status to %s'",
"%",
"(",
"entry",
".",
"id",
",",
"approved_text",
")",
")",
")",
"entry",
".",
"approved",
"=",
"approved"
] | manage pending entries . | train | false |
25,270 | def inthread(func):
@wraps(func)
def wrapped(*a, **kw):
return threads.deferToThread(func, *a, **kw)
return wrapped
| [
"def",
"inthread",
"(",
"func",
")",
":",
"@",
"wraps",
"(",
"func",
")",
"def",
"wrapped",
"(",
"*",
"a",
",",
"**",
"kw",
")",
":",
"return",
"threads",
".",
"deferToThread",
"(",
"func",
",",
"*",
"a",
",",
"**",
"kw",
")",
"return",
"wrapped"
] | decorator to call a function in a thread and return a deferred with the result . | train | false |
25,272 | def validateMasterOption(master):
try:
(hostname, port) = master.split(':')
port = int(port)
except (TypeError, ValueError):
raise usage.UsageError("master must have the form 'hostname:port'")
| [
"def",
"validateMasterOption",
"(",
"master",
")",
":",
"try",
":",
"(",
"hostname",
",",
"port",
")",
"=",
"master",
".",
"split",
"(",
"':'",
")",
"port",
"=",
"int",
"(",
"port",
")",
"except",
"(",
"TypeError",
",",
"ValueError",
")",
":",
"raise",
"usage",
".",
"UsageError",
"(",
"\"master must have the form 'hostname:port'\"",
")"
] | validate master command line option . | train | true |
25,273 | def is_cnf(expr):
return _is_form(expr, And, Or)
| [
"def",
"is_cnf",
"(",
"expr",
")",
":",
"return",
"_is_form",
"(",
"expr",
",",
"And",
",",
"Or",
")"
] | test whether or not an expression is in conjunctive normal form . | train | false |
25,275 | def cachedir_index_del(minion_id, base=None):
base = init_cachedir(base)
index_file = os.path.join(base, 'index.p')
lock_file(index_file)
if os.path.exists(index_file):
with salt.utils.fopen(index_file, 'r') as fh_:
index = msgpack.load(fh_)
else:
return
if (minion_id in index):
del index[minion_id]
with salt.utils.fopen(index_file, 'w') as fh_:
msgpack.dump(index, fh_)
unlock_file(index_file)
| [
"def",
"cachedir_index_del",
"(",
"minion_id",
",",
"base",
"=",
"None",
")",
":",
"base",
"=",
"init_cachedir",
"(",
"base",
")",
"index_file",
"=",
"os",
".",
"path",
".",
"join",
"(",
"base",
",",
"'index.p'",
")",
"lock_file",
"(",
"index_file",
")",
"if",
"os",
".",
"path",
".",
"exists",
"(",
"index_file",
")",
":",
"with",
"salt",
".",
"utils",
".",
"fopen",
"(",
"index_file",
",",
"'r'",
")",
"as",
"fh_",
":",
"index",
"=",
"msgpack",
".",
"load",
"(",
"fh_",
")",
"else",
":",
"return",
"if",
"(",
"minion_id",
"in",
"index",
")",
":",
"del",
"index",
"[",
"minion_id",
"]",
"with",
"salt",
".",
"utils",
".",
"fopen",
"(",
"index_file",
",",
"'w'",
")",
"as",
"fh_",
":",
"msgpack",
".",
"dump",
"(",
"index",
",",
"fh_",
")",
"unlock_file",
"(",
"index_file",
")"
] | delete an entry from the cachedir index . | train | false |
25,276 | def config(conf_path='/etc/rsyncd.conf'):
ret = ''
try:
with salt.utils.fopen(conf_path, 'r') as fp_:
for line in fp_:
ret += line
except IOError as exc:
if (exc.errno == errno.ENOENT):
raise CommandExecutionError('{0} does not exist'.format(conf_path))
elif (exc.errno == errno.EACCES):
raise CommandExecutionError('Unable to read {0}, access denied'.format(conf_path))
elif (exc.errno == errno.EISDIR):
raise CommandExecutionError('Unable to read {0}, path is a directory'.format(conf_path))
else:
raise CommandExecutionError('Error {0}: {1}'.format(exc.errno, exc.strerror))
else:
return ret
| [
"def",
"config",
"(",
"conf_path",
"=",
"'/etc/rsyncd.conf'",
")",
":",
"ret",
"=",
"''",
"try",
":",
"with",
"salt",
".",
"utils",
".",
"fopen",
"(",
"conf_path",
",",
"'r'",
")",
"as",
"fp_",
":",
"for",
"line",
"in",
"fp_",
":",
"ret",
"+=",
"line",
"except",
"IOError",
"as",
"exc",
":",
"if",
"(",
"exc",
".",
"errno",
"==",
"errno",
".",
"ENOENT",
")",
":",
"raise",
"CommandExecutionError",
"(",
"'{0} does not exist'",
".",
"format",
"(",
"conf_path",
")",
")",
"elif",
"(",
"exc",
".",
"errno",
"==",
"errno",
".",
"EACCES",
")",
":",
"raise",
"CommandExecutionError",
"(",
"'Unable to read {0}, access denied'",
".",
"format",
"(",
"conf_path",
")",
")",
"elif",
"(",
"exc",
".",
"errno",
"==",
"errno",
".",
"EISDIR",
")",
":",
"raise",
"CommandExecutionError",
"(",
"'Unable to read {0}, path is a directory'",
".",
"format",
"(",
"conf_path",
")",
")",
"else",
":",
"raise",
"CommandExecutionError",
"(",
"'Error {0}: {1}'",
".",
"format",
"(",
"exc",
".",
"errno",
",",
"exc",
".",
"strerror",
")",
")",
"else",
":",
"return",
"ret"
] | set traffic server configuration variable values . | train | true |
25,277 | def _createPluginDummy(entrypath, pluginContent, real, pluginModule):
entrypath.createDirectory()
pkg = entrypath.child('plugindummy')
pkg.createDirectory()
if real:
pkg.child('__init__.py').setContent('')
plugs = pkg.child('plugins')
plugs.createDirectory()
if real:
plugs.child('__init__.py').setContent(pluginInitFile)
plugs.child((pluginModule + '.py')).setContent(pluginContent)
return plugs
| [
"def",
"_createPluginDummy",
"(",
"entrypath",
",",
"pluginContent",
",",
"real",
",",
"pluginModule",
")",
":",
"entrypath",
".",
"createDirectory",
"(",
")",
"pkg",
"=",
"entrypath",
".",
"child",
"(",
"'plugindummy'",
")",
"pkg",
".",
"createDirectory",
"(",
")",
"if",
"real",
":",
"pkg",
".",
"child",
"(",
"'__init__.py'",
")",
".",
"setContent",
"(",
"''",
")",
"plugs",
"=",
"pkg",
".",
"child",
"(",
"'plugins'",
")",
"plugs",
".",
"createDirectory",
"(",
")",
"if",
"real",
":",
"plugs",
".",
"child",
"(",
"'__init__.py'",
")",
".",
"setContent",
"(",
"pluginInitFile",
")",
"plugs",
".",
"child",
"(",
"(",
"pluginModule",
"+",
"'.py'",
")",
")",
".",
"setContent",
"(",
"pluginContent",
")",
"return",
"plugs"
] | create a plugindummy package . | train | false |
25,278 | def one(method):
def loop(method, self, *args, **kwargs):
result = [method(rec, *args, **kwargs) for rec in self]
return aggregate(method, result, self)
wrapper = decorator(loop, method)
wrapper._api = 'one'
return wrapper
| [
"def",
"one",
"(",
"method",
")",
":",
"def",
"loop",
"(",
"method",
",",
"self",
",",
"*",
"args",
",",
"**",
"kwargs",
")",
":",
"result",
"=",
"[",
"method",
"(",
"rec",
",",
"*",
"args",
",",
"**",
"kwargs",
")",
"for",
"rec",
"in",
"self",
"]",
"return",
"aggregate",
"(",
"method",
",",
"result",
",",
"self",
")",
"wrapper",
"=",
"decorator",
"(",
"loop",
",",
"method",
")",
"wrapper",
".",
"_api",
"=",
"'one'",
"return",
"wrapper"
] | one returns a single game matching the given criteria . | train | false |
25,279 | def contains_exact_files(directory, expected_files, ignore_links=False):
return (sorted(expected_files) == sorted(exact_files(directory, ignore_links=ignore_links)))
| [
"def",
"contains_exact_files",
"(",
"directory",
",",
"expected_files",
",",
"ignore_links",
"=",
"False",
")",
":",
"return",
"(",
"sorted",
"(",
"expected_files",
")",
"==",
"sorted",
"(",
"exact_files",
"(",
"directory",
",",
"ignore_links",
"=",
"ignore_links",
")",
")",
")"
] | check if the only files which directory contains are expected_files . | train | false |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.