id_within_dataset
int64
1
55.5k
snippet
stringlengths
19
14.2k
tokens
listlengths
6
1.63k
nl
stringlengths
6
352
split_within_dataset
stringclasses
1 value
is_duplicated
bool
2 classes
25,142
def on_valid(valid_content_type, on_invalid=json): invalid_kwargs = introspect.generate_accepted_kwargs(on_invalid, 'request', 'response') invalid_takes_response = introspect.takes_all_arguments(on_invalid, 'response') def wrapper(function): valid_kwargs = introspect.generate_accepted_kwargs(function, 'request', 'response') valid_takes_response = introspect.takes_all_arguments(function, 'response') @content_type(valid_content_type) @wraps(function) def output_content(content, response, **kwargs): if ((type(content) == dict) and ('errors' in content)): response.content_type = on_invalid.content_type if invalid_takes_response: kwargs['response'] = response return on_invalid(content, **invalid_kwargs(kwargs)) if valid_takes_response: kwargs['response'] = response return function(content, **valid_kwargs(kwargs)) return output_content return wrapper
[ "def", "on_valid", "(", "valid_content_type", ",", "on_invalid", "=", "json", ")", ":", "invalid_kwargs", "=", "introspect", ".", "generate_accepted_kwargs", "(", "on_invalid", ",", "'request'", ",", "'response'", ")", "invalid_takes_response", "=", "introspect", ".", "takes_all_arguments", "(", "on_invalid", ",", "'response'", ")", "def", "wrapper", "(", "function", ")", ":", "valid_kwargs", "=", "introspect", ".", "generate_accepted_kwargs", "(", "function", ",", "'request'", ",", "'response'", ")", "valid_takes_response", "=", "introspect", ".", "takes_all_arguments", "(", "function", ",", "'response'", ")", "@", "content_type", "(", "valid_content_type", ")", "@", "wraps", "(", "function", ")", "def", "output_content", "(", "content", ",", "response", ",", "**", "kwargs", ")", ":", "if", "(", "(", "type", "(", "content", ")", "==", "dict", ")", "and", "(", "'errors'", "in", "content", ")", ")", ":", "response", ".", "content_type", "=", "on_invalid", ".", "content_type", "if", "invalid_takes_response", ":", "kwargs", "[", "'response'", "]", "=", "response", "return", "on_invalid", "(", "content", ",", "**", "invalid_kwargs", "(", "kwargs", ")", ")", "if", "valid_takes_response", ":", "kwargs", "[", "'response'", "]", "=", "response", "return", "function", "(", "content", ",", "**", "valid_kwargs", "(", "kwargs", ")", ")", "return", "output_content", "return", "wrapper" ]
renders as the specified content type only if no errors are found in the provided data object .
train
true
25,143
def getManipulatedPaths(close, elementNode, loop, prefix, sideLength): if (len(loop) < 1): return [[]] derivation = BottomDerivation(elementNode, prefix) targetMatrix = matrix.getBranchMatrixSetElementNode(elementNode) transformedLoop = matrix.getTransformedVector3s(matrix.getIdentityTetragrid(targetMatrix.tetragrid), loop) lift = ((derivation.altitude + derivation.getAdditionalPathLift()) - euclidean.getBottomByPath(transformedLoop)) for point in loop: point.z += lift return [loop]
[ "def", "getManipulatedPaths", "(", "close", ",", "elementNode", ",", "loop", ",", "prefix", ",", "sideLength", ")", ":", "if", "(", "len", "(", "loop", ")", "<", "1", ")", ":", "return", "[", "[", "]", "]", "derivation", "=", "BottomDerivation", "(", "elementNode", ",", "prefix", ")", "targetMatrix", "=", "matrix", ".", "getBranchMatrixSetElementNode", "(", "elementNode", ")", "transformedLoop", "=", "matrix", ".", "getTransformedVector3s", "(", "matrix", ".", "getIdentityTetragrid", "(", "targetMatrix", ".", "tetragrid", ")", ",", "loop", ")", "lift", "=", "(", "(", "derivation", ".", "altitude", "+", "derivation", ".", "getAdditionalPathLift", "(", ")", ")", "-", "euclidean", ".", "getBottomByPath", "(", "transformedLoop", ")", ")", "for", "point", "in", "loop", ":", "point", ".", "z", "+=", "lift", "return", "[", "loop", "]" ]
get round loop .
train
false
25,144
def create_comm_note(obj, version, author, body, note_type=comm.NO_ACTION, perms=None, attachments=None): perms = (perms or {}) if (('developer' not in perms) and (note_type in comm.REVIEWER_NOTE_TYPES)): perms['developer'] = False create_perms = dict(((('read_permission_%s' % key), has_perm) for (key, has_perm) in perms.iteritems())) version_param = {} if (obj.__class__ == Webapp): version_param['_version'] = version elif (obj.__class__ == Extension): version_param['_extension_version'] = version (thread, created_thread) = obj.threads.safer_get_or_create(defaults=create_perms, **version_param) note = thread.notes.create(note_type=note_type, body=body, author=author, **create_perms) if attachments: create_attachments(note, attachments) post_create_comm_note(note) return (thread, note)
[ "def", "create_comm_note", "(", "obj", ",", "version", ",", "author", ",", "body", ",", "note_type", "=", "comm", ".", "NO_ACTION", ",", "perms", "=", "None", ",", "attachments", "=", "None", ")", ":", "perms", "=", "(", "perms", "or", "{", "}", ")", "if", "(", "(", "'developer'", "not", "in", "perms", ")", "and", "(", "note_type", "in", "comm", ".", "REVIEWER_NOTE_TYPES", ")", ")", ":", "perms", "[", "'developer'", "]", "=", "False", "create_perms", "=", "dict", "(", "(", "(", "(", "'read_permission_%s'", "%", "key", ")", ",", "has_perm", ")", "for", "(", "key", ",", "has_perm", ")", "in", "perms", ".", "iteritems", "(", ")", ")", ")", "version_param", "=", "{", "}", "if", "(", "obj", ".", "__class__", "==", "Webapp", ")", ":", "version_param", "[", "'_version'", "]", "=", "version", "elif", "(", "obj", ".", "__class__", "==", "Extension", ")", ":", "version_param", "[", "'_extension_version'", "]", "=", "version", "(", "thread", ",", "created_thread", ")", "=", "obj", ".", "threads", ".", "safer_get_or_create", "(", "defaults", "=", "create_perms", ",", "**", "version_param", ")", "note", "=", "thread", ".", "notes", ".", "create", "(", "note_type", "=", "note_type", ",", "body", "=", "body", ",", "author", "=", "author", ",", "**", "create_perms", ")", "if", "attachments", ":", "create_attachments", "(", "note", ",", "attachments", ")", "post_create_comm_note", "(", "note", ")", "return", "(", "thread", ",", "note", ")" ]
creates a note on an obj versions thread .
train
false
25,145
def parse_partial_identifier(word): p = sqlparse.parse(word)[0] n_tok = len(p.tokens) if ((n_tok == 1) and isinstance(p.tokens[0], Identifier)): return p.tokens[0] elif p.token_next_by(m=(Error, '"'))[1]: return parse_partial_identifier((word + '"')) else: return None
[ "def", "parse_partial_identifier", "(", "word", ")", ":", "p", "=", "sqlparse", ".", "parse", "(", "word", ")", "[", "0", "]", "n_tok", "=", "len", "(", "p", ".", "tokens", ")", "if", "(", "(", "n_tok", "==", "1", ")", "and", "isinstance", "(", "p", ".", "tokens", "[", "0", "]", ",", "Identifier", ")", ")", ":", "return", "p", ".", "tokens", "[", "0", "]", "elif", "p", ".", "token_next_by", "(", "m", "=", "(", "Error", ",", "'\"'", ")", ")", "[", "1", "]", ":", "return", "parse_partial_identifier", "(", "(", "word", "+", "'\"'", ")", ")", "else", ":", "return", "None" ]
attempt to parse a word as an identifier word may include a schema qualification .
train
false
25,146
def stub_out_image_service(test): image_service = FakeImageService() test.stub_out('nova.image.glance.get_remote_image_service', (lambda x, y: (image_service, y))) test.stub_out('nova.image.glance.get_default_image_service', (lambda : image_service)) test.useFixture(nova_fixtures.ConfPatcher(group='glance', api_servers=['http://localhost:9292'])) return image_service
[ "def", "stub_out_image_service", "(", "test", ")", ":", "image_service", "=", "FakeImageService", "(", ")", "test", ".", "stub_out", "(", "'nova.image.glance.get_remote_image_service'", ",", "(", "lambda", "x", ",", "y", ":", "(", "image_service", ",", "y", ")", ")", ")", "test", ".", "stub_out", "(", "'nova.image.glance.get_default_image_service'", ",", "(", "lambda", ":", "image_service", ")", ")", "test", ".", "useFixture", "(", "nova_fixtures", ".", "ConfPatcher", "(", "group", "=", "'glance'", ",", "api_servers", "=", "[", "'http://localhost:9292'", "]", ")", ")", "return", "image_service" ]
stubs out the image service for the test with the fakeimageservice .
train
false
25,148
def reorder_title(title, articles=('the',), separators=(',', ', ')): ltitle = title.lower() for article in articles: for separator in separators: suffix = (separator + article) if (ltitle[(- len(suffix)):] == suffix): return ((title[((- len(suffix)) + len(separator)):] + ' ') + title[:(- len(suffix))]) return title
[ "def", "reorder_title", "(", "title", ",", "articles", "=", "(", "'the'", ",", ")", ",", "separators", "=", "(", "','", ",", "', '", ")", ")", ":", "ltitle", "=", "title", ".", "lower", "(", ")", "for", "article", "in", "articles", ":", "for", "separator", "in", "separators", ":", "suffix", "=", "(", "separator", "+", "article", ")", "if", "(", "ltitle", "[", "(", "-", "len", "(", "suffix", ")", ")", ":", "]", "==", "suffix", ")", ":", "return", "(", "(", "title", "[", "(", "(", "-", "len", "(", "suffix", ")", ")", "+", "len", "(", "separator", ")", ")", ":", "]", "+", "' '", ")", "+", "title", "[", ":", "(", "-", "len", "(", "suffix", ")", ")", "]", ")", "return", "title" ]
reorder the title .
train
false
25,149
def get_descrs(arrays, col_name_map): out_descrs = [] for (out_name, in_names) in six.iteritems(col_name_map): in_cols = [arr[name] for (arr, name) in zip(arrays, in_names) if (name is not None)] names = [name for name in in_names if (name is not None)] try: dtype = common_dtype(in_cols) except TableMergeError as tme: raise TableMergeError(u"The '{0}' columns have incompatible types: {1}".format(names[0], tme._incompat_types)) uniq_shapes = set((col.shape[1:] for col in in_cols)) if (len(uniq_shapes) != 1): raise TableMergeError(u'Key columns {0!r} have different shape'.format(name)) shape = uniq_shapes.pop() out_descrs.append((fix_column_name(out_name), dtype, shape)) return out_descrs
[ "def", "get_descrs", "(", "arrays", ",", "col_name_map", ")", ":", "out_descrs", "=", "[", "]", "for", "(", "out_name", ",", "in_names", ")", "in", "six", ".", "iteritems", "(", "col_name_map", ")", ":", "in_cols", "=", "[", "arr", "[", "name", "]", "for", "(", "arr", ",", "name", ")", "in", "zip", "(", "arrays", ",", "in_names", ")", "if", "(", "name", "is", "not", "None", ")", "]", "names", "=", "[", "name", "for", "name", "in", "in_names", "if", "(", "name", "is", "not", "None", ")", "]", "try", ":", "dtype", "=", "common_dtype", "(", "in_cols", ")", "except", "TableMergeError", "as", "tme", ":", "raise", "TableMergeError", "(", "u\"The '{0}' columns have incompatible types: {1}\"", ".", "format", "(", "names", "[", "0", "]", ",", "tme", ".", "_incompat_types", ")", ")", "uniq_shapes", "=", "set", "(", "(", "col", ".", "shape", "[", "1", ":", "]", "for", "col", "in", "in_cols", ")", ")", "if", "(", "len", "(", "uniq_shapes", ")", "!=", "1", ")", ":", "raise", "TableMergeError", "(", "u'Key columns {0!r} have different shape'", ".", "format", "(", "name", ")", ")", "shape", "=", "uniq_shapes", ".", "pop", "(", ")", "out_descrs", ".", "append", "(", "(", "fix_column_name", "(", "out_name", ")", ",", "dtype", ",", "shape", ")", ")", "return", "out_descrs" ]
find the dtypes descrs resulting from merging the list of arrays dtypes .
train
false
25,150
def generate_config(hass, passwd): from homeassistant.components.mqtt import PROTOCOL_311 config = {'listeners': {'default': {'max-connections': 50000, 'bind': '0.0.0.0:1883', 'type': 'tcp'}, 'ws-1': {'bind': '0.0.0.0:8080', 'type': 'ws'}}, 'auth': {'allow-anonymous': (hass.config.api.api_password is None)}, 'plugins': ['auth_anonymous']} if hass.config.api.api_password: username = 'homeassistant' password = hass.config.api.api_password from passlib.apps import custom_app_context passwd.write('homeassistant:{}\n'.format(custom_app_context.encrypt(hass.config.api.api_password)).encode('utf-8')) passwd.flush() config['auth']['password-file'] = passwd.name config['plugins'].append('auth_file') else: username = None password = None client_config = ('localhost', 1883, username, password, None, PROTOCOL_311) return (config, client_config)
[ "def", "generate_config", "(", "hass", ",", "passwd", ")", ":", "from", "homeassistant", ".", "components", ".", "mqtt", "import", "PROTOCOL_311", "config", "=", "{", "'listeners'", ":", "{", "'default'", ":", "{", "'max-connections'", ":", "50000", ",", "'bind'", ":", "'0.0.0.0:1883'", ",", "'type'", ":", "'tcp'", "}", ",", "'ws-1'", ":", "{", "'bind'", ":", "'0.0.0.0:8080'", ",", "'type'", ":", "'ws'", "}", "}", ",", "'auth'", ":", "{", "'allow-anonymous'", ":", "(", "hass", ".", "config", ".", "api", ".", "api_password", "is", "None", ")", "}", ",", "'plugins'", ":", "[", "'auth_anonymous'", "]", "}", "if", "hass", ".", "config", ".", "api", ".", "api_password", ":", "username", "=", "'homeassistant'", "password", "=", "hass", ".", "config", ".", "api", ".", "api_password", "from", "passlib", ".", "apps", "import", "custom_app_context", "passwd", ".", "write", "(", "'homeassistant:{}\\n'", ".", "format", "(", "custom_app_context", ".", "encrypt", "(", "hass", ".", "config", ".", "api", ".", "api_password", ")", ")", ".", "encode", "(", "'utf-8'", ")", ")", "passwd", ".", "flush", "(", ")", "config", "[", "'auth'", "]", "[", "'password-file'", "]", "=", "passwd", ".", "name", "config", "[", "'plugins'", "]", ".", "append", "(", "'auth_file'", ")", "else", ":", "username", "=", "None", "password", "=", "None", "client_config", "=", "(", "'localhost'", ",", "1883", ",", "username", ",", "password", ",", "None", ",", "PROTOCOL_311", ")", "return", "(", "config", ",", "client_config", ")" ]
generate a configuration based on current home assistant instance .
train
false
25,151
def upload_imageattachment(request, obj): return upload_media(request, ImageAttachmentUploadForm, create_imageattachment, obj=obj)
[ "def", "upload_imageattachment", "(", "request", ",", "obj", ")", ":", "return", "upload_media", "(", "request", ",", "ImageAttachmentUploadForm", ",", "create_imageattachment", ",", "obj", "=", "obj", ")" ]
uploads image attachments .
train
false
25,152
def simple_rnn(incoming, n_units, activation='sigmoid', dropout=None, bias=True, weights_init=None, return_seq=False, return_state=False, initial_state=None, dynamic=False, trainable=True, restore=True, reuse=False, scope=None, name='SimpleRNN'): cell = BasicRNNCell(n_units, activation=activation, bias=bias, weights_init=weights_init, trainable=trainable, restore=restore, reuse=reuse) x = _rnn_template(incoming, cell=cell, dropout=dropout, return_seq=return_seq, return_state=return_state, initial_state=initial_state, dynamic=dynamic, scope=scope, name=name) return x
[ "def", "simple_rnn", "(", "incoming", ",", "n_units", ",", "activation", "=", "'sigmoid'", ",", "dropout", "=", "None", ",", "bias", "=", "True", ",", "weights_init", "=", "None", ",", "return_seq", "=", "False", ",", "return_state", "=", "False", ",", "initial_state", "=", "None", ",", "dynamic", "=", "False", ",", "trainable", "=", "True", ",", "restore", "=", "True", ",", "reuse", "=", "False", ",", "scope", "=", "None", ",", "name", "=", "'SimpleRNN'", ")", ":", "cell", "=", "BasicRNNCell", "(", "n_units", ",", "activation", "=", "activation", ",", "bias", "=", "bias", ",", "weights_init", "=", "weights_init", ",", "trainable", "=", "trainable", ",", "restore", "=", "restore", ",", "reuse", "=", "reuse", ")", "x", "=", "_rnn_template", "(", "incoming", ",", "cell", "=", "cell", ",", "dropout", "=", "dropout", ",", "return_seq", "=", "return_seq", ",", "return_state", "=", "return_state", ",", "initial_state", "=", "initial_state", ",", "dynamic", "=", "dynamic", ",", "scope", "=", "scope", ",", "name", "=", "name", ")", "return", "x" ]
simple rnn .
train
false
25,157
def is_git_dir(d): if osp.isdir(d): if (osp.isdir(osp.join(d, 'objects')) and osp.isdir(osp.join(d, 'refs'))): headref = osp.join(d, 'HEAD') return (osp.isfile(headref) or (osp.islink(headref) and os.readlink(headref).startswith('refs'))) elif (osp.isfile(osp.join(d, 'gitdir')) and osp.isfile(osp.join(d, 'commondir')) and osp.isfile(osp.join(d, 'gitfile'))): raise WorkTreeRepositoryUnsupported(d) return False
[ "def", "is_git_dir", "(", "d", ")", ":", "if", "osp", ".", "isdir", "(", "d", ")", ":", "if", "(", "osp", ".", "isdir", "(", "osp", ".", "join", "(", "d", ",", "'objects'", ")", ")", "and", "osp", ".", "isdir", "(", "osp", ".", "join", "(", "d", ",", "'refs'", ")", ")", ")", ":", "headref", "=", "osp", ".", "join", "(", "d", ",", "'HEAD'", ")", "return", "(", "osp", ".", "isfile", "(", "headref", ")", "or", "(", "osp", ".", "islink", "(", "headref", ")", "and", "os", ".", "readlink", "(", "headref", ")", ".", "startswith", "(", "'refs'", ")", ")", ")", "elif", "(", "osp", ".", "isfile", "(", "osp", ".", "join", "(", "d", ",", "'gitdir'", ")", ")", "and", "osp", ".", "isfile", "(", "osp", ".", "join", "(", "d", ",", "'commondir'", ")", ")", "and", "osp", ".", "isfile", "(", "osp", ".", "join", "(", "d", ",", "'gitfile'", ")", ")", ")", ":", "raise", "WorkTreeRepositoryUnsupported", "(", "d", ")", "return", "False" ]
from gits setup .
train
true
25,158
def advantage(cpu_dt, gpu_dt): assert ((gpu_dt >= 0) and (cpu_dt >= 0)) if ((gpu_dt == 0) and (cpu_dt == 0)): return numpy.nan elif (gpu_dt == 0): return numpy.inf else: return (cpu_dt / gpu_dt)
[ "def", "advantage", "(", "cpu_dt", ",", "gpu_dt", ")", ":", "assert", "(", "(", "gpu_dt", ">=", "0", ")", "and", "(", "cpu_dt", ">=", "0", ")", ")", "if", "(", "(", "gpu_dt", "==", "0", ")", "and", "(", "cpu_dt", "==", "0", ")", ")", ":", "return", "numpy", ".", "nan", "elif", "(", "gpu_dt", "==", "0", ")", ":", "return", "numpy", ".", "inf", "else", ":", "return", "(", "cpu_dt", "/", "gpu_dt", ")" ]
return ratio of cpu_dt / gpu_dt .
train
false
25,161
def _validate_connectivity(image_dim, connectivity, offset): if (connectivity is None): connectivity = 1 if np.isscalar(connectivity): c_connectivity = ndi.generate_binary_structure(image_dim, connectivity) else: c_connectivity = np.array(connectivity, bool) if (c_connectivity.ndim != image_dim): raise ValueError('Connectivity dimension must be same as image') if (offset is None): if any([((x % 2) == 0) for x in c_connectivity.shape]): raise ValueError('Connectivity array must have an unambiguous center') offset = (np.array(c_connectivity.shape) // 2) return (c_connectivity, offset)
[ "def", "_validate_connectivity", "(", "image_dim", ",", "connectivity", ",", "offset", ")", ":", "if", "(", "connectivity", "is", "None", ")", ":", "connectivity", "=", "1", "if", "np", ".", "isscalar", "(", "connectivity", ")", ":", "c_connectivity", "=", "ndi", ".", "generate_binary_structure", "(", "image_dim", ",", "connectivity", ")", "else", ":", "c_connectivity", "=", "np", ".", "array", "(", "connectivity", ",", "bool", ")", "if", "(", "c_connectivity", ".", "ndim", "!=", "image_dim", ")", ":", "raise", "ValueError", "(", "'Connectivity dimension must be same as image'", ")", "if", "(", "offset", "is", "None", ")", ":", "if", "any", "(", "[", "(", "(", "x", "%", "2", ")", "==", "0", ")", "for", "x", "in", "c_connectivity", ".", "shape", "]", ")", ":", "raise", "ValueError", "(", "'Connectivity array must have an unambiguous center'", ")", "offset", "=", "(", "np", ".", "array", "(", "c_connectivity", ".", "shape", ")", "//", "2", ")", "return", "(", "c_connectivity", ",", "offset", ")" ]
convert any valid connectivity to a structuring element and offset .
train
false
25,164
def unlock_file(filename): log.trace('Removing lock for {0}'.format(filename)) lock = (filename + '.lock') try: os.remove(lock) except OSError as exc: log.trace('Unable to remove lock for {0}: {1}'.format(filename, exc))
[ "def", "unlock_file", "(", "filename", ")", ":", "log", ".", "trace", "(", "'Removing lock for {0}'", ".", "format", "(", "filename", ")", ")", "lock", "=", "(", "filename", "+", "'.lock'", ")", "try", ":", "os", ".", "remove", "(", "lock", ")", "except", "OSError", "as", "exc", ":", "log", ".", "trace", "(", "'Unable to remove lock for {0}: {1}'", ".", "format", "(", "filename", ",", "exc", ")", ")" ]
unlock file .
train
true
25,165
def _course_outline_json(request, course_module): return create_xblock_info(course_module, include_child_info=True, course_outline=True, include_children_predicate=(lambda xblock: (not (xblock.category == 'vertical'))), user=request.user)
[ "def", "_course_outline_json", "(", "request", ",", "course_module", ")", ":", "return", "create_xblock_info", "(", "course_module", ",", "include_child_info", "=", "True", ",", "course_outline", "=", "True", ",", "include_children_predicate", "=", "(", "lambda", "xblock", ":", "(", "not", "(", "xblock", ".", "category", "==", "'vertical'", ")", ")", ")", ",", "user", "=", "request", ".", "user", ")" ]
returns a json representation of the course module and recursively all of its children .
train
false
25,166
def ip_to_int(ip): ret = 0 for octet in ip.split('.'): ret = ((ret * 256) + int(octet)) return ret
[ "def", "ip_to_int", "(", "ip", ")", ":", "ret", "=", "0", "for", "octet", "in", "ip", ".", "split", "(", "'.'", ")", ":", "ret", "=", "(", "(", "ret", "*", "256", ")", "+", "int", "(", "octet", ")", ")", "return", "ret" ]
converts an ip address to an integer .
train
true
25,169
@testing.requires_testing_data @requires_mayavi @requires_mne @requires_freesurfer def test_subject_source_with_fsaverage(): from mne.gui._file_traits import MRISubjectSource tempdir = _TempDir() mri = MRISubjectSource() assert_false(mri.can_create_fsaverage) assert_raises(RuntimeError, mri.create_fsaverage) mri.subjects_dir = tempdir assert_true(mri.can_create_fsaverage) mri.create_fsaverage()
[ "@", "testing", ".", "requires_testing_data", "@", "requires_mayavi", "@", "requires_mne", "@", "requires_freesurfer", "def", "test_subject_source_with_fsaverage", "(", ")", ":", "from", "mne", ".", "gui", ".", "_file_traits", "import", "MRISubjectSource", "tempdir", "=", "_TempDir", "(", ")", "mri", "=", "MRISubjectSource", "(", ")", "assert_false", "(", "mri", ".", "can_create_fsaverage", ")", "assert_raises", "(", "RuntimeError", ",", "mri", ".", "create_fsaverage", ")", "mri", ".", "subjects_dir", "=", "tempdir", "assert_true", "(", "mri", ".", "can_create_fsaverage", ")", "mri", ".", "create_fsaverage", "(", ")" ]
test subjectselector .
train
false
25,170
def build_versions_form(project): attrs = {'project': project} versions_qs = project.versions.all() active = versions_qs.filter(active=True) if active.exists(): choices = [(version.slug, version.verbose_name) for version in active] attrs['default-version'] = forms.ChoiceField(label=_('Default Version'), choices=choices, initial=project.get_default_version()) for version in versions_qs: field_name = ('version-%s' % version.slug) privacy_name = ('privacy-%s' % version.slug) if (version.type == TAG): label = ('%s (%s)' % (version.verbose_name, version.identifier[:8])) else: label = version.verbose_name attrs[field_name] = forms.BooleanField(label=label, widget=DualCheckboxWidget(version), initial=version.active, required=False) attrs[privacy_name] = forms.ChoiceField(label='privacy', choices=constants.PRIVACY_CHOICES, initial=version.privacy_level) return type('VersionsForm', (BaseVersionsForm,), attrs)
[ "def", "build_versions_form", "(", "project", ")", ":", "attrs", "=", "{", "'project'", ":", "project", "}", "versions_qs", "=", "project", ".", "versions", ".", "all", "(", ")", "active", "=", "versions_qs", ".", "filter", "(", "active", "=", "True", ")", "if", "active", ".", "exists", "(", ")", ":", "choices", "=", "[", "(", "version", ".", "slug", ",", "version", ".", "verbose_name", ")", "for", "version", "in", "active", "]", "attrs", "[", "'default-version'", "]", "=", "forms", ".", "ChoiceField", "(", "label", "=", "_", "(", "'Default Version'", ")", ",", "choices", "=", "choices", ",", "initial", "=", "project", ".", "get_default_version", "(", ")", ")", "for", "version", "in", "versions_qs", ":", "field_name", "=", "(", "'version-%s'", "%", "version", ".", "slug", ")", "privacy_name", "=", "(", "'privacy-%s'", "%", "version", ".", "slug", ")", "if", "(", "version", ".", "type", "==", "TAG", ")", ":", "label", "=", "(", "'%s (%s)'", "%", "(", "version", ".", "verbose_name", ",", "version", ".", "identifier", "[", ":", "8", "]", ")", ")", "else", ":", "label", "=", "version", ".", "verbose_name", "attrs", "[", "field_name", "]", "=", "forms", ".", "BooleanField", "(", "label", "=", "label", ",", "widget", "=", "DualCheckboxWidget", "(", "version", ")", ",", "initial", "=", "version", ".", "active", ",", "required", "=", "False", ")", "attrs", "[", "privacy_name", "]", "=", "forms", ".", "ChoiceField", "(", "label", "=", "'privacy'", ",", "choices", "=", "constants", ".", "PRIVACY_CHOICES", ",", "initial", "=", "version", ".", "privacy_level", ")", "return", "type", "(", "'VersionsForm'", ",", "(", "BaseVersionsForm", ",", ")", ",", "attrs", ")" ]
versions form with a list of versions and version privacy levels .
train
false
25,172
def award(): mode = session.s3.hrm.mode def prep(r): if (mode is not None): auth.permission.fail() return True s3.prep = prep output = s3_rest_controller() return output
[ "def", "award", "(", ")", ":", "mode", "=", "session", ".", "s3", ".", "hrm", ".", "mode", "def", "prep", "(", "r", ")", ":", "if", "(", "mode", "is", "not", "None", ")", ":", "auth", ".", "permission", ".", "fail", "(", ")", "return", "True", "s3", ".", "prep", "=", "prep", "output", "=", "s3_rest_controller", "(", ")", "return", "output" ]
volunteer awards controller .
train
false
25,173
def is_iterable(x): try: iter(x) except TypeError: return False else: return True
[ "def", "is_iterable", "(", "x", ")", ":", "try", ":", "iter", "(", "x", ")", "except", "TypeError", ":", "return", "False", "else", ":", "return", "True" ]
are we being asked to look up a list of things .
train
false
25,174
def os_like_constants(): os_like = util.get_systemd_os_like() if os_like: for os_name in os_like: if (os_name in CLI_DEFAULTS.keys()): return CLI_DEFAULTS[os_name] return {}
[ "def", "os_like_constants", "(", ")", ":", "os_like", "=", "util", ".", "get_systemd_os_like", "(", ")", "if", "os_like", ":", "for", "os_name", "in", "os_like", ":", "if", "(", "os_name", "in", "CLI_DEFAULTS", ".", "keys", "(", ")", ")", ":", "return", "CLI_DEFAULTS", "[", "os_name", "]", "return", "{", "}" ]
try to get constants for distribution with similar layout and configuration .
train
false
25,176
def is_standard_module(modname, std_path=None): modname = modname.split('.')[0] try: filename = file_from_modpath([modname]) except ImportError: return False if (filename is None): return True filename = _normalize_path(filename) if filename.startswith(_cache_normalize_path(EXT_LIB_DIR)): return False if (std_path is None): std_path = STD_LIB_DIRS for path in std_path: if filename.startswith(_cache_normalize_path(path)): return True return False
[ "def", "is_standard_module", "(", "modname", ",", "std_path", "=", "None", ")", ":", "modname", "=", "modname", ".", "split", "(", "'.'", ")", "[", "0", "]", "try", ":", "filename", "=", "file_from_modpath", "(", "[", "modname", "]", ")", "except", "ImportError", ":", "return", "False", "if", "(", "filename", "is", "None", ")", ":", "return", "True", "filename", "=", "_normalize_path", "(", "filename", ")", "if", "filename", ".", "startswith", "(", "_cache_normalize_path", "(", "EXT_LIB_DIR", ")", ")", ":", "return", "False", "if", "(", "std_path", "is", "None", ")", ":", "std_path", "=", "STD_LIB_DIRS", "for", "path", "in", "std_path", ":", "if", "filename", ".", "startswith", "(", "_cache_normalize_path", "(", "path", ")", ")", ":", "return", "True", "return", "False" ]
try to guess if a module is a standard python module :type modname: str .
train
false
25,177
def test_bootstrap_ols(): ols_fit = (lambda X, y: np.dot(np.dot(np.linalg.inv(np.dot(X.T, X)), X.T), y)) X = np.column_stack((rs.randn(50, 4), np.ones(50))) w = [2, 4, 0, 3, 5] y_noisy = (np.dot(X, w) + (rs.randn(50) * 20)) y_lownoise = (np.dot(X, w) + rs.randn(50)) n_boot = 500 w_boot_noisy = algo.bootstrap(X, y_noisy, n_boot=n_boot, func=ols_fit) w_boot_lownoise = algo.bootstrap(X, y_lownoise, n_boot=n_boot, func=ols_fit) assert_equal(w_boot_noisy.shape, (n_boot, 5)) assert_equal(w_boot_lownoise.shape, (n_boot, 5)) nose.tools.assert_greater(w_boot_noisy.std(), w_boot_lownoise.std())
[ "def", "test_bootstrap_ols", "(", ")", ":", "ols_fit", "=", "(", "lambda", "X", ",", "y", ":", "np", ".", "dot", "(", "np", ".", "dot", "(", "np", ".", "linalg", ".", "inv", "(", "np", ".", "dot", "(", "X", ".", "T", ",", "X", ")", ")", ",", "X", ".", "T", ")", ",", "y", ")", ")", "X", "=", "np", ".", "column_stack", "(", "(", "rs", ".", "randn", "(", "50", ",", "4", ")", ",", "np", ".", "ones", "(", "50", ")", ")", ")", "w", "=", "[", "2", ",", "4", ",", "0", ",", "3", ",", "5", "]", "y_noisy", "=", "(", "np", ".", "dot", "(", "X", ",", "w", ")", "+", "(", "rs", ".", "randn", "(", "50", ")", "*", "20", ")", ")", "y_lownoise", "=", "(", "np", ".", "dot", "(", "X", ",", "w", ")", "+", "rs", ".", "randn", "(", "50", ")", ")", "n_boot", "=", "500", "w_boot_noisy", "=", "algo", ".", "bootstrap", "(", "X", ",", "y_noisy", ",", "n_boot", "=", "n_boot", ",", "func", "=", "ols_fit", ")", "w_boot_lownoise", "=", "algo", ".", "bootstrap", "(", "X", ",", "y_lownoise", ",", "n_boot", "=", "n_boot", ",", "func", "=", "ols_fit", ")", "assert_equal", "(", "w_boot_noisy", ".", "shape", ",", "(", "n_boot", ",", "5", ")", ")", "assert_equal", "(", "w_boot_lownoise", ".", "shape", ",", "(", "n_boot", ",", "5", ")", ")", "nose", ".", "tools", ".", "assert_greater", "(", "w_boot_noisy", ".", "std", "(", ")", ",", "w_boot_lownoise", ".", "std", "(", ")", ")" ]
test bootstrap of ols model fit .
train
false
25,179
@contextlib.contextmanager def report_scope(observation): current = _reporters[(-1)] old = current.observation current.observation = observation (yield) current.observation = old
[ "@", "contextlib", ".", "contextmanager", "def", "report_scope", "(", "observation", ")", ":", "current", "=", "_reporters", "[", "(", "-", "1", ")", "]", "old", "=", "current", ".", "observation", "current", ".", "observation", "=", "observation", "(", "yield", ")", "current", ".", "observation", "=", "old" ]
returns a report scope with the current reporter .
train
false
25,180
def test_masked_row_with_object_col(): t = table.Table([[1]], dtype=['O'], masked=True) if NUMPY_LT_1_8: with pytest.raises(ValueError): t['col0'].mask = False t[0].as_void() with pytest.raises(ValueError): t['col0'].mask = True t[0].as_void() else: t['col0'].mask = False assert (t[0]['col0'] == 1) t['col0'].mask = True assert (t[0]['col0'] is np.ma.masked)
[ "def", "test_masked_row_with_object_col", "(", ")", ":", "t", "=", "table", ".", "Table", "(", "[", "[", "1", "]", "]", ",", "dtype", "=", "[", "'O'", "]", ",", "masked", "=", "True", ")", "if", "NUMPY_LT_1_8", ":", "with", "pytest", ".", "raises", "(", "ValueError", ")", ":", "t", "[", "'col0'", "]", ".", "mask", "=", "False", "t", "[", "0", "]", ".", "as_void", "(", ")", "with", "pytest", ".", "raises", "(", "ValueError", ")", ":", "t", "[", "'col0'", "]", ".", "mask", "=", "True", "t", "[", "0", "]", ".", "as_void", "(", ")", "else", ":", "t", "[", "'col0'", "]", ".", "mask", "=", "False", "assert", "(", "t", "[", "0", "]", "[", "'col0'", "]", "==", "1", ")", "t", "[", "'col0'", "]", ".", "mask", "=", "True", "assert", "(", "t", "[", "0", "]", "[", "'col0'", "]", "is", "np", ".", "ma", ".", "masked", ")" ]
numpy < 1 .
train
false
25,182
def get_organization_courses(organization_id): if (not organizations_enabled()): return [] from organizations import api as organizations_api return organizations_api.get_organization_courses(organization_id)
[ "def", "get_organization_courses", "(", "organization_id", ")", ":", "if", "(", "not", "organizations_enabled", "(", ")", ")", ":", "return", "[", "]", "from", "organizations", "import", "api", "as", "organizations_api", "return", "organizations_api", ".", "get_organization_courses", "(", "organization_id", ")" ]
client api operation adapter/wrapper .
train
false
25,185
def supportsTeams(endpoint): return endpoint.usesExtension(ns_uri)
[ "def", "supportsTeams", "(", "endpoint", ")", ":", "return", "endpoint", ".", "usesExtension", "(", "ns_uri", ")" ]
does the given endpoint advertise support for launchpad teams? .
train
false
25,186
@pytest.mark.django_db @pytest.mark.parametrize('view,model,get_object,service_provider_attr', [(PaymentMethodEditView, PaymentMethod, get_default_payment_method, 'payment_processor'), (ShippingMethodEditView, ShippingMethod, get_default_shipping_method, 'carrier')]) def test_behavior_add_save(rf, admin_user, view, model, get_object, service_provider_attr): get_default_shop() with override_settings(LANGUAGES=[('en', 'en')]): object = get_object() view = view.as_view() service_provider_attr_field = ('base-%s' % service_provider_attr) data = get_default_data(object, service_provider_attr, service_provider_attr_field) components_before = object.behavior_components.count() assert (not components_before) request = apply_request_middleware(rf.post('/', data=data, user=admin_user)) view(request, pk=object.pk) components_after = object.behavior_components.count() assert (components_after == len(get_default_behavior_settings()))
[ "@", "pytest", ".", "mark", ".", "django_db", "@", "pytest", ".", "mark", ".", "parametrize", "(", "'view,model,get_object,service_provider_attr'", ",", "[", "(", "PaymentMethodEditView", ",", "PaymentMethod", ",", "get_default_payment_method", ",", "'payment_processor'", ")", ",", "(", "ShippingMethodEditView", ",", "ShippingMethod", ",", "get_default_shipping_method", ",", "'carrier'", ")", "]", ")", "def", "test_behavior_add_save", "(", "rf", ",", "admin_user", ",", "view", ",", "model", ",", "get_object", ",", "service_provider_attr", ")", ":", "get_default_shop", "(", ")", "with", "override_settings", "(", "LANGUAGES", "=", "[", "(", "'en'", ",", "'en'", ")", "]", ")", ":", "object", "=", "get_object", "(", ")", "view", "=", "view", ".", "as_view", "(", ")", "service_provider_attr_field", "=", "(", "'base-%s'", "%", "service_provider_attr", ")", "data", "=", "get_default_data", "(", "object", ",", "service_provider_attr", ",", "service_provider_attr_field", ")", "components_before", "=", "object", ".", "behavior_components", ".", "count", "(", ")", "assert", "(", "not", "components_before", ")", "request", "=", "apply_request_middleware", "(", "rf", ".", "post", "(", "'/'", ",", "data", "=", "data", ",", "user", "=", "admin_user", ")", ")", "view", "(", "request", ",", "pk", "=", "object", ".", "pk", ")", "components_after", "=", "object", ".", "behavior_components", ".", "count", "(", ")", "assert", "(", "components_after", "==", "len", "(", "get_default_behavior_settings", "(", ")", ")", ")" ]
to make things little bit more simple lets use only english as a language .
train
false
25,187
def preprocess_file(filename, cpp_path='cpp', cpp_args=''): path_list = [cpp_path] if isinstance(cpp_args, list): path_list += cpp_args elif (cpp_args != ''): path_list += [cpp_args] path_list += [filename] try: pipe = Popen(path_list, stdout=PIPE, universal_newlines=True) text = pipe.communicate()[0] except OSError as e: raise RuntimeError((("Unable to invoke 'cpp'. " + 'Make sure its path was passed correctly\n') + ('Original error: %s' % e))) return text
[ "def", "preprocess_file", "(", "filename", ",", "cpp_path", "=", "'cpp'", ",", "cpp_args", "=", "''", ")", ":", "path_list", "=", "[", "cpp_path", "]", "if", "isinstance", "(", "cpp_args", ",", "list", ")", ":", "path_list", "+=", "cpp_args", "elif", "(", "cpp_args", "!=", "''", ")", ":", "path_list", "+=", "[", "cpp_args", "]", "path_list", "+=", "[", "filename", "]", "try", ":", "pipe", "=", "Popen", "(", "path_list", ",", "stdout", "=", "PIPE", ",", "universal_newlines", "=", "True", ")", "text", "=", "pipe", ".", "communicate", "(", ")", "[", "0", "]", "except", "OSError", "as", "e", ":", "raise", "RuntimeError", "(", "(", "(", "\"Unable to invoke 'cpp'. \"", "+", "'Make sure its path was passed correctly\\n'", ")", "+", "(", "'Original error: %s'", "%", "e", ")", ")", ")", "return", "text" ]
preprocess a file using cpp .
train
false
25,188
@cli.command() @click.option('--username', prompt=True, help="The developer's shown username.") @click.option('--email', prompt='E-Mail', help="The developer's email address") @click.password_option(help='The login password.') @pass_repo def setuser(repo, username, email, password): repo.set_config('username', username) repo.set_config('email', email) repo.set_config('password', ('*' * len(password))) click.echo('Changed credentials.')
[ "@", "cli", ".", "command", "(", ")", "@", "click", ".", "option", "(", "'--username'", ",", "prompt", "=", "True", ",", "help", "=", "\"The developer's shown username.\"", ")", "@", "click", ".", "option", "(", "'--email'", ",", "prompt", "=", "'E-Mail'", ",", "help", "=", "\"The developer's email address\"", ")", "@", "click", ".", "password_option", "(", "help", "=", "'The login password.'", ")", "@", "pass_repo", "def", "setuser", "(", "repo", ",", "username", ",", "email", ",", "password", ")", ":", "repo", ".", "set_config", "(", "'username'", ",", "username", ")", "repo", ".", "set_config", "(", "'email'", ",", "email", ")", "repo", ".", "set_config", "(", "'password'", ",", "(", "'*'", "*", "len", "(", "password", ")", ")", ")", "click", ".", "echo", "(", "'Changed credentials.'", ")" ]
sets the user credentials .
train
false
25,189
def load_pytest_conf(path, parser): namespace = {} exec open(path, 'rb').read() in namespace if ('update_parser' in namespace): namespace['update_parser'](parser) return namespace.get('CustomPyTester', PyTester)
[ "def", "load_pytest_conf", "(", "path", ",", "parser", ")", ":", "namespace", "=", "{", "}", "exec", "open", "(", "path", ",", "'rb'", ")", ".", "read", "(", ")", "in", "namespace", "if", "(", "'update_parser'", "in", "namespace", ")", ":", "namespace", "[", "'update_parser'", "]", "(", "parser", ")", "return", "namespace", ".", "get", "(", "'CustomPyTester'", ",", "PyTester", ")" ]
loads a pytestconf .
train
false
25,190
def gen_lib_options(compiler, library_dirs, runtime_library_dirs, libraries): lib_opts = [] for dir in library_dirs: lib_opts.append(compiler.library_dir_option(dir)) for dir in runtime_library_dirs: opt = compiler.runtime_library_dir_option(dir) if isinstance(opt, list): lib_opts = (lib_opts + opt) else: lib_opts.append(opt) for lib in libraries: (lib_dir, lib_name) = os.path.split(lib) if lib_dir: lib_file = compiler.find_library_file([lib_dir], lib_name) if lib_file: lib_opts.append(lib_file) else: compiler.warn(("no library file corresponding to '%s' found (skipping)" % lib)) else: lib_opts.append(compiler.library_option(lib)) return lib_opts
[ "def", "gen_lib_options", "(", "compiler", ",", "library_dirs", ",", "runtime_library_dirs", ",", "libraries", ")", ":", "lib_opts", "=", "[", "]", "for", "dir", "in", "library_dirs", ":", "lib_opts", ".", "append", "(", "compiler", ".", "library_dir_option", "(", "dir", ")", ")", "for", "dir", "in", "runtime_library_dirs", ":", "opt", "=", "compiler", ".", "runtime_library_dir_option", "(", "dir", ")", "if", "isinstance", "(", "opt", ",", "list", ")", ":", "lib_opts", "=", "(", "lib_opts", "+", "opt", ")", "else", ":", "lib_opts", ".", "append", "(", "opt", ")", "for", "lib", "in", "libraries", ":", "(", "lib_dir", ",", "lib_name", ")", "=", "os", ".", "path", ".", "split", "(", "lib", ")", "if", "lib_dir", ":", "lib_file", "=", "compiler", ".", "find_library_file", "(", "[", "lib_dir", "]", ",", "lib_name", ")", "if", "lib_file", ":", "lib_opts", ".", "append", "(", "lib_file", ")", "else", ":", "compiler", ".", "warn", "(", "(", "\"no library file corresponding to '%s' found (skipping)\"", "%", "lib", ")", ")", "else", ":", "lib_opts", ".", "append", "(", "compiler", ".", "library_option", "(", "lib", ")", ")", "return", "lib_opts" ]
generate linker options for searching library directories and linking with specific libraries .
train
false
25,191
def py_encode(symb2freq): heap = [[wt, [sym, u'']] for (sym, wt) in symb2freq.items()] heapify(heap) while (len(heap) > 1): lo = heappop(heap) hi = heappop(heap) for pair in lo[1:]: pair[1] = (u'0' + pair[1]) for pair in hi[1:]: pair[1] = (u'1' + pair[1]) heappush(heap, (([(lo[0] + hi[0])] + lo[1:]) + hi[1:])) return dict(heappop(heap)[1:])
[ "def", "py_encode", "(", "symb2freq", ")", ":", "heap", "=", "[", "[", "wt", ",", "[", "sym", ",", "u''", "]", "]", "for", "(", "sym", ",", "wt", ")", "in", "symb2freq", ".", "items", "(", ")", "]", "heapify", "(", "heap", ")", "while", "(", "len", "(", "heap", ")", ">", "1", ")", ":", "lo", "=", "heappop", "(", "heap", ")", "hi", "=", "heappop", "(", "heap", ")", "for", "pair", "in", "lo", "[", "1", ":", "]", ":", "pair", "[", "1", "]", "=", "(", "u'0'", "+", "pair", "[", "1", "]", ")", "for", "pair", "in", "hi", "[", "1", ":", "]", ":", "pair", "[", "1", "]", "=", "(", "u'1'", "+", "pair", "[", "1", "]", ")", "heappush", "(", "heap", ",", "(", "(", "[", "(", "lo", "[", "0", "]", "+", "hi", "[", "0", "]", ")", "]", "+", "lo", "[", "1", ":", "]", ")", "+", "hi", "[", "1", ":", "]", ")", ")", "return", "dict", "(", "heappop", "(", "heap", ")", "[", "1", ":", "]", ")" ]
huffman encode the given dict mapping symbols to weights from rosetta code .
train
false
25,193
def sign_certificate_request(keypair, dn, request, serial, validity_period, digest, start=None, additional_extensions=()): if (start is None): start = datetime.datetime.utcnow() expire = (start + datetime.timedelta(seconds=validity_period)) start = start.strftime('%Y%m%d%H%M%SZ') expire = expire.strftime('%Y%m%d%H%M%SZ') req = request.original cert = crypto.X509() cert.set_issuer(dn) cert.set_subject(req.get_subject()) cert.set_pubkey(req.get_pubkey()) cert.set_notBefore(start) cert.set_notAfter(expire) cert.set_serial_number(serial) cert.add_extensions(additional_extensions) cert.sign(keypair.original, digest) return Certificate(cert)
[ "def", "sign_certificate_request", "(", "keypair", ",", "dn", ",", "request", ",", "serial", ",", "validity_period", ",", "digest", ",", "start", "=", "None", ",", "additional_extensions", "=", "(", ")", ")", ":", "if", "(", "start", "is", "None", ")", ":", "start", "=", "datetime", ".", "datetime", ".", "utcnow", "(", ")", "expire", "=", "(", "start", "+", "datetime", ".", "timedelta", "(", "seconds", "=", "validity_period", ")", ")", "start", "=", "start", ".", "strftime", "(", "'%Y%m%d%H%M%SZ'", ")", "expire", "=", "expire", ".", "strftime", "(", "'%Y%m%d%H%M%SZ'", ")", "req", "=", "request", ".", "original", "cert", "=", "crypto", ".", "X509", "(", ")", "cert", ".", "set_issuer", "(", "dn", ")", "cert", ".", "set_subject", "(", "req", ".", "get_subject", "(", ")", ")", "cert", ".", "set_pubkey", "(", "req", ".", "get_pubkey", "(", ")", ")", "cert", ".", "set_notBefore", "(", "start", ")", "cert", ".", "set_notAfter", "(", "expire", ")", "cert", ".", "set_serial_number", "(", "serial", ")", "cert", ".", "add_extensions", "(", "additional_extensions", ")", "cert", ".", "sign", "(", "keypair", ".", "original", ",", "digest", ")", "return", "Certificate", "(", "cert", ")" ]
sign a certificaterequest and return a certificate .
train
false
25,194
def getNewRepository(): return ExportRepository()
[ "def", "getNewRepository", "(", ")", ":", "return", "ExportRepository", "(", ")" ]
get the repository constructor .
train
false
25,195
def remove_profile_images(profile_image_names): storage = get_profile_image_storage() for name in profile_image_names.values(): storage.delete(name)
[ "def", "remove_profile_images", "(", "profile_image_names", ")", ":", "storage", "=", "get_profile_image_storage", "(", ")", "for", "name", "in", "profile_image_names", ".", "values", "(", ")", ":", "storage", ".", "delete", "(", "name", ")" ]
physically remove the image files specified in profile_image_names .
train
false
25,196
def profile_start(name): _profiles_running[name] = time.time() _profiles_stack.append(name)
[ "def", "profile_start", "(", "name", ")", ":", "_profiles_running", "[", "name", "]", "=", "time", ".", "time", "(", ")", "_profiles_stack", ".", "append", "(", "name", ")" ]
starts a profiling interval with specific name profiling data is sent to the client with next data batch .
train
false
25,197
def getLargestCenterOutsetLoopFromLoop(loop, radius, thresholdRatio=0.9): if (radius == 0.0): return loop radius = abs(radius) points = getPointsFromLoop(loop, radius, thresholdRatio) centers = getCentersFromPoints(points, (globalIntercircleMultiplier * radius)) largestCenterOutset = None largestOutsetArea = (-987654321.0) for center in centers: outset = getSimplifiedInsetFromClockwiseLoop(center, radius) if isLargeSameDirection(outset, center, radius): if (euclidean.isPathInsideLoop(loop, outset) != euclidean.isWiddershins(loop)): centerOutset = CenterOutset(center, outset) outsetArea = abs(euclidean.getAreaLoop(outset)) if (outsetArea > largestOutsetArea): largestOutsetArea = outsetArea largestCenterOutset = centerOutset if (largestCenterOutset == None): return None largestCenterOutset.center = euclidean.getSimplifiedLoop(largestCenterOutset.center, radius) return largestCenterOutset
[ "def", "getLargestCenterOutsetLoopFromLoop", "(", "loop", ",", "radius", ",", "thresholdRatio", "=", "0.9", ")", ":", "if", "(", "radius", "==", "0.0", ")", ":", "return", "loop", "radius", "=", "abs", "(", "radius", ")", "points", "=", "getPointsFromLoop", "(", "loop", ",", "radius", ",", "thresholdRatio", ")", "centers", "=", "getCentersFromPoints", "(", "points", ",", "(", "globalIntercircleMultiplier", "*", "radius", ")", ")", "largestCenterOutset", "=", "None", "largestOutsetArea", "=", "(", "-", "987654321.0", ")", "for", "center", "in", "centers", ":", "outset", "=", "getSimplifiedInsetFromClockwiseLoop", "(", "center", ",", "radius", ")", "if", "isLargeSameDirection", "(", "outset", ",", "center", ",", "radius", ")", ":", "if", "(", "euclidean", ".", "isPathInsideLoop", "(", "loop", ",", "outset", ")", "!=", "euclidean", ".", "isWiddershins", "(", "loop", ")", ")", ":", "centerOutset", "=", "CenterOutset", "(", "center", ",", "outset", ")", "outsetArea", "=", "abs", "(", "euclidean", ".", "getAreaLoop", "(", "outset", ")", ")", "if", "(", "outsetArea", ">", "largestOutsetArea", ")", ":", "largestOutsetArea", "=", "outsetArea", "largestCenterOutset", "=", "centerOutset", "if", "(", "largestCenterOutset", "==", "None", ")", ":", "return", "None", "largestCenterOutset", ".", "center", "=", "euclidean", ".", "getSimplifiedLoop", "(", "largestCenterOutset", ".", "center", ",", "radius", ")", "return", "largestCenterOutset" ]
get the largest circle outset loop from the loop .
train
false
25,198
def copy_data(data, dest, header=None, use_put=None): ret = None if use_put: udata = data else: udata = urllib.urlencode(data) if utils.is_url(dest): ret = copy_remote(udata, dest, use_put) if header: return ret[header] else: if header: ret = (dest + str(time.time())) dest = (ret + '/_task_result') copy_local(udata, dest, use_put) return ret
[ "def", "copy_data", "(", "data", ",", "dest", ",", "header", "=", "None", ",", "use_put", "=", "None", ")", ":", "ret", "=", "None", "if", "use_put", ":", "udata", "=", "data", "else", ":", "udata", "=", "urllib", ".", "urlencode", "(", "data", ")", "if", "utils", ".", "is_url", "(", "dest", ")", ":", "ret", "=", "copy_remote", "(", "udata", ",", "dest", ",", "use_put", ")", "if", "header", ":", "return", "ret", "[", "header", "]", "else", ":", "if", "header", ":", "ret", "=", "(", "dest", "+", "str", "(", "time", ".", "time", "(", ")", ")", ")", "dest", "=", "(", "ret", "+", "'/_task_result'", ")", "copy_local", "(", "udata", ",", "dest", ",", "use_put", ")", "return", "ret" ]
copy data to a destination to aid in debugging .
train
false
25,199
def _any_pandas_objects(terms): return any((isinstance(term.value, pd.core.generic.PandasObject) for term in terms))
[ "def", "_any_pandas_objects", "(", "terms", ")", ":", "return", "any", "(", "(", "isinstance", "(", "term", ".", "value", ",", "pd", ".", "core", ".", "generic", ".", "PandasObject", ")", "for", "term", "in", "terms", ")", ")" ]
check a sequence of terms for instances of pandasobject .
train
true
25,200
def make_dataset(X, y, sample_weight, random_state=None): rng = check_random_state(random_state) seed = rng.randint(1, np.iinfo(np.int32).max) if sp.issparse(X): dataset = CSRDataset(X.data, X.indptr, X.indices, y, sample_weight, seed=seed) intercept_decay = SPARSE_INTERCEPT_DECAY else: dataset = ArrayDataset(X, y, sample_weight, seed=seed) intercept_decay = 1.0 return (dataset, intercept_decay)
[ "def", "make_dataset", "(", "X", ",", "y", ",", "sample_weight", ",", "random_state", "=", "None", ")", ":", "rng", "=", "check_random_state", "(", "random_state", ")", "seed", "=", "rng", ".", "randint", "(", "1", ",", "np", ".", "iinfo", "(", "np", ".", "int32", ")", ".", "max", ")", "if", "sp", ".", "issparse", "(", "X", ")", ":", "dataset", "=", "CSRDataset", "(", "X", ".", "data", ",", "X", ".", "indptr", ",", "X", ".", "indices", ",", "y", ",", "sample_weight", ",", "seed", "=", "seed", ")", "intercept_decay", "=", "SPARSE_INTERCEPT_DECAY", "else", ":", "dataset", "=", "ArrayDataset", "(", "X", ",", "y", ",", "sample_weight", ",", "seed", "=", "seed", ")", "intercept_decay", "=", "1.0", "return", "(", "dataset", ",", "intercept_decay", ")" ]
create dataset abstraction for sparse and dense inputs .
train
false
25,201
def askUser(text, parent=None, help='', defaultno=False, msgfunc=None, title='Anki'): if (not parent): parent = aqt.mw.app.activeWindow() if (not msgfunc): msgfunc = QMessageBox.question sb = (QMessageBox.Yes | QMessageBox.No) if help: sb |= QMessageBox.Help while 1: if defaultno: default = QMessageBox.No else: default = QMessageBox.Yes r = msgfunc(parent, title, text, sb, default) if (r == QMessageBox.Help): openHelp(help) else: break return (r == QMessageBox.Yes)
[ "def", "askUser", "(", "text", ",", "parent", "=", "None", ",", "help", "=", "''", ",", "defaultno", "=", "False", ",", "msgfunc", "=", "None", ",", "title", "=", "'Anki'", ")", ":", "if", "(", "not", "parent", ")", ":", "parent", "=", "aqt", ".", "mw", ".", "app", ".", "activeWindow", "(", ")", "if", "(", "not", "msgfunc", ")", ":", "msgfunc", "=", "QMessageBox", ".", "question", "sb", "=", "(", "QMessageBox", ".", "Yes", "|", "QMessageBox", ".", "No", ")", "if", "help", ":", "sb", "|=", "QMessageBox", ".", "Help", "while", "1", ":", "if", "defaultno", ":", "default", "=", "QMessageBox", ".", "No", "else", ":", "default", "=", "QMessageBox", ".", "Yes", "r", "=", "msgfunc", "(", "parent", ",", "title", ",", "text", ",", "sb", ",", "default", ")", "if", "(", "r", "==", "QMessageBox", ".", "Help", ")", ":", "openHelp", "(", "help", ")", "else", ":", "break", "return", "(", "r", "==", "QMessageBox", ".", "Yes", ")" ]
show a yes/no question .
train
false
25,203
def broadcast_to_sharejs(action, sharejs_uuid, node=None, wiki_name='home', data=None): url = 'http://{host}:{port}/{action}/{id}/'.format(host=wiki_settings.SHAREJS_HOST, port=wiki_settings.SHAREJS_PORT, action=action, id=sharejs_uuid) if ((action == 'redirect') or (action == 'delete')): redirect_url = urllib.quote(node.web_url_for('project_wiki_view', wname=wiki_name, _guid=True), safe='') url = os.path.join(url, redirect_url) try: requests.post(url, json=data) except requests.ConnectionError: pass
[ "def", "broadcast_to_sharejs", "(", "action", ",", "sharejs_uuid", ",", "node", "=", "None", ",", "wiki_name", "=", "'home'", ",", "data", "=", "None", ")", ":", "url", "=", "'http://{host}:{port}/{action}/{id}/'", ".", "format", "(", "host", "=", "wiki_settings", ".", "SHAREJS_HOST", ",", "port", "=", "wiki_settings", ".", "SHAREJS_PORT", ",", "action", "=", "action", ",", "id", "=", "sharejs_uuid", ")", "if", "(", "(", "action", "==", "'redirect'", ")", "or", "(", "action", "==", "'delete'", ")", ")", ":", "redirect_url", "=", "urllib", ".", "quote", "(", "node", ".", "web_url_for", "(", "'project_wiki_view'", ",", "wname", "=", "wiki_name", ",", "_guid", "=", "True", ")", ",", "safe", "=", "''", ")", "url", "=", "os", ".", "path", ".", "join", "(", "url", ",", "redirect_url", ")", "try", ":", "requests", ".", "post", "(", "url", ",", "json", "=", "data", ")", "except", "requests", ".", "ConnectionError", ":", "pass" ]
broadcast an action to all documents connected to a wiki .
train
false
25,204
def dict_formatter(view, value): return json.dumps(value, ensure_ascii=False)
[ "def", "dict_formatter", "(", "view", ",", "value", ")", ":", "return", "json", ".", "dumps", "(", "value", ",", "ensure_ascii", "=", "False", ")" ]
removes unicode entities when displaying dict as string .
train
false
25,205
def test_eppstein_matching(): G = nx.Graph() G.add_nodes_from(['a', 2, 3, 4], bipartite=0) G.add_nodes_from([1, 'b', 'c'], bipartite=1) G.add_edges_from([('a', 1), ('a', 'b'), (2, 'b'), (2, 'c'), (3, 'c'), (4, 1)]) matching = eppstein_matching(G) assert_true((len(matching) == len(maximum_matching(G)))) assert all(((x in set(matching.keys())) for x in set(matching.values())))
[ "def", "test_eppstein_matching", "(", ")", ":", "G", "=", "nx", ".", "Graph", "(", ")", "G", ".", "add_nodes_from", "(", "[", "'a'", ",", "2", ",", "3", ",", "4", "]", ",", "bipartite", "=", "0", ")", "G", ".", "add_nodes_from", "(", "[", "1", ",", "'b'", ",", "'c'", "]", ",", "bipartite", "=", "1", ")", "G", ".", "add_edges_from", "(", "[", "(", "'a'", ",", "1", ")", ",", "(", "'a'", ",", "'b'", ")", ",", "(", "2", ",", "'b'", ")", ",", "(", "2", ",", "'c'", ")", ",", "(", "3", ",", "'c'", ")", ",", "(", "4", ",", "1", ")", "]", ")", "matching", "=", "eppstein_matching", "(", "G", ")", "assert_true", "(", "(", "len", "(", "matching", ")", "==", "len", "(", "maximum_matching", "(", "G", ")", ")", ")", ")", "assert", "all", "(", "(", "(", "x", "in", "set", "(", "matching", ".", "keys", "(", ")", ")", ")", "for", "x", "in", "set", "(", "matching", ".", "values", "(", ")", ")", ")", ")" ]
test in accordance to issue #1927 .
train
false
25,206
def clashing(): ns = {} exec_('from sympy import *', ns) clash1 = {} clash2 = {} while ns: (k, _) = ns.popitem() if (k in _greek): clash2[k] = Symbol(k) _greek.remove(k) elif (k in _latin): clash1[k] = Symbol(k) _latin.remove(k) clash = {} clash.update(clash1) clash.update(clash2) return (clash1, clash2, clash)
[ "def", "clashing", "(", ")", ":", "ns", "=", "{", "}", "exec_", "(", "'from sympy import *'", ",", "ns", ")", "clash1", "=", "{", "}", "clash2", "=", "{", "}", "while", "ns", ":", "(", "k", ",", "_", ")", "=", "ns", ".", "popitem", "(", ")", "if", "(", "k", "in", "_greek", ")", ":", "clash2", "[", "k", "]", "=", "Symbol", "(", "k", ")", "_greek", ".", "remove", "(", "k", ")", "elif", "(", "k", "in", "_latin", ")", ":", "clash1", "[", "k", "]", "=", "Symbol", "(", "k", ")", "_latin", ".", "remove", "(", "k", ")", "clash", "=", "{", "}", "clash", ".", "update", "(", "clash1", ")", "clash", ".", "update", "(", "clash2", ")", "return", "(", "clash1", ",", "clash2", ",", "clash", ")" ]
return the clashing-symbols dictionaries .
train
false
25,207
@handle_response_format @treeio_login_required def item_view(request, folderPath, itemPath, response_format='html'): try: item = KnowledgeItem.by_path(folderPath, itemPath) except KnowledgeItem.DoesNotExist: raise Http404 if (not item): raise Http404 items = Object.filter_permitted(manager=KnowledgeItem.objects, user=request.user.profile, mode='r') if (not request.user.profile.has_permission(item)): return user_denied(request, message="You don't have access to this Knowledge Item") context = _get_default_context(request) context.update({'items': items, 'item': item}) return render_to_response('knowledge/item_view', context, context_instance=RequestContext(request), response_format=response_format)
[ "@", "handle_response_format", "@", "treeio_login_required", "def", "item_view", "(", "request", ",", "folderPath", ",", "itemPath", ",", "response_format", "=", "'html'", ")", ":", "try", ":", "item", "=", "KnowledgeItem", ".", "by_path", "(", "folderPath", ",", "itemPath", ")", "except", "KnowledgeItem", ".", "DoesNotExist", ":", "raise", "Http404", "if", "(", "not", "item", ")", ":", "raise", "Http404", "items", "=", "Object", ".", "filter_permitted", "(", "manager", "=", "KnowledgeItem", ".", "objects", ",", "user", "=", "request", ".", "user", ".", "profile", ",", "mode", "=", "'r'", ")", "if", "(", "not", "request", ".", "user", ".", "profile", ".", "has_permission", "(", "item", ")", ")", ":", "return", "user_denied", "(", "request", ",", "message", "=", "\"You don't have access to this Knowledge Item\"", ")", "context", "=", "_get_default_context", "(", "request", ")", "context", ".", "update", "(", "{", "'items'", ":", "items", ",", "'item'", ":", "item", "}", ")", "return", "render_to_response", "(", "'knowledge/item_view'", ",", "context", ",", "context_instance", "=", "RequestContext", "(", "request", ")", ",", "response_format", "=", "response_format", ")" ]
item view .
train
false
25,208
@before.each_scenario def process_requires_tags(scenario): tag_re = re.compile('requires_stub_(?P<server>[^_]+)') for tag in scenario.tags: requires = tag_re.match(tag) if requires: if (requires.group('server') == 'youtube'): if (not is_youtube_available(YOUTUBE_API_URLS)): scenario.steps = [] return start_stub(requires.group('server'))
[ "@", "before", ".", "each_scenario", "def", "process_requires_tags", "(", "scenario", ")", ":", "tag_re", "=", "re", ".", "compile", "(", "'requires_stub_(?P<server>[^_]+)'", ")", "for", "tag", "in", "scenario", ".", "tags", ":", "requires", "=", "tag_re", ".", "match", "(", "tag", ")", "if", "requires", ":", "if", "(", "requires", ".", "group", "(", "'server'", ")", "==", "'youtube'", ")", ":", "if", "(", "not", "is_youtube_available", "(", "YOUTUBE_API_URLS", ")", ")", ":", "scenario", ".", "steps", "=", "[", "]", "return", "start_stub", "(", "requires", ".", "group", "(", "'server'", ")", ")" ]
process the scenario tags to make sure that any requirements are met prior to that scenario being executed .
train
false
25,209
def api_prefix(url_prefix=None, api_version=None): if (url_prefix is None): url_prefix = config.URL_PREFIX if (api_version is None): api_version = config.API_VERSION prefix = (('/%s' % url_prefix) if url_prefix else '') version = (('/%s' % api_version) if api_version else '') return (prefix + version)
[ "def", "api_prefix", "(", "url_prefix", "=", "None", ",", "api_version", "=", "None", ")", ":", "if", "(", "url_prefix", "is", "None", ")", ":", "url_prefix", "=", "config", ".", "URL_PREFIX", "if", "(", "api_version", "is", "None", ")", ":", "api_version", "=", "config", ".", "API_VERSION", "prefix", "=", "(", "(", "'/%s'", "%", "url_prefix", ")", "if", "url_prefix", "else", "''", ")", "version", "=", "(", "(", "'/%s'", "%", "api_version", ")", "if", "api_version", "else", "''", ")", "return", "(", "prefix", "+", "version", ")" ]
returns the prefix to api endpoints .
train
false
25,212
def test_human_readable_custom(): f = formatters.HumanReadable() assert (f(None) == u('\xe2\x88\x85')) f = formatters.HumanReadable(none_char='/') assert (f(None) == '/')
[ "def", "test_human_readable_custom", "(", ")", ":", "f", "=", "formatters", ".", "HumanReadable", "(", ")", "assert", "(", "f", "(", "None", ")", "==", "u", "(", "'\\xe2\\x88\\x85'", ")", ")", "f", "=", "formatters", ".", "HumanReadable", "(", "none_char", "=", "'/'", ")", "assert", "(", "f", "(", "None", ")", "==", "'/'", ")" ]
test human_readable formatter option .
train
false
25,213
def gather_bootstrap_script(bootstrap=None): if (not HAS_CLOUD): return (False, 'config.gather_bootstrap_script is unavailable') ret = salt.utils.cloud.update_bootstrap(__opts__, url=bootstrap) if (('Success' in ret) and (len(ret['Success']['Files updated']) > 0)): return ret['Success']['Files updated'][0]
[ "def", "gather_bootstrap_script", "(", "bootstrap", "=", "None", ")", ":", "if", "(", "not", "HAS_CLOUD", ")", ":", "return", "(", "False", ",", "'config.gather_bootstrap_script is unavailable'", ")", "ret", "=", "salt", ".", "utils", ".", "cloud", ".", "update_bootstrap", "(", "__opts__", ",", "url", "=", "bootstrap", ")", "if", "(", "(", "'Success'", "in", "ret", ")", "and", "(", "len", "(", "ret", "[", "'Success'", "]", "[", "'Files updated'", "]", ")", ">", "0", ")", ")", ":", "return", "ret", "[", "'Success'", "]", "[", "'Files updated'", "]", "[", "0", "]" ]
download the salt-bootstrap script .
train
true
25,214
def float_to_byte(value, mantissabits=5, zeroexp=2): fzero = ((63 - zeroexp) << mantissabits) bits = unpack('i', pack('f', value))[0] smallfloat = (bits >> (24 - mantissabits)) if (smallfloat < fzero): if (bits <= 0): result = chr(0) else: result = chr(1) elif (smallfloat >= (fzero + 256)): result = chr(255) else: result = chr((smallfloat - fzero)) return b(result)
[ "def", "float_to_byte", "(", "value", ",", "mantissabits", "=", "5", ",", "zeroexp", "=", "2", ")", ":", "fzero", "=", "(", "(", "63", "-", "zeroexp", ")", "<<", "mantissabits", ")", "bits", "=", "unpack", "(", "'i'", ",", "pack", "(", "'f'", ",", "value", ")", ")", "[", "0", "]", "smallfloat", "=", "(", "bits", ">>", "(", "24", "-", "mantissabits", ")", ")", "if", "(", "smallfloat", "<", "fzero", ")", ":", "if", "(", "bits", "<=", "0", ")", ":", "result", "=", "chr", "(", "0", ")", "else", ":", "result", "=", "chr", "(", "1", ")", "elif", "(", "smallfloat", ">=", "(", "fzero", "+", "256", ")", ")", ":", "result", "=", "chr", "(", "255", ")", "else", ":", "result", "=", "chr", "(", "(", "smallfloat", "-", "fzero", ")", ")", "return", "b", "(", "result", ")" ]
encodes a floating point number in a single byte .
train
false
25,216
def test_sigmoid(): def ref_sigmoid(x): if (x >= 0): return (1 / (1 + np.exp((- x)))) else: z = np.exp(x) return (z / (1 + z)) sigmoid = np.vectorize(ref_sigmoid) x = K.placeholder(ndim=2) f = K.function([x], [activations.sigmoid(x)]) test_values = get_standard_values() result = f([test_values])[0] expected = sigmoid(test_values) assert_allclose(result, expected, rtol=1e-05)
[ "def", "test_sigmoid", "(", ")", ":", "def", "ref_sigmoid", "(", "x", ")", ":", "if", "(", "x", ">=", "0", ")", ":", "return", "(", "1", "/", "(", "1", "+", "np", ".", "exp", "(", "(", "-", "x", ")", ")", ")", ")", "else", ":", "z", "=", "np", ".", "exp", "(", "x", ")", "return", "(", "z", "/", "(", "1", "+", "z", ")", ")", "sigmoid", "=", "np", ".", "vectorize", "(", "ref_sigmoid", ")", "x", "=", "K", ".", "placeholder", "(", "ndim", "=", "2", ")", "f", "=", "K", ".", "function", "(", "[", "x", "]", ",", "[", "activations", ".", "sigmoid", "(", "x", ")", "]", ")", "test_values", "=", "get_standard_values", "(", ")", "result", "=", "f", "(", "[", "test_values", "]", ")", "[", "0", "]", "expected", "=", "sigmoid", "(", "test_values", ")", "assert_allclose", "(", "result", ",", "expected", ",", "rtol", "=", "1e-05", ")" ]
test using a numerically stable reference sigmoid implementation .
train
false
25,217
def generate_tasks(**kwargs): tasks = _local_tasks() for (fun, args) in tasks: fun(*args) tasks.clear()
[ "def", "generate_tasks", "(", "**", "kwargs", ")", ":", "tasks", "=", "_local_tasks", "(", ")", "for", "(", "fun", ",", "args", ")", "in", "tasks", ":", "fun", "(", "*", "args", ")", "tasks", ".", "clear", "(", ")" ]
goes through thread local index update tasks set and generates celery tasks for all tasks in the set .
train
false
25,220
def get_cursor_position(fd=1): csbi = get_console_screen_buffer_info(fd=fd) coord = csbi.dwCursorPosition return (coord.X, coord.Y)
[ "def", "get_cursor_position", "(", "fd", "=", "1", ")", ":", "csbi", "=", "get_console_screen_buffer_info", "(", "fd", "=", "fd", ")", "coord", "=", "csbi", ".", "dwCursorPosition", "return", "(", "coord", ".", "X", ",", "coord", ".", "Y", ")" ]
gets the current cursor position as an tuple .
train
false
25,222
def get_local_facts_from_file(filename): local_facts = dict() try: ini_facts = configparser.SafeConfigParser() ini_facts.read(filename) for section in ini_facts.sections(): local_facts[section] = dict() for (key, value) in ini_facts.items(section): local_facts[section][key] = value except (configparser.MissingSectionHeaderError, configparser.ParsingError): try: with open(filename, 'r') as facts_file: local_facts = json.load(facts_file) except (ValueError, IOError): pass return local_facts
[ "def", "get_local_facts_from_file", "(", "filename", ")", ":", "local_facts", "=", "dict", "(", ")", "try", ":", "ini_facts", "=", "configparser", ".", "SafeConfigParser", "(", ")", "ini_facts", ".", "read", "(", "filename", ")", "for", "section", "in", "ini_facts", ".", "sections", "(", ")", ":", "local_facts", "[", "section", "]", "=", "dict", "(", ")", "for", "(", "key", ",", "value", ")", "in", "ini_facts", ".", "items", "(", "section", ")", ":", "local_facts", "[", "section", "]", "[", "key", "]", "=", "value", "except", "(", "configparser", ".", "MissingSectionHeaderError", ",", "configparser", ".", "ParsingError", ")", ":", "try", ":", "with", "open", "(", "filename", ",", "'r'", ")", "as", "facts_file", ":", "local_facts", "=", "json", ".", "load", "(", "facts_file", ")", "except", "(", "ValueError", ",", "IOError", ")", ":", "pass", "return", "local_facts" ]
retrieve local facts from fact file args: filename : local facts file returns: dict: the retrieved facts .
train
false
25,223
def miniEditImages(): return {'Select': BitmapImage(file='/usr/include/X11/bitmaps/left_ptr'), 'Switch': PhotoImage(data='\nR0lGODlhLgAgAPcAAB2ZxGq61imex4zH3RWWwmK41tzd3vn9/jCiyfX7/Q6SwFay0gBlmtnZ2snJ\nyr+2tAuMu6rY6D6kyfHx8XO/2Uqszjmly6DU5uXz+JLN4uz3+kSrzlKx0ZeZm2K21BuYw67a6QB9\nr+Xl5rW2uHW61On1+UGpzbrf6xiXwny9166vsMLCwgBdlAmHt8TFxgBwpNTs9C2hyO7t7ZnR5L/B\nw0yv0NXV1gBimKGjpABtoQBuoqKkpiaUvqWmqHbB2/j4+Pf39729vgB/sN7w9obH3hSMugCAsonJ\n4M/q8wBglgB6rCCaxLO0tX7C2wBqniGMuABzpuPl5f3+/v39/fr6+r7i7vP6/ABonV621LLc6zWk\nyrq6uq6wskGlyUaszp6gohmYw8HDxKaoqn3E3LGztWGuzcnLzKmrrOnp6gB1qCaex1q001ewz+Dg\n4QB3qrCxstHS09LR0dHR0s7Oz8zNzsfIyQaJuQB0pozL4YzI3re4uAGFtYDG3hOUwb+/wQB5rOvr\n6wB2qdju9TWfxgBpniOcxeLj48vn8dvc3VKuzwB2qp6fos/Q0aXV6D+jxwB7rsXHyLu8vb27vCSc\nxSGZwxyZxH3A2RuUv0+uzz+ozCedxgCDtABnnABroKutr/7+/n2/2LTd6wBvo9bX2OLo6lGv0C6d\nxS6avjmmzLTR2uzr6m651RuXw4jF3CqfxySaxSadyAuRv9bd4cPExRiMuDKjyUWevNPS0sXl8BeY\nxKytr8G/wABypXvC23vD3O73+3vE3cvU2PH5+7S1t7q7vCGVwO/v8JfM3zymyyyZwrWys+Hy90Ki\nxK6qqg+TwBKXxMvMzaWtsK7U4jemzLXEygBxpW++2aCho97Z18bP0/T09fX29vb19ViuzdDR0crf\n51qd01y00ujo6Onq6hCDs2Gpw3i71CqWv3S71nO92M/h52m207bJ0AN6rPPz9Nrh5Nvo7K/b6oTI\n37Td7ABqneHi4yScxo/M4RiWwRqVwcro8n3B2lGoylStzszMzAAAACH5BAEAAP8ALAAAAAAuACAA\nBwj/AP8JHEjw3wEkEY74WOjrQhUNBSNKnCjRSoYKCOwJcKWpEAACBFBRGEKxZMkDjRAg2OBlQyYL\nWhDEcOWxDwofv0zqHIhhDYIFC2p4MYFMS62ZaiYVWlJJAYIqO00KMlEjABYOQokaRbp0CYBKffpE\niDpxSKYC1gqswToUmYVaCFyp6QrgwwcCscaSJZhgQYBeAdRyqFBhgwWkGyct8WoXRZ8Ph/YOxMOB\nCIUAHsBxwGQBAII1YwpMI5Brcd0PKFA4Q2ZFMgYteZqkwxyu1KQNJzQc+CdFCrxypyqdRoEPX6x7\nki/n2TfbAxtNRHYTVCWpWTRbuRoX7yMgZ9QSFQa0/7LU/BXygjIWXVOBTR2sxp7BxGpENgKbY+PR\nreqyIOKnOh0M445AjTjDCgrPSBNFKt9w8wMVU5g0Bg8kDAAKOutQAkNEQNBwDRAEeVEcAV6w84Ay\nKowQSRhmzNGAASIAYow2IP6DySPk8ANKCv1wINE2cpjxCUEgOIOPAKicQMMbKnhyhhg97HDNF4vs\nIEYkNkzwjwSP/PHIE2VIgIdEnxjAiBwNGIKGDKS8I0sw2VAzApNOQimGLlyMAIkDw2yhZTF/KKGE\nlxCEMtEPBtDhACQurLDCLkFIsoUeZLyRpx8OmEGHN3AEcU0HkFAhUDFulDroJvOU5M44iDjgDTQO\n1P/hzRw2IFJPGw3AAY0LI/SAwxc7jEKQI2mkEUipRoxp0g821AMIGlG0McockMzihx5c1LkDDmSg\nUVAiafACRbGPVKDTFG3MYUYdLoThRxDE6DEMGUww8eQONGwTER9piFINFOPasaFJVIjTwC1xzOGP\nA3HUKoIMDTwJR4QRgdBOJzq8UM0Lj5QihU5ZdGMOCSSYUwYzAwwkDhNtUKTBOZ10koMOoohihDwm\nHZKPEDwb4fMe9An0g5Yl+SDKFTHnkMMLLQAjXUTxUCLEIyH0bIQAwuxVQhEMcEIIIUmHUEsWGCQg\nxQEaIFGAHV0+QnUIIWwyg2T/3MPLDQwwcAUhTjiswYsQl1SAxQKmbBJCIMe6ISjVmXwsWQKJEJJE\n3l1/TY8O4wZyh8ZQ3IF4qX9cggTdAmEwCAMs3IB311fsDfbMGv97BxSBQBAP6QMN0QUhLCSRhOp5\ne923zDpk/EIaRdyO+0C/eHBHEiz0vjrrfMfciSKD4LJ8RBEk88IN0ff+O/CEVEPLGK1tH1ECM7Dx\nRDWdcMLJFTpUQ44jfCyjvlShZNDE/0QAgT6ypr6AAAA7\n '), 'LegacySwitch': PhotoImage(data='\nR0lGODlhMgAYAPcAAAEBAXmDjbe4uAE5cjF7xwFWq2Sa0S9biSlrrdTW1k2Ly02a5xUvSQFHjmep\n6bfI2Q5SlQIYLwFfvj6M3Jaan8fHyDuFzwFp0Vah60uU3AEiRhFgrgFRogFr10N9uTFrpytHYQFM\nmGWt9wIwX+bm5kaT4gtFgR1cnJPF9yt80CF0yAIMGHmp2c/P0AEoUb/P4Fei7qK4zgpLjgFkyQlf\nt1mf5jKD1WWJrQ86ZwFAgBhYmVOa4MPV52uv8y+A0iR3ywFbtUyX5ECI0Q1UmwIcOUGQ3RBXoQI0\naRJbpr3BxVeJvQUJDafH5wIlS2aq7xBmv52lr7fH12el5Wml3097ph1ru7vM3HCz91Ke6lid40KQ\n4GSQvgQGClFnfwVJjszMzVCX3hljrdPT1AFLlBRnutPf6yd5zjeI2QE9eRBdrBNVl+3v70mV4ydf\nlwMVKwErVlul8AFChTGB1QE3bsTFxQImTVmAp0FjiUSM1k+b6QQvWQ1SlxMgLgFixEqU3xJhsgFT\npn2Xs5OluZ+1yz1Xb6HN+Td9wy1zuYClykV5r0x2oeDh4qmvt8LDwxhuxRlLfyRioo2124mft9bi\n71mDr7fT79nl8Z2hpQs9b7vN4QMQIOPj5XOPrU2Jx32z6xtvwzeBywFFikFnjwcPFa29yxJjuFmP\nxQFv3qGxwRc/Z8vb6wsRGBNqwqmpqTdvqQIbNQFPngMzZAEfP0mQ13mHlQFYsAFnznOXu2mPtQxj\nvQ1Vn4Ot1+/x8my0/CJgnxNNh8DT5CdJaWyx+AELFWmt8QxPkxBZpwMFB015pgFduGCNuyx7zdnZ\n2WKm6h1xyOPp8aW70QtPkUmM0LrCyr/FyztljwFPm0OJzwFny7/L1xFjswE/e12i50iR2VR8o2Gf\n3xszS2eTvz2BxSlloQdJiwMHDzF3u7bJ3T2I1WCp8+Xt80FokQFJklef6mORw2ap7SJ1y77Q47nN\n3wFfu1Kb5cXJyxdhrdDR0wlNkTSF11Oa4yp4yQEuW0WQ3QIDBQI7dSH5BAEAAAAALAAAAAAyABgA\nBwj/AAEIHDjKF6SDvhImPMHwhA6HOiLqUENRDYSLEIplxBcNHz4Z5GTI8BLKS5OBA1Ply2fDhxwf\nPlLITGFmmRkzP+DlVKHCmU9nnz45csSqKKsn9gileZKrVC4aRFACOGZu5UobNuRohRkzhc2b+36o\nqCaqrFmzZEV1ERBg3BOmMl5JZTBhwhm7ZyycYZnvJdeuNl21qkCHTiPDhxspTtKoQgUKCJ6wehMV\n5QctWupeo6TkjOd8e1lmdQkTGbTTMaDFiDGINeskX6YhEicUiQa5A/kUKaFFwQ0oXzjZ8Tbcm3Hj\nirwpMtTSgg9QMJf5WEZ9375AiED19ImpSQSUB4Kw/8HFSMyiRWJaqG/xhf2X91+oCbmq1e/MFD/2\nEcApVkWVJhp8J9AqsywQxDfAbLJJPAy+kMkL8shjxTkUnhOJZ5+JVp8cKfhwxwdf4fQLgG4MFAwW\nKOZRAxM81EAPPQvoE0QQfrDhx4399OMBMjz2yCMVivCoCAWXKLKMTPvoUYcsKwi0RCcwYCAlFjU0\nA6OBM4pXAhsl8FYELYWFWZhiZCbRQgIC2AGTLy408coxAoEDx5wwtGPALTVg0E4NKC7gp4FsBKoA\nKi8U+oIVmVih6DnZPMBMAlGwIARWOLiggSYC+ZNIOulwY4AkSZCyxaikbqHMqaeaIp4+rAaxQxBg\n2P+IozuRzvLZIS4syYVAfMAhwhSC1EPCGoskIIYY9yS7Hny75OFnEIAGyiVvWkjjRxF11fXIG3WU\nKNA6wghDTCW88PKMJZOkm24Z7LarSjPtoIjFn1lKyyVmmBVhwRtvaDDMgFL0Eu4VhaiDwhXCXNFD\nD8QQw7ATEDsBw8RSxotFHs7CKJ60XWrRBj91EOGPQCA48c7J7zTjSTPctOzynjVkkYU+O9S8Axg4\nZ6BzBt30003Ps+AhNB5C4PCGC5gKJMMTZJBRytOl/CH1HxvQkMbVVxujtdZGGKGL17rsEfYQe+xR\nzNnFcGQCv7LsKlAtp8R9Sgd0032BLXjPoPcMffTd3YcEgAMOxOBA1GJ4AYgXAMjiHDTgggveCgRI\n3RfcnffefgcOeDKEG3444osDwgEspMNiTQhx5FoOShxcrrfff0uQjOycD+554qFzMHrpp4cwBju/\n5+CmVNbArnntndeCO+O689777+w0IH0o1P/TRJMohRA4EJwn47nyiocOSOmkn/57COxE3wD11Mfh\nfg45zCGyVF4Ufvvyze8ewv5jQK9++6FwXxzglwM0GPAfR8AeSo4gwAHCbxsQNCAa/kHBAVhwAHPI\n4BE2eIRYeHAEIBwBP0Y4Qn41YWRSCQgAOw==\n '), 'LegacyRouter': PhotoImage(data='\nR0lGODlhMgAYAPcAAAEBAXZ8gQNAgL29vQNctjl/xVSa4j1dfCF+3QFq1DmL3wJMmAMzZZW11dnZ\n2SFrtyNdmTSO6gIZMUKa8gJVqEOHzR9Pf5W74wFjxgFx4jltn+np6Eyi+DuT6qKiohdtwwUPGWiq\n6ymF4LHH3Rh11CV81kKT5AMoUA9dq1ap/mV0gxdXlytRdR1ptRNPjTt9vwNgvwJZsX+69gsXJQFH\njTtjizF0tvHx8VOm9z2V736Dhz2N3QM2acPZ70qe8gFo0HS19wVRnTiR6hMpP0eP1i6J5iNlqAtg\ntktjfQFu3TNxryx4xAMTIzOE1XqAh1uf5SWC4AcfNy1XgQJny93n8a2trRh312Gt+VGm/AQIDTmB\nyAF37QJasydzvxM/ayF3zhdLf8zLywFdu4i56gFlyi2J4yV/1w8wUo2/8j+X8D2Q5Eee9jeR7Uia\n7DpeggFt2QNPm97e3jRong9bpziH2DuT7aipqQoVICmG45vI9R5720eT4Q1hs1er/yVVhwJJktPh\n70tfdbHP7Xev5xs5V7W1sz9jhz11rUVZcQ9WoCVVhQk7cRdtwWuw9QYOFyFHbSBnr0dznxtWkS18\nzKfP9wwcLAMHCwFFiS5UeqGtuRNNiwMfPS1hlQMtWRE5XzGM5yhxusLCwCljnwMdOFWh7cve8pG/\n7Tlxp+Tr8g9bpXF3f0lheStrrYu13QEXLS1ppTV3uUuR1RMjNTF3vU2X4TZupwRSolNne4nB+T+L\n2YGz4zJ/zYe99YGHjRdDcT95sx09XQldsgMLEwMrVc/X3yN3yQ1JhTRbggsdMQNfu9HPz6WlpW2t\n7RctQ0GFyeHh4dvl8SBZklCb5kOO2kWR3Vmt/zdjkQIQHi90uvPz8wIVKBp42SV5zbfT7wtXpStV\nfwFWrBVvyTt3swFz5kGBv2+1/QlbrVFjdQM7d1+j54i67UmX51qn9i1vsy+D2TuR5zddhQsjOR1t\nu0GV6ghbsDVZf4+76RRisent8Xd9hQFBgwFNmwJLlcPDwwFr1z2T5yH5BAEAAAAALAAAAAAyABgA\nBwj/AAEIHEiQYJY7Qwg9UsTplRIbENuxEiXJgpcz8e5YKsixY8Essh7JcbbOBwcOa1JOmJAmTY4c\nHeoIabJrCShI0XyB8YRso0eOjoAdWpciBZajJ1GuWcnSZY46Ed5N8hPATqEBoRB9gVJsxRlhPwHI\n0kDkVywcRpGe9LF0adOnMpt8CxDnxg1o9lphKoEACoIvmlxxvHOKVg0n/Tzku2WoVoU2J1P6WNkS\nrtwADuxCG/MOjwgRUEIjGG3FhaOBzaThiDSCil27G8Isc3LLjZwXsA6YYJmDjhTMmseoKQIFDx7R\noxHo2abnwygAlUj1mV6tWjlelEpRwfd6gzI7VeJQ/2vZoVaDUqigqftXpH0R46H9Kl++zUo4JnKq\n9dGvv09RHFhcIUMe0NiFDyql0OJUHWywMc87TXRhhCRGiHAccvNZUR8JxpDTH38p9HEUFhxgMSAv\njbBjQge8PSXEC6uo0IsHA6gAAShmgCbffNtsQwIJifhRHX/TpUUiSijlUk8AqgQixSwdNBjCa7CF\noVggmEgCyRf01WcFCYvYUgB104k4YlK5HONEXXfpokYdMrXRAzMhmNINNNzB9p0T57AgyZckpKKP\nGFNgw06ZWKR10jTw6MAmFWj4AJcQQkQQwSefvFeGCemMIQggeaJywSQ/wgHOAmJskQEfWqBlFBEH\n1P/QaGY3QOpDZXA2+A6m7hl3IRQKGDCIAj6iwE8yGKC6xbJv8IHNHgACQQybN2QiTi5NwdlBpZdi\nisd7vyanByOJ7CMGGRhgwE+qyy47DhnBPLDLEzLIAEQjBtChRmVPNWgpr+Be+Nc9icARww9TkIEu\nDAsQ0O7DzGIQzD2QdDEJHTsIAROc3F7qWQncyHPPHN5QQAAG/vjzw8oKp8sPPxDH3O44/kwBQzLB\nxBCMOTzzHEMMBMBARgJvZJBBEm/4k0ACKydMBgwYoKNNEjJXbTXE42Q9jtFIp8z0Dy1jQMA1AGzi\nz9VoW7310V0znYDTGMQgwUDXLDBO2nhvoTXbbyRk/XXL+pxWkAT8UJ331WsbnbTSK8MggDZhCTOM\nLQkcjvXeSPedAAw0nABWWARZIgEDfyTzxt15Z53BG1PEcEknrvgEelhZMDHKCTwI8EcQFHBBAAFc\ngGPLHwLwcMIo12Qxu0ABAQA7\n '), 'Controller': PhotoImage(data='\n R0lGODlhMAAwAPcAAAEBAWfNAYWFhcfHx+3t6/f390lJUaWlpfPz8/Hx72lpaZGRke/v77m5uc0B\n AeHh4e/v7WNjY3t7e5eXlyMjI4mJidPT0+3t7f///09PT7Ozs/X19fHx8ZWTk8HBwX9/fwAAAAAA\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAACH5BAEAAAAALAAAAAAwADAA\n Bwj/AAEIHEiwoMGDCBMqXMiwocOHECNKnEixosWLGAEIeMCxo8ePHwVkBGABg8mTKFOmtDByAIYN\n MGPCRCCzQIENNzEMGOkBAwIKQIMKpYCgKAIHCDB4GNkAA4OnUJ9++CDhQ1QGFzA0GKkBA4GvYMOK\n BYtBA1cNaNOqXcuWq8q3b81m7Cqdbk2bMMu6/Tl0qFEEAZLKxdj1KlSqVA3rnet1rOOwiwmznUzZ\n LdzLJgdfpIv3pmebN2Pm1GyRbocNp1PLNMDaAM3Im1/alQk4gO28pCt2RdCBt+/eRg8IP1AUdmmf\n f5MrL56bYlcOvaP7Xo6Ag3HdGDho3869u/YE1507t+3AgLz58ujPMwg/sTBUCAzgy49PH0LW5u0x\n XFiwvz////5dcJ9bjxVIAHsSdUXAAgs2yOCDDn6FYEQaFGDgYxNCpEFfHHKIX4IDhCjiiCSS+CGF\n FlCmogYpcnVABTDGKGOMAlRQYwUHnKjhAjX2aOOPN8LImgAL6PiQBhLMqCSNAThQgQRGOqRBBD1W\n aaOVAggnQARRNqRBBxmEKeaYZIrZQZcMKbDiigqM5OabcMYp55x01ilnQAA7\n '), 'Host': PhotoImage(data='\n R0lGODlhIAAYAPcAMf//////zP//mf//Zv//M///AP/M///MzP/M\n mf/MZv/MM//MAP+Z//+ZzP+Zmf+ZZv+ZM/+ZAP9m//9mzP9mmf9m\n Zv9mM/9mAP8z//8zzP8zmf8zZv8zM/8zAP8A//8AzP8Amf8AZv8A\n M/8AAMz//8z/zMz/mcz/Zsz/M8z/AMzM/8zMzMzMmczMZszMM8zM\n AMyZ/8yZzMyZmcyZZsyZM8yZAMxm/8xmzMxmmcxmZsxmM8xmAMwz\n /8wzzMwzmcwzZswzM8wzAMwA/8wAzMwAmcwAZswAM8wAAJn//5n/\n zJn/mZn/Zpn/M5n/AJnM/5nMzJnMmZnMZpnMM5nMAJmZ/5mZzJmZ\n mZmZZpmZM5mZAJlm/5lmzJlmmZlmZplmM5lmAJkz/5kzzJkzmZkz\n ZpkzM5kzAJkA/5kAzJkAmZkAZpkAM5kAAGb//2b/zGb/mWb/Zmb/\n M2b/AGbM/2bMzGbMmWbMZmbMM2bMAGaZ/2aZzGaZmWaZZmaZM2aZ\n AGZm/2ZmzGZmmWZmZmZmM2ZmAGYz/2YzzGYzmWYzZmYzM2YzAGYA\n /2YAzGYAmWYAZmYAM2YAADP//zP/zDP/mTP/ZjP/MzP/ADPM/zPM\n zDPMmTPMZjPMMzPMADOZ/zOZzDOZmTOZZjOZMzOZADNm/zNmzDNm\n mTNmZjNmMzNmADMz/zMzzDMzmTMzZjMzMzMzADMA/zMAzDMAmTMA\n ZjMAMzMAAAD//wD/zAD/mQD/ZgD/MwD/AADM/wDMzADMmQDMZgDM\n MwDMAACZ/wCZzACZmQCZZgCZMwCZAABm/wBmzABmmQBmZgBmMwBm\n AAAz/wAzzAAzmQAzZgAzMwAzAAAA/wAAzAAAmQAAZgAAM+4AAN0A\n ALsAAKoAAIgAAHcAAFUAAEQAACIAABEAAADuAADdAAC7AACqAACI\n AAB3AABVAABEAAAiAAARAAAA7gAA3QAAuwAAqgAAiAAAdwAAVQAA\n RAAAIgAAEe7u7t3d3bu7u6qqqoiIiHd3d1VVVURERCIiIhEREQAA\n ACH5BAEAAAAALAAAAAAgABgAAAiNAAH8G0iwoMGDCAcKTMiw4UBw\n BPXVm0ixosWLFvVBHFjPoUeC9Tb+6/jRY0iQ/8iVbHiS40CVKxG2\n HEkQZsyCM0mmvGkw50uePUV2tEnOZkyfQA8iTYpTKNOgKJ+C3AhO\n p9SWVaVOfWj1KdauTL9q5UgVbFKsEjGqXVtP40NwcBnCjXtw7tx/\n C8cSBBAQADs=\n '), 'OldSwitch': PhotoImage(data='\n R0lGODlhIAAYAPcAMf//////zP//mf//Zv//M///AP/M///MzP/M\n mf/MZv/MM//MAP+Z//+ZzP+Zmf+ZZv+ZM/+ZAP9m//9mzP9mmf9m\n Zv9mM/9mAP8z//8zzP8zmf8zZv8zM/8zAP8A//8AzP8Amf8AZv8A\n M/8AAMz//8z/zMz/mcz/Zsz/M8z/AMzM/8zMzMzMmczMZszMM8zM\n AMyZ/8yZzMyZmcyZZsyZM8yZAMxm/8xmzMxmmcxmZsxmM8xmAMwz\n /8wzzMwzmcwzZswzM8wzAMwA/8wAzMwAmcwAZswAM8wAAJn//5n/\n zJn/mZn/Zpn/M5n/AJnM/5nMzJnMmZnMZpnMM5nMAJmZ/5mZzJmZ\n mZmZZpmZM5mZAJlm/5lmzJlmmZlmZplmM5lmAJkz/5kzzJkzmZkz\n ZpkzM5kzAJkA/5kAzJkAmZkAZpkAM5kAAGb//2b/zGb/mWb/Zmb/\n M2b/AGbM/2bMzGbMmWbMZmbMM2bMAGaZ/2aZzGaZmWaZZmaZM2aZ\n AGZm/2ZmzGZmmWZmZmZmM2ZmAGYz/2YzzGYzmWYzZmYzM2YzAGYA\n /2YAzGYAmWYAZmYAM2YAADP//zP/zDP/mTP/ZjP/MzP/ADPM/zPM\n zDPMmTPMZjPMMzPMADOZ/zOZzDOZmTOZZjOZMzOZADNm/zNmzDNm\n mTNmZjNmMzNmADMz/zMzzDMzmTMzZjMzMzMzADMA/zMAzDMAmTMA\n ZjMAMzMAAAD//wD/zAD/mQD/ZgD/MwD/AADM/wDMzADMmQDMZgDM\n MwDMAACZ/wCZzACZmQCZZgCZMwCZAABm/wBmzABmmQBmZgBmMwBm\n AAAz/wAzzAAzmQAzZgAzMwAzAAAA/wAAzAAAmQAAZgAAM+4AAN0A\n ALsAAKoAAIgAAHcAAFUAAEQAACIAABEAAADuAADdAAC7AACqAACI\n AAB3AABVAABEAAAiAAARAAAA7gAA3QAAuwAAqgAAiAAAdwAAVQAA\n RAAAIgAAEe7u7t3d3bu7u6qqqoiIiHd3d1VVVURERCIiIhEREQAA\n ACH5BAEAAAAALAAAAAAgABgAAAhwAAEIHEiwoMGDCBMqXMiwocOH\n ECNKnEixosWB3zJq3Mixo0eNAL7xG0mypMmTKPl9Cznyn8uWL/m5\n /AeTpsyYI1eKlBnO5r+eLYHy9Ck0J8ubPmPOrMmUpM6UUKMa/Ui1\n 6saLWLNq3cq1q9evYB0GBAA7\n '), 'NetLink': PhotoImage(data='\n R0lGODlhFgAWAPcAMf//////zP//mf//Zv//M///AP/M///MzP/M\n mf/MZv/MM//MAP+Z//+ZzP+Zmf+ZZv+ZM/+ZAP9m//9mzP9mmf9m\n Zv9mM/9mAP8z//8zzP8zmf8zZv8zM/8zAP8A//8AzP8Amf8AZv8A\n M/8AAMz//8z/zMz/mcz/Zsz/M8z/AMzM/8zMzMzMmczMZszMM8zM\n AMyZ/8yZzMyZmcyZZsyZM8yZAMxm/8xmzMxmmcxmZsxmM8xmAMwz\n /8wzzMwzmcwzZswzM8wzAMwA/8wAzMwAmcwAZswAM8wAAJn//5n/\n zJn/mZn/Zpn/M5n/AJnM/5nMzJnMmZnMZpnMM5nMAJmZ/5mZzJmZ\n mZmZZpmZM5mZAJlm/5lmzJlmmZlmZplmM5lmAJkz/5kzzJkzmZkz\n ZpkzM5kzAJkA/5kAzJkAmZkAZpkAM5kAAGb//2b/zGb/mWb/Zmb/\n M2b/AGbM/2bMzGbMmWbMZmbMM2bMAGaZ/2aZzGaZmWaZZmaZM2aZ\n AGZm/2ZmzGZmmWZmZmZmM2ZmAGYz/2YzzGYzmWYzZmYzM2YzAGYA\n /2YAzGYAmWYAZmYAM2YAADP//zP/zDP/mTP/ZjP/MzP/ADPM/zPM\n zDPMmTPMZjPMMzPMADOZ/zOZzDOZmTOZZjOZMzOZADNm/zNmzDNm\n mTNmZjNmMzNmADMz/zMzzDMzmTMzZjMzMzMzADMA/zMAzDMAmTMA\n ZjMAMzMAAAD//wD/zAD/mQD/ZgD/MwD/AADM/wDMzADMmQDMZgDM\n MwDMAACZ/wCZzACZmQCZZgCZMwCZAABm/wBmzABmmQBmZgBmMwBm\n AAAz/wAzzAAzmQAzZgAzMwAzAAAA/wAAzAAAmQAAZgAAM+4AAN0A\n ALsAAKoAAIgAAHcAAFUAAEQAACIAABEAAADuAADdAAC7AACqAACI\n AAB3AABVAABEAAAiAAARAAAA7gAA3QAAuwAAqgAAiAAAdwAAVQAA\n RAAAIgAAEe7u7t3d3bu7u6qqqoiIiHd3d1VVVURERCIiIhEREQAA\n ACH5BAEAAAAALAAAAAAWABYAAAhIAAEIHEiwoEGBrhIeXEgwoUKG\n Cx0+hGhQoiuKBy1irChxY0GNHgeCDAlgZEiTHlFuVImRJUWXEGEy\n lBmxI8mSNknm1Dnx5sCAADs=\n ')}
[ "def", "miniEditImages", "(", ")", ":", "return", "{", "'Select'", ":", "BitmapImage", "(", "file", "=", "'/usr/include/X11/bitmaps/left_ptr'", ")", ",", "'Switch'", ":", "PhotoImage", "(", "data", "=", "'\\nR0lGODlhLgAgAPcAAB2ZxGq61imex4zH3RWWwmK41tzd3vn9/jCiyfX7/Q6SwFay0gBlmtnZ2snJ\\nyr+2tAuMu6rY6D6kyfHx8XO/2Uqszjmly6DU5uXz+JLN4uz3+kSrzlKx0ZeZm2K21BuYw67a6QB9\\nr+Xl5rW2uHW61On1+UGpzbrf6xiXwny9166vsMLCwgBdlAmHt8TFxgBwpNTs9C2hyO7t7ZnR5L/B\\nw0yv0NXV1gBimKGjpABtoQBuoqKkpiaUvqWmqHbB2/j4+Pf39729vgB/sN7w9obH3hSMugCAsonJ\\n4M/q8wBglgB6rCCaxLO0tX7C2wBqniGMuABzpuPl5f3+/v39/fr6+r7i7vP6/ABonV621LLc6zWk\\nyrq6uq6wskGlyUaszp6gohmYw8HDxKaoqn3E3LGztWGuzcnLzKmrrOnp6gB1qCaex1q001ewz+Dg\\n4QB3qrCxstHS09LR0dHR0s7Oz8zNzsfIyQaJuQB0pozL4YzI3re4uAGFtYDG3hOUwb+/wQB5rOvr\\n6wB2qdju9TWfxgBpniOcxeLj48vn8dvc3VKuzwB2qp6fos/Q0aXV6D+jxwB7rsXHyLu8vb27vCSc\\nxSGZwxyZxH3A2RuUv0+uzz+ozCedxgCDtABnnABroKutr/7+/n2/2LTd6wBvo9bX2OLo6lGv0C6d\\nxS6avjmmzLTR2uzr6m651RuXw4jF3CqfxySaxSadyAuRv9bd4cPExRiMuDKjyUWevNPS0sXl8BeY\\nxKytr8G/wABypXvC23vD3O73+3vE3cvU2PH5+7S1t7q7vCGVwO/v8JfM3zymyyyZwrWys+Hy90Ki\\nxK6qqg+TwBKXxMvMzaWtsK7U4jemzLXEygBxpW++2aCho97Z18bP0/T09fX29vb19ViuzdDR0crf\\n51qd01y00ujo6Onq6hCDs2Gpw3i71CqWv3S71nO92M/h52m207bJ0AN6rPPz9Nrh5Nvo7K/b6oTI\\n37Td7ABqneHi4yScxo/M4RiWwRqVwcro8n3B2lGoylStzszMzAAAACH5BAEAAP8ALAAAAAAuACAA\\nBwj/AP8JHEjw3wEkEY74WOjrQhUNBSNKnCjRSoYKCOwJcKWpEAACBFBRGEKxZMkDjRAg2OBlQyYL\\nWhDEcOWxDwofv0zqHIhhDYIFC2p4MYFMS62ZaiYVWlJJAYIqO00KMlEjABYOQokaRbp0CYBKffpE\\niDpxSKYC1gqswToUmYVaCFyp6QrgwwcCscaSJZhgQYBeAdRyqFBhgwWkGyct8WoXRZ8Ph/YOxMOB\\nCIUAHsBxwGQBAII1YwpMI5Brcd0PKFA4Q2ZFMgYteZqkwxyu1KQNJzQc+CdFCrxypyqdRoEPX6x7\\nki/n2TfbAxtNRHYTVCWpWTRbuRoX7yMgZ9QSFQa0/7LU/BXygjIWXVOBTR2sxp7BxGpENgKbY+PR\\nreqyIOKnOh0M445AjTjDCgrPSBNFKt9w8wMVU5g0Bg8kDAAKOutQAkNEQNBwDRAEeVEcAV6w84Ay\\nKowQSRhmzNGAASIAYow2IP6DySPk8ANKCv1wINE2cpjxCUEgOIOPAKicQMMbKnhyhhg97HDNF4vs\\nIEYkNkzwjwSP/PHIE2VIgIdEnxjAiBwNGIKGDKS8I0sw2VAzApNOQimGLlyMAIkDw2yhZTF/KKGE\\nlxCEMtEPBtDhACQurLDCLkFIsoUeZLyRpx8OmEGHN3AEcU0HkFAhUDFulDroJvOU5M44iDjgDTQO\\n1P/hzRw2IFJPGw3AAY0LI/SAwxc7jEKQI2mkEUipRoxp0g821AMIGlG0McockMzihx5c1LkDDmSg\\nUVAiafACRbGPVKDTFG3MYUYdLoThRxDE6DEMGUww8eQONGwTER9piFINFOPasaFJVIjTwC1xzOGP\\nA3HUKoIMDTwJR4QRgdBOJzq8UM0Lj5QihU5ZdGMOCSSYUwYzAwwkDhNtUKTBOZ10koMOoohihDwm\\nHZKPEDwb4fMe9An0g5Yl+SDKFTHnkMMLLQAjXUTxUCLEIyH0bIQAwuxVQhEMcEIIIUmHUEsWGCQg\\nxQEaIFGAHV0+QnUIIWwyg2T/3MPLDQwwcAUhTjiswYsQl1SAxQKmbBJCIMe6ISjVmXwsWQKJEJJE\\n3l1/TY8O4wZyh8ZQ3IF4qX9cggTdAmEwCAMs3IB311fsDfbMGv97BxSBQBAP6QMN0QUhLCSRhOp5\\ne923zDpk/EIaRdyO+0C/eHBHEiz0vjrrfMfciSKD4LJ8RBEk88IN0ff+O/CEVEPLGK1tH1ECM7Dx\\nRDWdcMLJFTpUQ44jfCyjvlShZNDE/0QAgT6ypr6AAAA7\\n '", ")", ",", "'LegacySwitch'", ":", "PhotoImage", "(", "data", "=", "'\\nR0lGODlhMgAYAPcAAAEBAXmDjbe4uAE5cjF7xwFWq2Sa0S9biSlrrdTW1k2Ly02a5xUvSQFHjmep\\n6bfI2Q5SlQIYLwFfvj6M3Jaan8fHyDuFzwFp0Vah60uU3AEiRhFgrgFRogFr10N9uTFrpytHYQFM\\nmGWt9wIwX+bm5kaT4gtFgR1cnJPF9yt80CF0yAIMGHmp2c/P0AEoUb/P4Fei7qK4zgpLjgFkyQlf\\nt1mf5jKD1WWJrQ86ZwFAgBhYmVOa4MPV52uv8y+A0iR3ywFbtUyX5ECI0Q1UmwIcOUGQ3RBXoQI0\\naRJbpr3BxVeJvQUJDafH5wIlS2aq7xBmv52lr7fH12el5Wml3097ph1ru7vM3HCz91Ke6lid40KQ\\n4GSQvgQGClFnfwVJjszMzVCX3hljrdPT1AFLlBRnutPf6yd5zjeI2QE9eRBdrBNVl+3v70mV4ydf\\nlwMVKwErVlul8AFChTGB1QE3bsTFxQImTVmAp0FjiUSM1k+b6QQvWQ1SlxMgLgFixEqU3xJhsgFT\\npn2Xs5OluZ+1yz1Xb6HN+Td9wy1zuYClykV5r0x2oeDh4qmvt8LDwxhuxRlLfyRioo2124mft9bi\\n71mDr7fT79nl8Z2hpQs9b7vN4QMQIOPj5XOPrU2Jx32z6xtvwzeBywFFikFnjwcPFa29yxJjuFmP\\nxQFv3qGxwRc/Z8vb6wsRGBNqwqmpqTdvqQIbNQFPngMzZAEfP0mQ13mHlQFYsAFnznOXu2mPtQxj\\nvQ1Vn4Ot1+/x8my0/CJgnxNNh8DT5CdJaWyx+AELFWmt8QxPkxBZpwMFB015pgFduGCNuyx7zdnZ\\n2WKm6h1xyOPp8aW70QtPkUmM0LrCyr/FyztljwFPm0OJzwFny7/L1xFjswE/e12i50iR2VR8o2Gf\\n3xszS2eTvz2BxSlloQdJiwMHDzF3u7bJ3T2I1WCp8+Xt80FokQFJklef6mORw2ap7SJ1y77Q47nN\\n3wFfu1Kb5cXJyxdhrdDR0wlNkTSF11Oa4yp4yQEuW0WQ3QIDBQI7dSH5BAEAAAAALAAAAAAyABgA\\nBwj/AAEIHDjKF6SDvhImPMHwhA6HOiLqUENRDYSLEIplxBcNHz4Z5GTI8BLKS5OBA1Ply2fDhxwf\\nPlLITGFmmRkzP+DlVKHCmU9nnz45csSqKKsn9gileZKrVC4aRFACOGZu5UobNuRohRkzhc2b+36o\\nqCaqrFmzZEV1ERBg3BOmMl5JZTBhwhm7ZyycYZnvJdeuNl21qkCHTiPDhxspTtKoQgUKCJ6wehMV\\n5QctWupeo6TkjOd8e1lmdQkTGbTTMaDFiDGINeskX6YhEicUiQa5A/kUKaFFwQ0oXzjZ8Tbcm3Hj\\nirwpMtTSgg9QMJf5WEZ9375AiED19ImpSQSUB4Kw/8HFSMyiRWJaqG/xhf2X91+oCbmq1e/MFD/2\\nEcApVkWVJhp8J9AqsywQxDfAbLJJPAy+kMkL8shjxTkUnhOJZ5+JVp8cKfhwxwdf4fQLgG4MFAwW\\nKOZRAxM81EAPPQvoE0QQfrDhx4399OMBMjz2yCMVivCoCAWXKLKMTPvoUYcsKwi0RCcwYCAlFjU0\\nA6OBM4pXAhsl8FYELYWFWZhiZCbRQgIC2AGTLy408coxAoEDx5wwtGPALTVg0E4NKC7gp4FsBKoA\\nKi8U+oIVmVih6DnZPMBMAlGwIARWOLiggSYC+ZNIOulwY4AkSZCyxaikbqHMqaeaIp4+rAaxQxBg\\n2P+IozuRzvLZIS4syYVAfMAhwhSC1EPCGoskIIYY9yS7Hny75OFnEIAGyiVvWkjjRxF11fXIG3WU\\nKNA6wghDTCW88PKMJZOkm24Z7LarSjPtoIjFn1lKyyVmmBVhwRtvaDDMgFL0Eu4VhaiDwhXCXNFD\\nD8QQw7ATEDsBw8RSxotFHs7CKJ60XWrRBj91EOGPQCA48c7J7zTjSTPctOzynjVkkYU+O9S8Axg4\\nZ6BzBt30003Ps+AhNB5C4PCGC5gKJMMTZJBRytOl/CH1HxvQkMbVVxujtdZGGKGL17rsEfYQe+xR\\nzNnFcGQCv7LsKlAtp8R9Sgd0032BLXjPoPcMffTd3YcEgAMOxOBA1GJ4AYgXAMjiHDTgggveCgRI\\n3RfcnffefgcOeDKEG3444osDwgEspMNiTQhx5FoOShxcrrfff0uQjOycD+554qFzMHrpp4cwBju/\\n5+CmVNbArnntndeCO+O689777+w0IH0o1P/TRJMohRA4EJwn47nyiocOSOmkn/57COxE3wD11Mfh\\nfg45zCGyVF4Ufvvyze8ewv5jQK9++6FwXxzglwM0GPAfR8AeSo4gwAHCbxsQNCAa/kHBAVhwAHPI\\n4BE2eIRYeHAEIBwBP0Y4Qn41YWRSCQgAOw==\\n '", ")", ",", "'LegacyRouter'", ":", "PhotoImage", "(", "data", "=", "'\\nR0lGODlhMgAYAPcAAAEBAXZ8gQNAgL29vQNctjl/xVSa4j1dfCF+3QFq1DmL3wJMmAMzZZW11dnZ\\n2SFrtyNdmTSO6gIZMUKa8gJVqEOHzR9Pf5W74wFjxgFx4jltn+np6Eyi+DuT6qKiohdtwwUPGWiq\\n6ymF4LHH3Rh11CV81kKT5AMoUA9dq1ap/mV0gxdXlytRdR1ptRNPjTt9vwNgvwJZsX+69gsXJQFH\\njTtjizF0tvHx8VOm9z2V736Dhz2N3QM2acPZ70qe8gFo0HS19wVRnTiR6hMpP0eP1i6J5iNlqAtg\\ntktjfQFu3TNxryx4xAMTIzOE1XqAh1uf5SWC4AcfNy1XgQJny93n8a2trRh312Gt+VGm/AQIDTmB\\nyAF37QJasydzvxM/ayF3zhdLf8zLywFdu4i56gFlyi2J4yV/1w8wUo2/8j+X8D2Q5Eee9jeR7Uia\\n7DpeggFt2QNPm97e3jRong9bpziH2DuT7aipqQoVICmG45vI9R5720eT4Q1hs1er/yVVhwJJktPh\\n70tfdbHP7Xev5xs5V7W1sz9jhz11rUVZcQ9WoCVVhQk7cRdtwWuw9QYOFyFHbSBnr0dznxtWkS18\\nzKfP9wwcLAMHCwFFiS5UeqGtuRNNiwMfPS1hlQMtWRE5XzGM5yhxusLCwCljnwMdOFWh7cve8pG/\\n7Tlxp+Tr8g9bpXF3f0lheStrrYu13QEXLS1ppTV3uUuR1RMjNTF3vU2X4TZupwRSolNne4nB+T+L\\n2YGz4zJ/zYe99YGHjRdDcT95sx09XQldsgMLEwMrVc/X3yN3yQ1JhTRbggsdMQNfu9HPz6WlpW2t\\n7RctQ0GFyeHh4dvl8SBZklCb5kOO2kWR3Vmt/zdjkQIQHi90uvPz8wIVKBp42SV5zbfT7wtXpStV\\nfwFWrBVvyTt3swFz5kGBv2+1/QlbrVFjdQM7d1+j54i67UmX51qn9i1vsy+D2TuR5zddhQsjOR1t\\nu0GV6ghbsDVZf4+76RRisent8Xd9hQFBgwFNmwJLlcPDwwFr1z2T5yH5BAEAAAAALAAAAAAyABgA\\nBwj/AAEIHEiQYJY7Qwg9UsTplRIbENuxEiXJgpcz8e5YKsixY8Essh7JcbbOBwcOa1JOmJAmTY4c\\nHeoIabJrCShI0XyB8YRso0eOjoAdWpciBZajJ1GuWcnSZY46Ed5N8hPATqEBoRB9gVJsxRlhPwHI\\n0kDkVywcRpGe9LF0adOnMpt8CxDnxg1o9lphKoEACoIvmlxxvHOKVg0n/Tzku2WoVoU2J1P6WNkS\\nrtwADuxCG/MOjwgRUEIjGG3FhaOBzaThiDSCil27G8Isc3LLjZwXsA6YYJmDjhTMmseoKQIFDx7R\\noxHo2abnwygAlUj1mV6tWjlelEpRwfd6gzI7VeJQ/2vZoVaDUqigqftXpH0R46H9Kl++zUo4JnKq\\n9dGvv09RHFhcIUMe0NiFDyql0OJUHWywMc87TXRhhCRGiHAccvNZUR8JxpDTH38p9HEUFhxgMSAv\\njbBjQge8PSXEC6uo0IsHA6gAAShmgCbffNtsQwIJifhRHX/TpUUiSijlUk8AqgQixSwdNBjCa7CF\\noVggmEgCyRf01WcFCYvYUgB104k4YlK5HONEXXfpokYdMrXRAzMhmNINNNzB9p0T57AgyZckpKKP\\nGFNgw06ZWKR10jTw6MAmFWj4AJcQQkQQwSefvFeGCemMIQggeaJywSQ/wgHOAmJskQEfWqBlFBEH\\n1P/QaGY3QOpDZXA2+A6m7hl3IRQKGDCIAj6iwE8yGKC6xbJv8IHNHgACQQybN2QiTi5NwdlBpZdi\\nisd7vyanByOJ7CMGGRhgwE+qyy47DhnBPLDLEzLIAEQjBtChRmVPNWgpr+Be+Nc9icARww9TkIEu\\nDAsQ0O7DzGIQzD2QdDEJHTsIAROc3F7qWQncyHPPHN5QQAAG/vjzw8oKp8sPPxDH3O44/kwBQzLB\\nxBCMOTzzHEMMBMBARgJvZJBBEm/4k0ACKydMBgwYoKNNEjJXbTXE42Q9jtFIp8z0Dy1jQMA1AGzi\\nz9VoW7310V0znYDTGMQgwUDXLDBO2nhvoTXbbyRk/XXL+pxWkAT8UJ331WsbnbTSK8MggDZhCTOM\\nLQkcjvXeSPedAAw0nABWWARZIgEDfyTzxt15Z53BG1PEcEknrvgEelhZMDHKCTwI8EcQFHBBAAFc\\ngGPLHwLwcMIo12Qxu0ABAQA7\\n '", ")", ",", "'Controller'", ":", "PhotoImage", "(", "data", "=", "'\\n R0lGODlhMAAwAPcAAAEBAWfNAYWFhcfHx+3t6/f390lJUaWlpfPz8/Hx72lpaZGRke/v77m5uc0B\\n AeHh4e/v7WNjY3t7e5eXlyMjI4mJidPT0+3t7f///09PT7Ozs/X19fHx8ZWTk8HBwX9/fwAAAAAA\\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA\\n AAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAACH5BAEAAAAALAAAAAAwADAA\\n Bwj/AAEIHEiwoMGDCBMqXMiwocOHECNKnEixosWLGAEIeMCxo8ePHwVkBGABg8mTKFOmtDByAIYN\\n MGPCRCCzQIENNzEMGOkBAwIKQIMKpYCgKAIHCDB4GNkAA4OnUJ9++CDhQ1QGFzA0GKkBA4GvYMOK\\n BYtBA1cNaNOqXcuWq8q3b81m7Cqdbk2bMMu6/Tl0qFEEAZLKxdj1KlSqVA3rnet1rOOwiwmznUzZ\\n LdzLJgdfpIv3pmebN2Pm1GyRbocNp1PLNMDaAM3Im1/alQk4gO28pCt2RdCBt+/eRg8IP1AUdmmf\\n f5MrL56bYlcOvaP7Xo6Ag3HdGDho3869u/YE1507t+3AgLz58ujPMwg/sTBUCAzgy49PH0LW5u0x\\n XFiwvz////5dcJ9bjxVIAHsSdUXAAgs2yOCDDn6FYEQaFGDgYxNCpEFfHHKIX4IDhCjiiCSS+CGF\\n FlCmogYpcnVABTDGKGOMAlRQYwUHnKjhAjX2aOOPN8LImgAL6PiQBhLMqCSNAThQgQRGOqRBBD1W\\n aaOVAggnQARRNqRBBxmEKeaYZIrZQZcMKbDiigqM5OabcMYp55x01ilnQAA7\\n '", ")", ",", "'Host'", ":", "PhotoImage", "(", "data", "=", "'\\n R0lGODlhIAAYAPcAMf//////zP//mf//Zv//M///AP/M///MzP/M\\n mf/MZv/MM//MAP+Z//+ZzP+Zmf+ZZv+ZM/+ZAP9m//9mzP9mmf9m\\n Zv9mM/9mAP8z//8zzP8zmf8zZv8zM/8zAP8A//8AzP8Amf8AZv8A\\n M/8AAMz//8z/zMz/mcz/Zsz/M8z/AMzM/8zMzMzMmczMZszMM8zM\\n AMyZ/8yZzMyZmcyZZsyZM8yZAMxm/8xmzMxmmcxmZsxmM8xmAMwz\\n /8wzzMwzmcwzZswzM8wzAMwA/8wAzMwAmcwAZswAM8wAAJn//5n/\\n zJn/mZn/Zpn/M5n/AJnM/5nMzJnMmZnMZpnMM5nMAJmZ/5mZzJmZ\\n mZmZZpmZM5mZAJlm/5lmzJlmmZlmZplmM5lmAJkz/5kzzJkzmZkz\\n ZpkzM5kzAJkA/5kAzJkAmZkAZpkAM5kAAGb//2b/zGb/mWb/Zmb/\\n M2b/AGbM/2bMzGbMmWbMZmbMM2bMAGaZ/2aZzGaZmWaZZmaZM2aZ\\n AGZm/2ZmzGZmmWZmZmZmM2ZmAGYz/2YzzGYzmWYzZmYzM2YzAGYA\\n /2YAzGYAmWYAZmYAM2YAADP//zP/zDP/mTP/ZjP/MzP/ADPM/zPM\\n zDPMmTPMZjPMMzPMADOZ/zOZzDOZmTOZZjOZMzOZADNm/zNmzDNm\\n mTNmZjNmMzNmADMz/zMzzDMzmTMzZjMzMzMzADMA/zMAzDMAmTMA\\n ZjMAMzMAAAD//wD/zAD/mQD/ZgD/MwD/AADM/wDMzADMmQDMZgDM\\n MwDMAACZ/wCZzACZmQCZZgCZMwCZAABm/wBmzABmmQBmZgBmMwBm\\n AAAz/wAzzAAzmQAzZgAzMwAzAAAA/wAAzAAAmQAAZgAAM+4AAN0A\\n ALsAAKoAAIgAAHcAAFUAAEQAACIAABEAAADuAADdAAC7AACqAACI\\n AAB3AABVAABEAAAiAAARAAAA7gAA3QAAuwAAqgAAiAAAdwAAVQAA\\n RAAAIgAAEe7u7t3d3bu7u6qqqoiIiHd3d1VVVURERCIiIhEREQAA\\n ACH5BAEAAAAALAAAAAAgABgAAAiNAAH8G0iwoMGDCAcKTMiw4UBw\\n BPXVm0ixosWLFvVBHFjPoUeC9Tb+6/jRY0iQ/8iVbHiS40CVKxG2\\n HEkQZsyCM0mmvGkw50uePUV2tEnOZkyfQA8iTYpTKNOgKJ+C3AhO\\n p9SWVaVOfWj1KdauTL9q5UgVbFKsEjGqXVtP40NwcBnCjXtw7tx/\\n C8cSBBAQADs=\\n '", ")", ",", "'OldSwitch'", ":", "PhotoImage", "(", "data", "=", "'\\n R0lGODlhIAAYAPcAMf//////zP//mf//Zv//M///AP/M///MzP/M\\n mf/MZv/MM//MAP+Z//+ZzP+Zmf+ZZv+ZM/+ZAP9m//9mzP9mmf9m\\n Zv9mM/9mAP8z//8zzP8zmf8zZv8zM/8zAP8A//8AzP8Amf8AZv8A\\n M/8AAMz//8z/zMz/mcz/Zsz/M8z/AMzM/8zMzMzMmczMZszMM8zM\\n AMyZ/8yZzMyZmcyZZsyZM8yZAMxm/8xmzMxmmcxmZsxmM8xmAMwz\\n /8wzzMwzmcwzZswzM8wzAMwA/8wAzMwAmcwAZswAM8wAAJn//5n/\\n zJn/mZn/Zpn/M5n/AJnM/5nMzJnMmZnMZpnMM5nMAJmZ/5mZzJmZ\\n mZmZZpmZM5mZAJlm/5lmzJlmmZlmZplmM5lmAJkz/5kzzJkzmZkz\\n ZpkzM5kzAJkA/5kAzJkAmZkAZpkAM5kAAGb//2b/zGb/mWb/Zmb/\\n M2b/AGbM/2bMzGbMmWbMZmbMM2bMAGaZ/2aZzGaZmWaZZmaZM2aZ\\n AGZm/2ZmzGZmmWZmZmZmM2ZmAGYz/2YzzGYzmWYzZmYzM2YzAGYA\\n /2YAzGYAmWYAZmYAM2YAADP//zP/zDP/mTP/ZjP/MzP/ADPM/zPM\\n zDPMmTPMZjPMMzPMADOZ/zOZzDOZmTOZZjOZMzOZADNm/zNmzDNm\\n mTNmZjNmMzNmADMz/zMzzDMzmTMzZjMzMzMzADMA/zMAzDMAmTMA\\n ZjMAMzMAAAD//wD/zAD/mQD/ZgD/MwD/AADM/wDMzADMmQDMZgDM\\n MwDMAACZ/wCZzACZmQCZZgCZMwCZAABm/wBmzABmmQBmZgBmMwBm\\n AAAz/wAzzAAzmQAzZgAzMwAzAAAA/wAAzAAAmQAAZgAAM+4AAN0A\\n ALsAAKoAAIgAAHcAAFUAAEQAACIAABEAAADuAADdAAC7AACqAACI\\n AAB3AABVAABEAAAiAAARAAAA7gAA3QAAuwAAqgAAiAAAdwAAVQAA\\n RAAAIgAAEe7u7t3d3bu7u6qqqoiIiHd3d1VVVURERCIiIhEREQAA\\n ACH5BAEAAAAALAAAAAAgABgAAAhwAAEIHEiwoMGDCBMqXMiwocOH\\n ECNKnEixosWB3zJq3Mixo0eNAL7xG0mypMmTKPl9Cznyn8uWL/m5\\n /AeTpsyYI1eKlBnO5r+eLYHy9Ck0J8ubPmPOrMmUpM6UUKMa/Ui1\\n 6saLWLNq3cq1q9evYB0GBAA7\\n '", ")", ",", "'NetLink'", ":", "PhotoImage", "(", "data", "=", "'\\n R0lGODlhFgAWAPcAMf//////zP//mf//Zv//M///AP/M///MzP/M\\n mf/MZv/MM//MAP+Z//+ZzP+Zmf+ZZv+ZM/+ZAP9m//9mzP9mmf9m\\n Zv9mM/9mAP8z//8zzP8zmf8zZv8zM/8zAP8A//8AzP8Amf8AZv8A\\n M/8AAMz//8z/zMz/mcz/Zsz/M8z/AMzM/8zMzMzMmczMZszMM8zM\\n AMyZ/8yZzMyZmcyZZsyZM8yZAMxm/8xmzMxmmcxmZsxmM8xmAMwz\\n /8wzzMwzmcwzZswzM8wzAMwA/8wAzMwAmcwAZswAM8wAAJn//5n/\\n zJn/mZn/Zpn/M5n/AJnM/5nMzJnMmZnMZpnMM5nMAJmZ/5mZzJmZ\\n mZmZZpmZM5mZAJlm/5lmzJlmmZlmZplmM5lmAJkz/5kzzJkzmZkz\\n ZpkzM5kzAJkA/5kAzJkAmZkAZpkAM5kAAGb//2b/zGb/mWb/Zmb/\\n M2b/AGbM/2bMzGbMmWbMZmbMM2bMAGaZ/2aZzGaZmWaZZmaZM2aZ\\n AGZm/2ZmzGZmmWZmZmZmM2ZmAGYz/2YzzGYzmWYzZmYzM2YzAGYA\\n /2YAzGYAmWYAZmYAM2YAADP//zP/zDP/mTP/ZjP/MzP/ADPM/zPM\\n zDPMmTPMZjPMMzPMADOZ/zOZzDOZmTOZZjOZMzOZADNm/zNmzDNm\\n mTNmZjNmMzNmADMz/zMzzDMzmTMzZjMzMzMzADMA/zMAzDMAmTMA\\n ZjMAMzMAAAD//wD/zAD/mQD/ZgD/MwD/AADM/wDMzADMmQDMZgDM\\n MwDMAACZ/wCZzACZmQCZZgCZMwCZAABm/wBmzABmmQBmZgBmMwBm\\n AAAz/wAzzAAzmQAzZgAzMwAzAAAA/wAAzAAAmQAAZgAAM+4AAN0A\\n ALsAAKoAAIgAAHcAAFUAAEQAACIAABEAAADuAADdAAC7AACqAACI\\n AAB3AABVAABEAAAiAAARAAAA7gAA3QAAuwAAqgAAiAAAdwAAVQAA\\n RAAAIgAAEe7u7t3d3bu7u6qqqoiIiHd3d1VVVURERCIiIhEREQAA\\n ACH5BAEAAAAALAAAAAAWABYAAAhIAAEIHEiwoEGBrhIeXEgwoUKG\\n Cx0+hGhQoiuKBy1irChxY0GNHgeCDAlgZEiTHlFuVImRJUWXEGEy\\n lBmxI8mSNknm1Dnx5sCAADs=\\n '", ")", "}" ]
create and return images for miniedit .
train
false
25,224
def unpack_ext(ext_path): modcache = os.path.join(OPTIONS.saltdir, 'running_data', 'var', 'cache', 'salt', 'minion', 'extmods') tfile = tarfile.TarFile.gzopen(ext_path) old_umask = os.umask(63) tfile.extractall(path=modcache) tfile.close() os.umask(old_umask) os.unlink(ext_path) ver_path = os.path.join(modcache, 'ext_version') ver_dst = os.path.join(OPTIONS.saltdir, 'ext_version') shutil.move(ver_path, ver_dst)
[ "def", "unpack_ext", "(", "ext_path", ")", ":", "modcache", "=", "os", ".", "path", ".", "join", "(", "OPTIONS", ".", "saltdir", ",", "'running_data'", ",", "'var'", ",", "'cache'", ",", "'salt'", ",", "'minion'", ",", "'extmods'", ")", "tfile", "=", "tarfile", ".", "TarFile", ".", "gzopen", "(", "ext_path", ")", "old_umask", "=", "os", ".", "umask", "(", "63", ")", "tfile", ".", "extractall", "(", "path", "=", "modcache", ")", "tfile", ".", "close", "(", ")", "os", ".", "umask", "(", "old_umask", ")", "os", ".", "unlink", "(", "ext_path", ")", "ver_path", "=", "os", ".", "path", ".", "join", "(", "modcache", ",", "'ext_version'", ")", "ver_dst", "=", "os", ".", "path", ".", "join", "(", "OPTIONS", ".", "saltdir", ",", "'ext_version'", ")", "shutil", ".", "move", "(", "ver_path", ",", "ver_dst", ")" ]
unpack the external modules .
train
true
25,225
def _stringsToObjects(strings, arglist, proto): objects = {} myStrings = strings.copy() for (argname, argparser) in arglist: argparser.fromBox(argname, myStrings, objects, proto) return objects
[ "def", "_stringsToObjects", "(", "strings", ",", "arglist", ",", "proto", ")", ":", "objects", "=", "{", "}", "myStrings", "=", "strings", ".", "copy", "(", ")", "for", "(", "argname", ",", "argparser", ")", "in", "arglist", ":", "argparser", ".", "fromBox", "(", "argname", ",", "myStrings", ",", "objects", ",", "proto", ")", "return", "objects" ]
convert an ampbox to a dictionary of python objects .
train
true
25,226
def _find_closest_point_on_path(lc, point): ds = np.sum(((lc - point[None, :]) ** 2), 1) imin = np.argmin(ds) dmin = np.inf xcmin = None legmin = (None, None) closed = mlab.is_closed_polygon(lc) legs = [] if ((imin > 0) or closed): legs.append((((imin - 1) % len(lc)), imin)) if ((imin < (len(lc) - 1)) or closed): legs.append((imin, ((imin + 1) % len(lc)))) for leg in legs: (d, xc) = _find_closest_point_on_leg(lc[leg[0]], lc[leg[1]], point) if (d < dmin): dmin = d xcmin = xc legmin = leg return (dmin, xcmin, legmin)
[ "def", "_find_closest_point_on_path", "(", "lc", ",", "point", ")", ":", "ds", "=", "np", ".", "sum", "(", "(", "(", "lc", "-", "point", "[", "None", ",", ":", "]", ")", "**", "2", ")", ",", "1", ")", "imin", "=", "np", ".", "argmin", "(", "ds", ")", "dmin", "=", "np", ".", "inf", "xcmin", "=", "None", "legmin", "=", "(", "None", ",", "None", ")", "closed", "=", "mlab", ".", "is_closed_polygon", "(", "lc", ")", "legs", "=", "[", "]", "if", "(", "(", "imin", ">", "0", ")", "or", "closed", ")", ":", "legs", ".", "append", "(", "(", "(", "(", "imin", "-", "1", ")", "%", "len", "(", "lc", ")", ")", ",", "imin", ")", ")", "if", "(", "(", "imin", "<", "(", "len", "(", "lc", ")", "-", "1", ")", ")", "or", "closed", ")", ":", "legs", ".", "append", "(", "(", "imin", ",", "(", "(", "imin", "+", "1", ")", "%", "len", "(", "lc", ")", ")", ")", ")", "for", "leg", "in", "legs", ":", "(", "d", ",", "xc", ")", "=", "_find_closest_point_on_leg", "(", "lc", "[", "leg", "[", "0", "]", "]", ",", "lc", "[", "leg", "[", "1", "]", "]", ",", "point", ")", "if", "(", "d", "<", "dmin", ")", ":", "dmin", "=", "d", "xcmin", "=", "xc", "legmin", "=", "leg", "return", "(", "dmin", ",", "xcmin", ",", "legmin", ")" ]
lc: coordinates of vertices point: coordinates of test point .
train
false
25,227
def flushdb(host=None, port=None, db=None, password=None): server = _connect(host, port, db, password) return server.flushdb()
[ "def", "flushdb", "(", "host", "=", "None", ",", "port", "=", "None", ",", "db", "=", "None", ",", "password", "=", "None", ")", ":", "server", "=", "_connect", "(", "host", ",", "port", ",", "db", ",", "password", ")", "return", "server", ".", "flushdb", "(", ")" ]
remove all keys from the selected database cli example: .
train
true
25,228
def _addHeaderToRequest(request, header): requestHeaders = request.requestHeaders (name, value) = header values = requestHeaders.getRawHeaders(name) if (values is not None): values.append(value) else: requestHeaders.setRawHeaders(name, [value]) if (name == 'content-length'): request.gotLength(int(value)) return True return False
[ "def", "_addHeaderToRequest", "(", "request", ",", "header", ")", ":", "requestHeaders", "=", "request", ".", "requestHeaders", "(", "name", ",", "value", ")", "=", "header", "values", "=", "requestHeaders", ".", "getRawHeaders", "(", "name", ")", "if", "(", "values", "is", "not", "None", ")", ":", "values", ".", "append", "(", "value", ")", "else", ":", "requestHeaders", ".", "setRawHeaders", "(", "name", ",", "[", "value", "]", ")", "if", "(", "name", "==", "'content-length'", ")", ":", "request", ".", "gotLength", "(", "int", "(", "value", ")", ")", "return", "True", "return", "False" ]
add a header tuple to a request header object .
train
false
25,229
def project_update(project_id=None, name=None, description=None, enabled=None, profile=None, **connection_args): auth(profile, **connection_args) if (_OS_IDENTITY_API_VERSION > 2): return tenant_update(tenant_id=project_id, name=name, description=description, enabled=enabled, profile=profile, **connection_args) else: return False
[ "def", "project_update", "(", "project_id", "=", "None", ",", "name", "=", "None", ",", "description", "=", "None", ",", "enabled", "=", "None", ",", "profile", "=", "None", ",", "**", "connection_args", ")", ":", "auth", "(", "profile", ",", "**", "connection_args", ")", "if", "(", "_OS_IDENTITY_API_VERSION", ">", "2", ")", ":", "return", "tenant_update", "(", "tenant_id", "=", "project_id", ",", "name", "=", "name", ",", "description", "=", "description", ",", "enabled", "=", "enabled", ",", "profile", "=", "profile", ",", "**", "connection_args", ")", "else", ":", "return", "False" ]
update a tenants information the following fields may be updated: name .
train
true
25,230
def v7(method_v7): frame = currentframe().f_back return frame.f_locals.get(method_v7.__name__, method_v7)
[ "def", "v7", "(", "method_v7", ")", ":", "frame", "=", "currentframe", "(", ")", ".", "f_back", "return", "frame", ".", "f_locals", ".", "get", "(", "method_v7", ".", "__name__", ",", "method_v7", ")" ]
decorate a method that supports the old-style api only .
train
false
25,232
def format_jnlp_file_lines(web_flag, url, tep_fp): lines = [jnlp_top_block] if web_flag: lines += ['http://topiaryexplorer.sourceforge.net/app/'] else: topiaryexplorer_project_dir = load_qiime_config()['topiaryexplorer_project_dir'] if topiaryexplorer_project_dir: lines += [('file:' + topiaryexplorer_project_dir)] else: print "WARNING: Couldn't create jnlp file - topiaryexplorer_project_dir is not defined in your qiime_config. tep file was created sucessfully." lines += [jnlp_middle_block] if url: lines += [url] else: lines += [abspath(tep_fp)] lines += [jnlp_bottom_block] return lines
[ "def", "format_jnlp_file_lines", "(", "web_flag", ",", "url", ",", "tep_fp", ")", ":", "lines", "=", "[", "jnlp_top_block", "]", "if", "web_flag", ":", "lines", "+=", "[", "'http://topiaryexplorer.sourceforge.net/app/'", "]", "else", ":", "topiaryexplorer_project_dir", "=", "load_qiime_config", "(", ")", "[", "'topiaryexplorer_project_dir'", "]", "if", "topiaryexplorer_project_dir", ":", "lines", "+=", "[", "(", "'file:'", "+", "topiaryexplorer_project_dir", ")", "]", "else", ":", "print", "\"WARNING: Couldn't create jnlp file - topiaryexplorer_project_dir is not defined in your qiime_config. tep file was created sucessfully.\"", "lines", "+=", "[", "jnlp_middle_block", "]", "if", "url", ":", "lines", "+=", "[", "url", "]", "else", ":", "lines", "+=", "[", "abspath", "(", "tep_fp", ")", "]", "lines", "+=", "[", "jnlp_bottom_block", "]", "return", "lines" ]
format the jnlp file for topiaryexplorer .
train
false
25,233
@core_helper def sorted_extras(package_extras, auto_clean=False, subs=None, exclude=None): if (not exclude): exclude = config.get('package_hide_extras', []) output = [] for extra in sorted(package_extras, key=(lambda x: x['key'])): if (extra.get('state') == 'deleted'): continue (k, v) = (extra['key'], extra['value']) if (k in exclude): continue if (subs and (k in subs)): k = subs[k] elif auto_clean: k = k.replace('_', ' ').replace('-', ' ').title() if isinstance(v, (list, tuple)): v = ', '.join(map(unicode, v)) output.append((k, v)) return output
[ "@", "core_helper", "def", "sorted_extras", "(", "package_extras", ",", "auto_clean", "=", "False", ",", "subs", "=", "None", ",", "exclude", "=", "None", ")", ":", "if", "(", "not", "exclude", ")", ":", "exclude", "=", "config", ".", "get", "(", "'package_hide_extras'", ",", "[", "]", ")", "output", "=", "[", "]", "for", "extra", "in", "sorted", "(", "package_extras", ",", "key", "=", "(", "lambda", "x", ":", "x", "[", "'key'", "]", ")", ")", ":", "if", "(", "extra", ".", "get", "(", "'state'", ")", "==", "'deleted'", ")", ":", "continue", "(", "k", ",", "v", ")", "=", "(", "extra", "[", "'key'", "]", ",", "extra", "[", "'value'", "]", ")", "if", "(", "k", "in", "exclude", ")", ":", "continue", "if", "(", "subs", "and", "(", "k", "in", "subs", ")", ")", ":", "k", "=", "subs", "[", "k", "]", "elif", "auto_clean", ":", "k", "=", "k", ".", "replace", "(", "'_'", ",", "' '", ")", ".", "replace", "(", "'-'", ",", "' '", ")", ".", "title", "(", ")", "if", "isinstance", "(", "v", ",", "(", "list", ",", "tuple", ")", ")", ":", "v", "=", "', '", ".", "join", "(", "map", "(", "unicode", ",", "v", ")", ")", "output", ".", "append", "(", "(", "k", ",", "v", ")", ")", "return", "output" ]
used for outputting package extras .
train
false
25,236
def getitem(a, b): return a[b]
[ "def", "getitem", "(", "a", ",", "b", ")", ":", "return", "a", "[", "b", "]" ]
get item from an iterable at index .
train
false
25,237
def normalize_spaces(s): return ' '.join(s.split())
[ "def", "normalize_spaces", "(", "s", ")", ":", "return", "' '", ".", "join", "(", "s", ".", "split", "(", ")", ")" ]
replace any sequence of whitespace characters with a single space .
train
false
25,238
def test_solve_rational(): assert (solve(((x - (y ** 3)) / ((y ** 2) * sqrt((1 - (y ** 2))))), x) == [(y ** 3)])
[ "def", "test_solve_rational", "(", ")", ":", "assert", "(", "solve", "(", "(", "(", "x", "-", "(", "y", "**", "3", ")", ")", "/", "(", "(", "y", "**", "2", ")", "*", "sqrt", "(", "(", "1", "-", "(", "y", "**", "2", ")", ")", ")", ")", ")", ",", "x", ")", "==", "[", "(", "y", "**", "3", ")", "]", ")" ]
test solve for rational functions .
train
false
25,239
def backup_create(context, values): return IMPL.backup_create(context, values)
[ "def", "backup_create", "(", "context", ",", "values", ")", ":", "return", "IMPL", ".", "backup_create", "(", "context", ",", "values", ")" ]
create a backup from the values dictionary .
train
false
25,240
def _delete_asset(course_key, asset_key_string): if asset_key_string: try: asset_key = AssetKey.from_string(asset_key_string) except InvalidKeyError: if ('/' == asset_key_string[0]): asset_key_string = asset_key_string[1:] try: asset_key = AssetKey.from_string(asset_key_string) except InvalidKeyError: LOGGER.info('In course %r, unable to parse asset key %r, not attempting to delete signatory.', course_key, asset_key_string) return else: LOGGER.info('In course %r, unable to parse asset key %r, not attempting to delete signatory.', course_key, asset_key_string) return try: delete_asset(course_key, asset_key) except AssetNotFoundException: pass
[ "def", "_delete_asset", "(", "course_key", ",", "asset_key_string", ")", ":", "if", "asset_key_string", ":", "try", ":", "asset_key", "=", "AssetKey", ".", "from_string", "(", "asset_key_string", ")", "except", "InvalidKeyError", ":", "if", "(", "'/'", "==", "asset_key_string", "[", "0", "]", ")", ":", "asset_key_string", "=", "asset_key_string", "[", "1", ":", "]", "try", ":", "asset_key", "=", "AssetKey", ".", "from_string", "(", "asset_key_string", ")", "except", "InvalidKeyError", ":", "LOGGER", ".", "info", "(", "'In course %r, unable to parse asset key %r, not attempting to delete signatory.'", ",", "course_key", ",", "asset_key_string", ")", "return", "else", ":", "LOGGER", ".", "info", "(", "'In course %r, unable to parse asset key %r, not attempting to delete signatory.'", ",", "course_key", ",", "asset_key_string", ")", "return", "try", ":", "delete_asset", "(", "course_key", ",", "asset_key", ")", "except", "AssetNotFoundException", ":", "pass" ]
internal method used to create asset key from string and remove asset by calling delete_asset method of assets module .
train
false
25,241
def no_os_popen(logical_line): if ('os.popen(' in logical_line): (yield (0, 'N348 Deprecated library function os.popen(). Replace it using subprocess module. '))
[ "def", "no_os_popen", "(", "logical_line", ")", ":", "if", "(", "'os.popen('", "in", "logical_line", ")", ":", "(", "yield", "(", "0", ",", "'N348 Deprecated library function os.popen(). Replace it using subprocess module. '", ")", ")" ]
disallow os .
train
false
25,243
def instance_metadata_update(context, instance_uuid, metadata, delete): return IMPL.instance_metadata_update(context, instance_uuid, metadata, delete)
[ "def", "instance_metadata_update", "(", "context", ",", "instance_uuid", ",", "metadata", ",", "delete", ")", ":", "return", "IMPL", ".", "instance_metadata_update", "(", "context", ",", "instance_uuid", ",", "metadata", ",", "delete", ")" ]
update metadata if it exists .
train
false
25,244
@utils.arg('fixed_ip', metavar='<fixed_ip>', help=_('Fixed IP Address.')) @deprecated_network def do_fixed_ip_reserve(cs, args): cs.fixed_ips.reserve(args.fixed_ip)
[ "@", "utils", ".", "arg", "(", "'fixed_ip'", ",", "metavar", "=", "'<fixed_ip>'", ",", "help", "=", "_", "(", "'Fixed IP Address.'", ")", ")", "@", "deprecated_network", "def", "do_fixed_ip_reserve", "(", "cs", ",", "args", ")", ":", "cs", ".", "fixed_ips", ".", "reserve", "(", "args", ".", "fixed_ip", ")" ]
reserve a fixed ip .
train
false
25,245
def getWrappedInteger(integer, modulo): if (integer >= modulo): return modulo if (integer >= 0): return integer return (integer % modulo)
[ "def", "getWrappedInteger", "(", "integer", ",", "modulo", ")", ":", "if", "(", "integer", ">=", "modulo", ")", ":", "return", "modulo", "if", "(", "integer", ">=", "0", ")", ":", "return", "integer", "return", "(", "integer", "%", "modulo", ")" ]
get wrapped integer .
train
false
25,246
def Std(xs, mu=None, ddof=0): var = Var(xs, mu, ddof) return math.sqrt(var)
[ "def", "Std", "(", "xs", ",", "mu", "=", "None", ",", "ddof", "=", "0", ")", ":", "var", "=", "Var", "(", "xs", ",", "mu", ",", "ddof", ")", "return", "math", ".", "sqrt", "(", "var", ")" ]
computes standard deviation .
train
false
25,247
@hook.command() def shorten(text): args = text.split() url = args[0] custom = (args[1] if (len(args) > 1) else None) try: return web.shorten(url, custom=custom) except web.ServiceError as e: return e.message
[ "@", "hook", ".", "command", "(", ")", "def", "shorten", "(", "text", ")", ":", "args", "=", "text", ".", "split", "(", ")", "url", "=", "args", "[", "0", "]", "custom", "=", "(", "args", "[", "1", "]", "if", "(", "len", "(", "args", ")", ">", "1", ")", "else", "None", ")", "try", ":", "return", "web", ".", "shorten", "(", "url", ",", "custom", "=", "custom", ")", "except", "web", ".", "ServiceError", "as", "e", ":", "return", "e", ".", "message" ]
smartly shorten a given string .
train
false
25,248
def clientresp(data, **kwargs): creds = config.get_credentials() cfg = config.get_config() dumps_kwargs = {'sort_keys': True, 'cls': utils.PlotlyJSONEncoder} payload = {'platform': 'python', 'version': version.__version__, 'args': _json.dumps(data, **dumps_kwargs), 'un': creds['username'], 'key': creds['api_key'], 'origin': 'plot', 'kwargs': _json.dumps(kwargs, **dumps_kwargs)} url = '{plotly_domain}/clientresp'.format(**cfg) response = request('post', url, data=payload) parsed_content = response.json() if parsed_content.get('warning'): warnings.warn(parsed_content['warning']) if parsed_content.get('message'): print parsed_content['message'] return response
[ "def", "clientresp", "(", "data", ",", "**", "kwargs", ")", ":", "creds", "=", "config", ".", "get_credentials", "(", ")", "cfg", "=", "config", ".", "get_config", "(", ")", "dumps_kwargs", "=", "{", "'sort_keys'", ":", "True", ",", "'cls'", ":", "utils", ".", "PlotlyJSONEncoder", "}", "payload", "=", "{", "'platform'", ":", "'python'", ",", "'version'", ":", "version", ".", "__version__", ",", "'args'", ":", "_json", ".", "dumps", "(", "data", ",", "**", "dumps_kwargs", ")", ",", "'un'", ":", "creds", "[", "'username'", "]", ",", "'key'", ":", "creds", "[", "'api_key'", "]", ",", "'origin'", ":", "'plot'", ",", "'kwargs'", ":", "_json", ".", "dumps", "(", "kwargs", ",", "**", "dumps_kwargs", ")", "}", "url", "=", "'{plotly_domain}/clientresp'", ".", "format", "(", "**", "cfg", ")", "response", "=", "request", "(", "'post'", ",", "url", ",", "data", "=", "payload", ")", "parsed_content", "=", "response", ".", "json", "(", ")", "if", "parsed_content", ".", "get", "(", "'warning'", ")", ":", "warnings", ".", "warn", "(", "parsed_content", "[", "'warning'", "]", ")", "if", "parsed_content", ".", "get", "(", "'message'", ")", ":", "print", "parsed_content", "[", "'message'", "]", "return", "response" ]
deprecated endpoint .
train
false
25,250
def _check_header_magic_bytes(file_obj): file_obj.seek(0, 0) magic = file_obj.read(4) return (magic == 'PAR1')
[ "def", "_check_header_magic_bytes", "(", "file_obj", ")", ":", "file_obj", ".", "seek", "(", "0", ",", "0", ")", "magic", "=", "file_obj", ".", "read", "(", "4", ")", "return", "(", "magic", "==", "'PAR1'", ")" ]
check if the file-like obj has the par1 magic bytes at the header .
train
false
25,252
def cliques_containing_node(G, nodes=None, cliques=None): if (cliques is None): cliques = list(find_cliques(G)) if (nodes is None): nodes = list(G.nodes()) if (not isinstance(nodes, list)): v = nodes vcliques = [c for c in cliques if (v in c)] else: vcliques = {} for v in nodes: vcliques[v] = [c for c in cliques if (v in c)] return vcliques
[ "def", "cliques_containing_node", "(", "G", ",", "nodes", "=", "None", ",", "cliques", "=", "None", ")", ":", "if", "(", "cliques", "is", "None", ")", ":", "cliques", "=", "list", "(", "find_cliques", "(", "G", ")", ")", "if", "(", "nodes", "is", "None", ")", ":", "nodes", "=", "list", "(", "G", ".", "nodes", "(", ")", ")", "if", "(", "not", "isinstance", "(", "nodes", ",", "list", ")", ")", ":", "v", "=", "nodes", "vcliques", "=", "[", "c", "for", "c", "in", "cliques", "if", "(", "v", "in", "c", ")", "]", "else", ":", "vcliques", "=", "{", "}", "for", "v", "in", "nodes", ":", "vcliques", "[", "v", "]", "=", "[", "c", "for", "c", "in", "cliques", "if", "(", "v", "in", "c", ")", "]", "return", "vcliques" ]
returns a list of cliques containing the given node .
train
false
25,253
def _add_retry_host(filter_properties, host, node): retry = filter_properties.get('retry', None) if (not retry): return hosts = retry['hosts'] hosts.append([host, node])
[ "def", "_add_retry_host", "(", "filter_properties", ",", "host", ",", "node", ")", ":", "retry", "=", "filter_properties", ".", "get", "(", "'retry'", ",", "None", ")", "if", "(", "not", "retry", ")", ":", "return", "hosts", "=", "retry", "[", "'hosts'", "]", "hosts", ".", "append", "(", "[", "host", ",", "node", "]", ")" ]
add a retry entry for the selected compute node .
train
false
25,254
def mirror_y(clip, apply_to='mask'): return clip.fl_image((lambda f: f[::(-1)]), apply_to=apply_to)
[ "def", "mirror_y", "(", "clip", ",", "apply_to", "=", "'mask'", ")", ":", "return", "clip", ".", "fl_image", "(", "(", "lambda", "f", ":", "f", "[", ":", ":", "(", "-", "1", ")", "]", ")", ",", "apply_to", "=", "apply_to", ")" ]
flips the clip vertically .
train
false
25,256
def make_derivative_operator(M, z): def doit(C): r = ((z * C.diff(z)) + (C * M)) r = r.applyfunc(make_simp(z)) return r return doit
[ "def", "make_derivative_operator", "(", "M", ",", "z", ")", ":", "def", "doit", "(", "C", ")", ":", "r", "=", "(", "(", "z", "*", "C", ".", "diff", "(", "z", ")", ")", "+", "(", "C", "*", "M", ")", ")", "r", "=", "r", ".", "applyfunc", "(", "make_simp", "(", "z", ")", ")", "return", "r", "return", "doit" ]
create a derivative operator .
train
false
25,257
def hmc_move(s_rng, positions, energy_fn, stepsize, n_steps): initial_vel = s_rng.normal(size=positions.shape) (final_pos, final_vel) = simulate_dynamics(initial_pos=positions, initial_vel=initial_vel, stepsize=stepsize, n_steps=n_steps, energy_fn=energy_fn) accept = metropolis_hastings_accept(energy_prev=hamiltonian(positions, initial_vel, energy_fn), energy_next=hamiltonian(final_pos, final_vel, energy_fn), s_rng=s_rng) return (accept, final_pos)
[ "def", "hmc_move", "(", "s_rng", ",", "positions", ",", "energy_fn", ",", "stepsize", ",", "n_steps", ")", ":", "initial_vel", "=", "s_rng", ".", "normal", "(", "size", "=", "positions", ".", "shape", ")", "(", "final_pos", ",", "final_vel", ")", "=", "simulate_dynamics", "(", "initial_pos", "=", "positions", ",", "initial_vel", "=", "initial_vel", ",", "stepsize", "=", "stepsize", ",", "n_steps", "=", "n_steps", ",", "energy_fn", "=", "energy_fn", ")", "accept", "=", "metropolis_hastings_accept", "(", "energy_prev", "=", "hamiltonian", "(", "positions", ",", "initial_vel", ",", "energy_fn", ")", ",", "energy_next", "=", "hamiltonian", "(", "final_pos", ",", "final_vel", ",", "energy_fn", ")", ",", "s_rng", "=", "s_rng", ")", "return", "(", "accept", ",", "final_pos", ")" ]
this function performs one-step of hybrid monte-carlo sampling .
train
false
25,258
def _zpklp2lp(z, p, k, wo=1): degree = _relative_degree(z, p) z_lp = [(wo * z1) for z1 in z] p_lp = [(wo * p1) for p1 in p] k_lp = (k * (wo ** degree)) return (z_lp, p_lp, k_lp)
[ "def", "_zpklp2lp", "(", "z", ",", "p", ",", "k", ",", "wo", "=", "1", ")", ":", "degree", "=", "_relative_degree", "(", "z", ",", "p", ")", "z_lp", "=", "[", "(", "wo", "*", "z1", ")", "for", "z1", "in", "z", "]", "p_lp", "=", "[", "(", "wo", "*", "p1", ")", "for", "p1", "in", "p", "]", "k_lp", "=", "(", "k", "*", "(", "wo", "**", "degree", ")", ")", "return", "(", "z_lp", ",", "p_lp", ",", "k_lp", ")" ]
transform a lowpass filter prototype to a different frequency .
train
false
25,260
def __get_tags(vm_): t = config.get_cloud_config_value('tags', vm_, __opts__, default='[]', search_global=False) try: tags = literal_eval(t) except Exception: tags = None if ((not tags) or (not isinstance(tags, list))): tags = None return tags
[ "def", "__get_tags", "(", "vm_", ")", ":", "t", "=", "config", ".", "get_cloud_config_value", "(", "'tags'", ",", "vm_", ",", "__opts__", ",", "default", "=", "'[]'", ",", "search_global", "=", "False", ")", "try", ":", "tags", "=", "literal_eval", "(", "t", ")", "except", "Exception", ":", "tags", "=", "None", "if", "(", "(", "not", "tags", ")", "or", "(", "not", "isinstance", "(", "tags", ",", "list", ")", ")", ")", ":", "tags", "=", "None", "return", "tags" ]
get configured tags .
train
true
25,261
def reset(): _runtime.reset()
[ "def", "reset", "(", ")", ":", "_runtime", ".", "reset", "(", ")" ]
interface to git-reset(1)_ .
train
false
25,262
def test_allknn_sk_estimator(): check_estimator(AllKNN)
[ "def", "test_allknn_sk_estimator", "(", ")", ":", "check_estimator", "(", "AllKNN", ")" ]
test the sklearn estimator compatibility .
train
false
25,263
def compareSegmentLength(endpoint, otherEndpoint): if (endpoint.segmentLength > otherEndpoint.segmentLength): return 1 if (endpoint.segmentLength < otherEndpoint.segmentLength): return (-1) return 0
[ "def", "compareSegmentLength", "(", "endpoint", ",", "otherEndpoint", ")", ":", "if", "(", "endpoint", ".", "segmentLength", ">", "otherEndpoint", ".", "segmentLength", ")", ":", "return", "1", "if", "(", "endpoint", ".", "segmentLength", "<", "otherEndpoint", ".", "segmentLength", ")", ":", "return", "(", "-", "1", ")", "return", "0" ]
get comparison in order to sort endpoints in ascending order of segment length .
train
false
25,265
def _stdin_ready_other(): return True
[ "def", "_stdin_ready_other", "(", ")", ":", "return", "True" ]
return true .
train
false
25,267
def get_bind_data(zone_id, profile): conn = _get_driver(profile=profile) zone = conn.get_zone(zone_id) return conn.export_zone_to_bind_format(zone)
[ "def", "get_bind_data", "(", "zone_id", ",", "profile", ")", ":", "conn", "=", "_get_driver", "(", "profile", "=", "profile", ")", "zone", "=", "conn", ".", "get_zone", "(", "zone_id", ")", "return", "conn", ".", "export_zone_to_bind_format", "(", "zone", ")" ]
export zone to the bind compatible format .
train
true
25,269
def manage_entries(options, selection, approved): approved_text = (u'approved' if approved else u'pending') with Session() as session: if (selection == u'all'): entries = list_pending_entries(session=session, approved=(not approved)) else: try: entry = get_entry_by_id(session, selection) if (entry.approved is approved): console((colorize(u'red', u'ERROR: ') + (u'Entry with ID %s is already %s' % (entry.id, approved_text)))) sys.exit(1) except NoResultFound: console((u'Pending entry with ID %s does not exist' % selection)) sys.exit(1) else: entries = [entry] if (not entries): console((u'All entries are already %s' % approved_text)) return for entry in entries: if (entry.approved is not approved): console((u'Setting pending entry with ID %s status to %s' % (entry.id, approved_text))) entry.approved = approved
[ "def", "manage_entries", "(", "options", ",", "selection", ",", "approved", ")", ":", "approved_text", "=", "(", "u'approved'", "if", "approved", "else", "u'pending'", ")", "with", "Session", "(", ")", "as", "session", ":", "if", "(", "selection", "==", "u'all'", ")", ":", "entries", "=", "list_pending_entries", "(", "session", "=", "session", ",", "approved", "=", "(", "not", "approved", ")", ")", "else", ":", "try", ":", "entry", "=", "get_entry_by_id", "(", "session", ",", "selection", ")", "if", "(", "entry", ".", "approved", "is", "approved", ")", ":", "console", "(", "(", "colorize", "(", "u'red'", ",", "u'ERROR: '", ")", "+", "(", "u'Entry with ID %s is already %s'", "%", "(", "entry", ".", "id", ",", "approved_text", ")", ")", ")", ")", "sys", ".", "exit", "(", "1", ")", "except", "NoResultFound", ":", "console", "(", "(", "u'Pending entry with ID %s does not exist'", "%", "selection", ")", ")", "sys", ".", "exit", "(", "1", ")", "else", ":", "entries", "=", "[", "entry", "]", "if", "(", "not", "entries", ")", ":", "console", "(", "(", "u'All entries are already %s'", "%", "approved_text", ")", ")", "return", "for", "entry", "in", "entries", ":", "if", "(", "entry", ".", "approved", "is", "not", "approved", ")", ":", "console", "(", "(", "u'Setting pending entry with ID %s status to %s'", "%", "(", "entry", ".", "id", ",", "approved_text", ")", ")", ")", "entry", ".", "approved", "=", "approved" ]
manage pending entries .
train
false
25,270
def inthread(func): @wraps(func) def wrapped(*a, **kw): return threads.deferToThread(func, *a, **kw) return wrapped
[ "def", "inthread", "(", "func", ")", ":", "@", "wraps", "(", "func", ")", "def", "wrapped", "(", "*", "a", ",", "**", "kw", ")", ":", "return", "threads", ".", "deferToThread", "(", "func", ",", "*", "a", ",", "**", "kw", ")", "return", "wrapped" ]
decorator to call a function in a thread and return a deferred with the result .
train
false
25,272
def validateMasterOption(master): try: (hostname, port) = master.split(':') port = int(port) except (TypeError, ValueError): raise usage.UsageError("master must have the form 'hostname:port'")
[ "def", "validateMasterOption", "(", "master", ")", ":", "try", ":", "(", "hostname", ",", "port", ")", "=", "master", ".", "split", "(", "':'", ")", "port", "=", "int", "(", "port", ")", "except", "(", "TypeError", ",", "ValueError", ")", ":", "raise", "usage", ".", "UsageError", "(", "\"master must have the form 'hostname:port'\"", ")" ]
validate master command line option .
train
true
25,273
def is_cnf(expr): return _is_form(expr, And, Or)
[ "def", "is_cnf", "(", "expr", ")", ":", "return", "_is_form", "(", "expr", ",", "And", ",", "Or", ")" ]
test whether or not an expression is in conjunctive normal form .
train
false
25,275
def cachedir_index_del(minion_id, base=None): base = init_cachedir(base) index_file = os.path.join(base, 'index.p') lock_file(index_file) if os.path.exists(index_file): with salt.utils.fopen(index_file, 'r') as fh_: index = msgpack.load(fh_) else: return if (minion_id in index): del index[minion_id] with salt.utils.fopen(index_file, 'w') as fh_: msgpack.dump(index, fh_) unlock_file(index_file)
[ "def", "cachedir_index_del", "(", "minion_id", ",", "base", "=", "None", ")", ":", "base", "=", "init_cachedir", "(", "base", ")", "index_file", "=", "os", ".", "path", ".", "join", "(", "base", ",", "'index.p'", ")", "lock_file", "(", "index_file", ")", "if", "os", ".", "path", ".", "exists", "(", "index_file", ")", ":", "with", "salt", ".", "utils", ".", "fopen", "(", "index_file", ",", "'r'", ")", "as", "fh_", ":", "index", "=", "msgpack", ".", "load", "(", "fh_", ")", "else", ":", "return", "if", "(", "minion_id", "in", "index", ")", ":", "del", "index", "[", "minion_id", "]", "with", "salt", ".", "utils", ".", "fopen", "(", "index_file", ",", "'w'", ")", "as", "fh_", ":", "msgpack", ".", "dump", "(", "index", ",", "fh_", ")", "unlock_file", "(", "index_file", ")" ]
delete an entry from the cachedir index .
train
false
25,276
def config(conf_path='/etc/rsyncd.conf'): ret = '' try: with salt.utils.fopen(conf_path, 'r') as fp_: for line in fp_: ret += line except IOError as exc: if (exc.errno == errno.ENOENT): raise CommandExecutionError('{0} does not exist'.format(conf_path)) elif (exc.errno == errno.EACCES): raise CommandExecutionError('Unable to read {0}, access denied'.format(conf_path)) elif (exc.errno == errno.EISDIR): raise CommandExecutionError('Unable to read {0}, path is a directory'.format(conf_path)) else: raise CommandExecutionError('Error {0}: {1}'.format(exc.errno, exc.strerror)) else: return ret
[ "def", "config", "(", "conf_path", "=", "'/etc/rsyncd.conf'", ")", ":", "ret", "=", "''", "try", ":", "with", "salt", ".", "utils", ".", "fopen", "(", "conf_path", ",", "'r'", ")", "as", "fp_", ":", "for", "line", "in", "fp_", ":", "ret", "+=", "line", "except", "IOError", "as", "exc", ":", "if", "(", "exc", ".", "errno", "==", "errno", ".", "ENOENT", ")", ":", "raise", "CommandExecutionError", "(", "'{0} does not exist'", ".", "format", "(", "conf_path", ")", ")", "elif", "(", "exc", ".", "errno", "==", "errno", ".", "EACCES", ")", ":", "raise", "CommandExecutionError", "(", "'Unable to read {0}, access denied'", ".", "format", "(", "conf_path", ")", ")", "elif", "(", "exc", ".", "errno", "==", "errno", ".", "EISDIR", ")", ":", "raise", "CommandExecutionError", "(", "'Unable to read {0}, path is a directory'", ".", "format", "(", "conf_path", ")", ")", "else", ":", "raise", "CommandExecutionError", "(", "'Error {0}: {1}'", ".", "format", "(", "exc", ".", "errno", ",", "exc", ".", "strerror", ")", ")", "else", ":", "return", "ret" ]
set traffic server configuration variable values .
train
true
25,277
def _createPluginDummy(entrypath, pluginContent, real, pluginModule): entrypath.createDirectory() pkg = entrypath.child('plugindummy') pkg.createDirectory() if real: pkg.child('__init__.py').setContent('') plugs = pkg.child('plugins') plugs.createDirectory() if real: plugs.child('__init__.py').setContent(pluginInitFile) plugs.child((pluginModule + '.py')).setContent(pluginContent) return plugs
[ "def", "_createPluginDummy", "(", "entrypath", ",", "pluginContent", ",", "real", ",", "pluginModule", ")", ":", "entrypath", ".", "createDirectory", "(", ")", "pkg", "=", "entrypath", ".", "child", "(", "'plugindummy'", ")", "pkg", ".", "createDirectory", "(", ")", "if", "real", ":", "pkg", ".", "child", "(", "'__init__.py'", ")", ".", "setContent", "(", "''", ")", "plugs", "=", "pkg", ".", "child", "(", "'plugins'", ")", "plugs", ".", "createDirectory", "(", ")", "if", "real", ":", "plugs", ".", "child", "(", "'__init__.py'", ")", ".", "setContent", "(", "pluginInitFile", ")", "plugs", ".", "child", "(", "(", "pluginModule", "+", "'.py'", ")", ")", ".", "setContent", "(", "pluginContent", ")", "return", "plugs" ]
create a plugindummy package .
train
false
25,278
def one(method): def loop(method, self, *args, **kwargs): result = [method(rec, *args, **kwargs) for rec in self] return aggregate(method, result, self) wrapper = decorator(loop, method) wrapper._api = 'one' return wrapper
[ "def", "one", "(", "method", ")", ":", "def", "loop", "(", "method", ",", "self", ",", "*", "args", ",", "**", "kwargs", ")", ":", "result", "=", "[", "method", "(", "rec", ",", "*", "args", ",", "**", "kwargs", ")", "for", "rec", "in", "self", "]", "return", "aggregate", "(", "method", ",", "result", ",", "self", ")", "wrapper", "=", "decorator", "(", "loop", ",", "method", ")", "wrapper", ".", "_api", "=", "'one'", "return", "wrapper" ]
one returns a single game matching the given criteria .
train
false
25,279
def contains_exact_files(directory, expected_files, ignore_links=False): return (sorted(expected_files) == sorted(exact_files(directory, ignore_links=ignore_links)))
[ "def", "contains_exact_files", "(", "directory", ",", "expected_files", ",", "ignore_links", "=", "False", ")", ":", "return", "(", "sorted", "(", "expected_files", ")", "==", "sorted", "(", "exact_files", "(", "directory", ",", "ignore_links", "=", "ignore_links", ")", ")", ")" ]
check if the only files which directory contains are expected_files .
train
false