id_within_dataset
int64
1
55.5k
snippet
stringlengths
19
14.2k
tokens
listlengths
6
1.63k
nl
stringlengths
6
352
split_within_dataset
stringclasses
1 value
is_duplicated
bool
2 classes
39,171
def vms(nictag): ret = {} nictagadm = _check_nictagadm() cmd = '{nictagadm} vms {nictag}'.format(nictagadm=nictagadm, nictag=nictag) res = __salt__['cmd.run_all'](cmd) retcode = res['retcode'] if (retcode != 0): ret['Error'] = (res['stderr'] if ('stderr' in res) else 'Failed to get list of vms.') else: ret = res['stdout'].splitlines() return ret
[ "def", "vms", "(", "nictag", ")", ":", "ret", "=", "{", "}", "nictagadm", "=", "_check_nictagadm", "(", ")", "cmd", "=", "'{nictagadm} vms {nictag}'", ".", "format", "(", "nictagadm", "=", "nictagadm", ",", "nictag", "=", "nictag", ")", "res", "=", "__salt__", "[", "'cmd.run_all'", "]", "(", "cmd", ")", "retcode", "=", "res", "[", "'retcode'", "]", "if", "(", "retcode", "!=", "0", ")", ":", "ret", "[", "'Error'", "]", "=", "(", "res", "[", "'stderr'", "]", "if", "(", "'stderr'", "in", "res", ")", "else", "'Failed to get list of vms.'", ")", "else", ":", "ret", "=", "res", "[", "'stdout'", "]", ".", "splitlines", "(", ")", "return", "ret" ]
list all vms connect to nictag nictag : string name of nictag cli example: .
train
true
39,172
def ilshift(a, b): a <<= b return a
[ "def", "ilshift", "(", "a", ",", "b", ")", ":", "a", "<<=", "b", "return", "a" ]
same as a <<= b .
train
false
39,173
def ipv6(value, options=None): return _ipv_filter(value, 6, options=options)
[ "def", "ipv6", "(", "value", ",", "options", "=", "None", ")", ":", "return", "_ipv_filter", "(", "value", ",", "6", ",", "options", "=", "options", ")" ]
filters a list and returns ipv6 values only .
train
false
39,174
def stSpectralFlux(X, Xprev): sumX = numpy.sum((X + eps)) sumPrevX = numpy.sum((Xprev + eps)) F = numpy.sum((((X / sumX) - (Xprev / sumPrevX)) ** 2)) return F
[ "def", "stSpectralFlux", "(", "X", ",", "Xprev", ")", ":", "sumX", "=", "numpy", ".", "sum", "(", "(", "X", "+", "eps", ")", ")", "sumPrevX", "=", "numpy", ".", "sum", "(", "(", "Xprev", "+", "eps", ")", ")", "F", "=", "numpy", ".", "sum", "(", "(", "(", "(", "X", "/", "sumX", ")", "-", "(", "Xprev", "/", "sumPrevX", ")", ")", "**", "2", ")", ")", "return", "F" ]
computes the spectral flux feature of the current frame arguments: x: the abs of the current frame xpre: the abs of the previous frame .
train
true
39,175
def _choose_tagged_tests(tests, tags): selected = [] tags = set(tags) for test in _flatten_suite(tests): assert isinstance(test, unittest.TestCase) func = getattr(test, test._testMethodName) try: func = func.im_func except AttributeError: pass try: if (func.tags & tags): selected.append(test) except AttributeError: pass return unittest.TestSuite(selected)
[ "def", "_choose_tagged_tests", "(", "tests", ",", "tags", ")", ":", "selected", "=", "[", "]", "tags", "=", "set", "(", "tags", ")", "for", "test", "in", "_flatten_suite", "(", "tests", ")", ":", "assert", "isinstance", "(", "test", ",", "unittest", ".", "TestCase", ")", "func", "=", "getattr", "(", "test", ",", "test", ".", "_testMethodName", ")", "try", ":", "func", "=", "func", ".", "im_func", "except", "AttributeError", ":", "pass", "try", ":", "if", "(", "func", ".", "tags", "&", "tags", ")", ":", "selected", ".", "append", "(", "test", ")", "except", "AttributeError", ":", "pass", "return", "unittest", ".", "TestSuite", "(", "selected", ")" ]
select tests that are tagged with at least one of the given tags .
train
false
39,176
def _clientJobsDB(): return cjdao.ClientJobsDAO.get()
[ "def", "_clientJobsDB", "(", ")", ":", "return", "cjdao", ".", "ClientJobsDAO", ".", "get", "(", ")" ]
returns: the shared cjdao .
train
false
39,177
def wait_for_stack_status(stack_id, target_status, aws_config): def predicate(): stack_report = get_stack_report(stack_id, aws_config) current_status = stack_report['StackStatus'] Message.log(function='wait_for_stack_status', stack_id=stack_id, target_status=target_status, current_status=current_status) if (current_status == target_status): return stack_report return loop_until(reactor, predicate, repeat(10, 120))
[ "def", "wait_for_stack_status", "(", "stack_id", ",", "target_status", ",", "aws_config", ")", ":", "def", "predicate", "(", ")", ":", "stack_report", "=", "get_stack_report", "(", "stack_id", ",", "aws_config", ")", "current_status", "=", "stack_report", "[", "'StackStatus'", "]", "Message", ".", "log", "(", "function", "=", "'wait_for_stack_status'", ",", "stack_id", "=", "stack_id", ",", "target_status", "=", "target_status", ",", "current_status", "=", "current_status", ")", "if", "(", "current_status", "==", "target_status", ")", ":", "return", "stack_report", "return", "loop_until", "(", "reactor", ",", "predicate", ",", "repeat", "(", "10", ",", "120", ")", ")" ]
poll the status of a cloudformation stack .
train
false
39,179
@snippet def client_list_jobs(client, _): def do_something_with(_): pass job_iterator = client.list_jobs() for job in job_iterator: do_something_with(job)
[ "@", "snippet", "def", "client_list_jobs", "(", "client", ",", "_", ")", ":", "def", "do_something_with", "(", "_", ")", ":", "pass", "job_iterator", "=", "client", ".", "list_jobs", "(", ")", "for", "job", "in", "job_iterator", ":", "do_something_with", "(", "job", ")" ]
list jobs for a project .
train
false
39,180
def _get_subproject_base(subproject): entrypoint = _get_installed_entrypoint(subproject) return entrypoint.module_name.split('.')[0]
[ "def", "_get_subproject_base", "(", "subproject", ")", ":", "entrypoint", "=", "_get_installed_entrypoint", "(", "subproject", ")", "return", "entrypoint", ".", "module_name", ".", "split", "(", "'.'", ")", "[", "0", "]" ]
get the import base name for the installed subproject .
train
false
39,181
def check_user_support(user_obj): if (not is_authenticated(user_obj)): if (settings.ANONYMOUS_USER_NAME is None): return (False, user_obj) User = get_user_model() lookup = {User.USERNAME_FIELD: settings.ANONYMOUS_USER_NAME} user_obj = User.objects.get(**lookup) return (True, user_obj)
[ "def", "check_user_support", "(", "user_obj", ")", ":", "if", "(", "not", "is_authenticated", "(", "user_obj", ")", ")", ":", "if", "(", "settings", ".", "ANONYMOUS_USER_NAME", "is", "None", ")", ":", "return", "(", "False", ",", "user_obj", ")", "User", "=", "get_user_model", "(", ")", "lookup", "=", "{", "User", ".", "USERNAME_FIELD", ":", "settings", ".", "ANONYMOUS_USER_NAME", "}", "user_obj", "=", "User", ".", "objects", ".", "get", "(", "**", "lookup", ")", "return", "(", "True", ",", "user_obj", ")" ]
returns a tuple of checkresult and user_obj which should be used for permission checks checks if the given user is supported .
train
false
39,182
def mock_open(mock=None, read_data=''): def _readlines_side_effect(*args, **kwargs): if (handle.readlines.return_value is not None): return handle.readlines.return_value return list(_data) def _read_side_effect(*args, **kwargs): if (handle.read.return_value is not None): return handle.read.return_value return ''.join(_data) def _readline_side_effect(): if (handle.readline.return_value is not None): while True: (yield handle.readline.return_value) for line in _data: (yield line) global file_spec if (file_spec is None): import _io file_spec = list(set(dir(_io.TextIOWrapper)).union(set(dir(_io.BytesIO)))) if (mock is None): mock = MagicMock(name='open', spec=open) handle = MagicMock(spec=file_spec) handle.__enter__.return_value = handle _data = _iterate_read_data(read_data) handle.write.return_value = None handle.read.return_value = None handle.readline.return_value = None handle.readlines.return_value = None handle.read.side_effect = _read_side_effect handle.readline.side_effect = _readline_side_effect() handle.readlines.side_effect = _readlines_side_effect mock.return_value = handle return mock
[ "def", "mock_open", "(", "mock", "=", "None", ",", "read_data", "=", "''", ")", ":", "def", "_readlines_side_effect", "(", "*", "args", ",", "**", "kwargs", ")", ":", "if", "(", "handle", ".", "readlines", ".", "return_value", "is", "not", "None", ")", ":", "return", "handle", ".", "readlines", ".", "return_value", "return", "list", "(", "_data", ")", "def", "_read_side_effect", "(", "*", "args", ",", "**", "kwargs", ")", ":", "if", "(", "handle", ".", "read", ".", "return_value", "is", "not", "None", ")", ":", "return", "handle", ".", "read", ".", "return_value", "return", "''", ".", "join", "(", "_data", ")", "def", "_readline_side_effect", "(", ")", ":", "if", "(", "handle", ".", "readline", ".", "return_value", "is", "not", "None", ")", ":", "while", "True", ":", "(", "yield", "handle", ".", "readline", ".", "return_value", ")", "for", "line", "in", "_data", ":", "(", "yield", "line", ")", "global", "file_spec", "if", "(", "file_spec", "is", "None", ")", ":", "import", "_io", "file_spec", "=", "list", "(", "set", "(", "dir", "(", "_io", ".", "TextIOWrapper", ")", ")", ".", "union", "(", "set", "(", "dir", "(", "_io", ".", "BytesIO", ")", ")", ")", ")", "if", "(", "mock", "is", "None", ")", ":", "mock", "=", "MagicMock", "(", "name", "=", "'open'", ",", "spec", "=", "open", ")", "handle", "=", "MagicMock", "(", "spec", "=", "file_spec", ")", "handle", ".", "__enter__", ".", "return_value", "=", "handle", "_data", "=", "_iterate_read_data", "(", "read_data", ")", "handle", ".", "write", ".", "return_value", "=", "None", "handle", ".", "read", ".", "return_value", "=", "None", "handle", ".", "readline", ".", "return_value", "=", "None", "handle", ".", "readlines", ".", "return_value", "=", "None", "handle", ".", "read", ".", "side_effect", "=", "_read_side_effect", "handle", ".", "readline", ".", "side_effect", "=", "_readline_side_effect", "(", ")", "handle", ".", "readlines", ".", "side_effect", "=", "_readlines_side_effect", "mock", ".", "return_value", "=", "handle", "return", "mock" ]
a helper function to create a mock to replace the use of open .
train
true
39,184
def kivy_configure(): for callback in __kivy_post_configuration: callback()
[ "def", "kivy_configure", "(", ")", ":", "for", "callback", "in", "__kivy_post_configuration", ":", "callback", "(", ")" ]
call post-configuration of kivy .
train
false
39,185
def safe_get_user_model(): (user_app, user_model) = settings.AUTH_USER_MODEL.split('.') return apps.get_registered_model(user_app, user_model)
[ "def", "safe_get_user_model", "(", ")", ":", "(", "user_app", ",", "user_model", ")", "=", "settings", ".", "AUTH_USER_MODEL", ".", "split", "(", "'.'", ")", "return", "apps", ".", "get_registered_model", "(", "user_app", ",", "user_model", ")" ]
safe loading of the user model .
train
true
39,186
def post_structure(entry, site): author = entry.authors.all()[0] return {'title': entry.title, 'description': six.text_type(entry.html_content), 'link': ('%s://%s%s' % (PROTOCOL, site.domain, entry.get_absolute_url())), 'permaLink': ('%s://%s%s' % (PROTOCOL, site.domain, entry.get_absolute_url())), 'categories': [cat.title for cat in entry.categories.all()], 'dateCreated': DateTime(entry.creation_date.isoformat()), 'postid': entry.pk, 'userid': author.get_username(), 'mt_excerpt': entry.excerpt, 'mt_allow_comments': int(entry.comment_enabled), 'mt_allow_pings': (int(entry.pingback_enabled) or int(entry.trackback_enabled)), 'mt_keywords': entry.tags, 'wp_author': author.get_username(), 'wp_author_id': author.pk, 'wp_author_display_name': author.__str__(), 'wp_password': entry.password, 'wp_slug': entry.slug, 'sticky': entry.featured}
[ "def", "post_structure", "(", "entry", ",", "site", ")", ":", "author", "=", "entry", ".", "authors", ".", "all", "(", ")", "[", "0", "]", "return", "{", "'title'", ":", "entry", ".", "title", ",", "'description'", ":", "six", ".", "text_type", "(", "entry", ".", "html_content", ")", ",", "'link'", ":", "(", "'%s://%s%s'", "%", "(", "PROTOCOL", ",", "site", ".", "domain", ",", "entry", ".", "get_absolute_url", "(", ")", ")", ")", ",", "'permaLink'", ":", "(", "'%s://%s%s'", "%", "(", "PROTOCOL", ",", "site", ".", "domain", ",", "entry", ".", "get_absolute_url", "(", ")", ")", ")", ",", "'categories'", ":", "[", "cat", ".", "title", "for", "cat", "in", "entry", ".", "categories", ".", "all", "(", ")", "]", ",", "'dateCreated'", ":", "DateTime", "(", "entry", ".", "creation_date", ".", "isoformat", "(", ")", ")", ",", "'postid'", ":", "entry", ".", "pk", ",", "'userid'", ":", "author", ".", "get_username", "(", ")", ",", "'mt_excerpt'", ":", "entry", ".", "excerpt", ",", "'mt_allow_comments'", ":", "int", "(", "entry", ".", "comment_enabled", ")", ",", "'mt_allow_pings'", ":", "(", "int", "(", "entry", ".", "pingback_enabled", ")", "or", "int", "(", "entry", ".", "trackback_enabled", ")", ")", ",", "'mt_keywords'", ":", "entry", ".", "tags", ",", "'wp_author'", ":", "author", ".", "get_username", "(", ")", ",", "'wp_author_id'", ":", "author", ".", "pk", ",", "'wp_author_display_name'", ":", "author", ".", "__str__", "(", ")", ",", "'wp_password'", ":", "entry", ".", "password", ",", "'wp_slug'", ":", "entry", ".", "slug", ",", "'sticky'", ":", "entry", ".", "featured", "}" ]
a post structure with extensions .
train
true
39,188
def _composed(*decorators): def final_decorator(f): for d in decorators: if (not hasattr(d, '__code__')): setattr(d, '__code__', Fakecode()) f = d(f) return f return final_decorator
[ "def", "_composed", "(", "*", "decorators", ")", ":", "def", "final_decorator", "(", "f", ")", ":", "for", "d", "in", "decorators", ":", "if", "(", "not", "hasattr", "(", "d", ",", "'__code__'", ")", ")", ":", "setattr", "(", "d", ",", "'__code__'", ",", "Fakecode", "(", ")", ")", "f", "=", "d", "(", "f", ")", "return", "f", "return", "final_decorator" ]
takes a list of decorators and returns a single decorator .
train
false
39,190
@safe_filter(error_output=u'auto') @register.filter def background_margin(file_, geometry_string): if ((not file_) or sorl_settings.THUMBNAIL_DUMMY): return u'auto' margin = [0, 0] image_file = default.kvstore.get_or_set(ImageFile(file_)) (x, y) = parse_geometry(geometry_string, image_file.ratio) ex = (x - image_file.x) margin[0] = (ex / 2) ey = (y - image_file.y) margin[1] = (ey / 2) return u' '.join([(u'%spx' % n) for n in margin])
[ "@", "safe_filter", "(", "error_output", "=", "u'auto'", ")", "@", "register", ".", "filter", "def", "background_margin", "(", "file_", ",", "geometry_string", ")", ":", "if", "(", "(", "not", "file_", ")", "or", "sorl_settings", ".", "THUMBNAIL_DUMMY", ")", ":", "return", "u'auto'", "margin", "=", "[", "0", ",", "0", "]", "image_file", "=", "default", ".", "kvstore", ".", "get_or_set", "(", "ImageFile", "(", "file_", ")", ")", "(", "x", ",", "y", ")", "=", "parse_geometry", "(", "geometry_string", ",", "image_file", ".", "ratio", ")", "ex", "=", "(", "x", "-", "image_file", ".", "x", ")", "margin", "[", "0", "]", "=", "(", "ex", "/", "2", ")", "ey", "=", "(", "y", "-", "image_file", ".", "y", ")", "margin", "[", "1", "]", "=", "(", "ey", "/", "2", ")", "return", "u' '", ".", "join", "(", "[", "(", "u'%spx'", "%", "n", ")", "for", "n", "in", "margin", "]", ")" ]
returns the calculated margin for a background image and geometry .
train
false
39,191
def upretty(expr): return xpretty(expr, use_unicode=True, wrap_line=False)
[ "def", "upretty", "(", "expr", ")", ":", "return", "xpretty", "(", "expr", ",", "use_unicode", "=", "True", ",", "wrap_line", "=", "False", ")" ]
unicode pretty-printing .
train
false
39,192
def _FilterExcludedFiles(filenames): exclude_paths = [os.path.abspath(f) for f in _excludes] return [f for f in filenames if (os.path.abspath(f) not in exclude_paths)]
[ "def", "_FilterExcludedFiles", "(", "filenames", ")", ":", "exclude_paths", "=", "[", "os", ".", "path", ".", "abspath", "(", "f", ")", "for", "f", "in", "_excludes", "]", "return", "[", "f", "for", "f", "in", "filenames", "if", "(", "os", ".", "path", ".", "abspath", "(", "f", ")", "not", "in", "exclude_paths", ")", "]" ]
filters out files listed in the --exclude command line switch .
train
true
39,194
def sdm_sort(f, O): return sorted(f, key=(lambda term: O(term[0])), reverse=True)
[ "def", "sdm_sort", "(", "f", ",", "O", ")", ":", "return", "sorted", "(", "f", ",", "key", "=", "(", "lambda", "term", ":", "O", "(", "term", "[", "0", "]", ")", ")", ",", "reverse", "=", "True", ")" ]
sort terms in f using the given monomial order o .
train
false
39,196
def p_item_expr(p): p[0] = ('', p[1])
[ "def", "p_item_expr", "(", "p", ")", ":", "p", "[", "0", "]", "=", "(", "''", ",", "p", "[", "1", "]", ")" ]
pitem : expr .
train
false
39,197
def _psrdp(cmd): rdp = '$RDP = Get-WmiObject -Class Win32_TerminalServiceSetting -Namespace root\\CIMV2\\TerminalServices -Computer . -Authentication 6 -ErrorAction Stop' return __salt__['cmd.run']('{0} ; {1}'.format(rdp, cmd), shell='powershell', python_shell=True)
[ "def", "_psrdp", "(", "cmd", ")", ":", "rdp", "=", "'$RDP = Get-WmiObject -Class Win32_TerminalServiceSetting -Namespace root\\\\CIMV2\\\\TerminalServices -Computer . -Authentication 6 -ErrorAction Stop'", "return", "__salt__", "[", "'cmd.run'", "]", "(", "'{0} ; {1}'", ".", "format", "(", "rdp", ",", "cmd", ")", ",", "shell", "=", "'powershell'", ",", "python_shell", "=", "True", ")" ]
create a win32_terminalservicesetting wmi object as $rdp and execute the command cmd returns the stdout of the command .
train
true
39,198
def test_flattener_layer_state_separation_for_conv(): conv1 = ConvElemwise(8, [2, 2], 'sf1', SigmoidConvNonlinearity(), 0.1) conv2 = ConvElemwise(8, [2, 2], 'sf2', SigmoidConvNonlinearity(), 0.1) mlp = MLP(layers=[FlattenerLayer(CompositeLayer('comp', [conv1, conv2]))], input_space=Conv2DSpace(shape=[5, 5], num_channels=2)) topo_view = np.random.rand(10, 5, 5, 2).astype(theano.config.floatX) y = np.random.rand(10, 256).astype(theano.config.floatX) dataset = DenseDesignMatrix(topo_view=topo_view, y=y) train = Train(dataset, mlp, SGD(0.1, batch_size=5, monitoring_dataset=dataset)) train.algorithm.termination_criterion = EpochCounter(1) train.main_loop()
[ "def", "test_flattener_layer_state_separation_for_conv", "(", ")", ":", "conv1", "=", "ConvElemwise", "(", "8", ",", "[", "2", ",", "2", "]", ",", "'sf1'", ",", "SigmoidConvNonlinearity", "(", ")", ",", "0.1", ")", "conv2", "=", "ConvElemwise", "(", "8", ",", "[", "2", ",", "2", "]", ",", "'sf2'", ",", "SigmoidConvNonlinearity", "(", ")", ",", "0.1", ")", "mlp", "=", "MLP", "(", "layers", "=", "[", "FlattenerLayer", "(", "CompositeLayer", "(", "'comp'", ",", "[", "conv1", ",", "conv2", "]", ")", ")", "]", ",", "input_space", "=", "Conv2DSpace", "(", "shape", "=", "[", "5", ",", "5", "]", ",", "num_channels", "=", "2", ")", ")", "topo_view", "=", "np", ".", "random", ".", "rand", "(", "10", ",", "5", ",", "5", ",", "2", ")", ".", "astype", "(", "theano", ".", "config", ".", "floatX", ")", "y", "=", "np", ".", "random", ".", "rand", "(", "10", ",", "256", ")", ".", "astype", "(", "theano", ".", "config", ".", "floatX", ")", "dataset", "=", "DenseDesignMatrix", "(", "topo_view", "=", "topo_view", ",", "y", "=", "y", ")", "train", "=", "Train", "(", "dataset", ",", "mlp", ",", "SGD", "(", "0.1", ",", "batch_size", "=", "5", ",", "monitoring_dataset", "=", "dataset", ")", ")", "train", ".", "algorithm", ".", "termination_criterion", "=", "EpochCounter", "(", "1", ")", "train", ".", "main_loop", "(", ")" ]
creates a compositelayer wrapping two conv layers and ensures that state gets correctly picked apart .
train
false
39,199
def combine_stains(stains, conv_matrix): from ..exposure import rescale_intensity stains = dtype.img_as_float(stains) logrgb2 = np.dot((- np.reshape(stains, ((-1), 3))), conv_matrix) rgb2 = np.exp(logrgb2) return rescale_intensity(np.reshape((rgb2 - 2), stains.shape), in_range=((-1), 1))
[ "def", "combine_stains", "(", "stains", ",", "conv_matrix", ")", ":", "from", ".", ".", "exposure", "import", "rescale_intensity", "stains", "=", "dtype", ".", "img_as_float", "(", "stains", ")", "logrgb2", "=", "np", ".", "dot", "(", "(", "-", "np", ".", "reshape", "(", "stains", ",", "(", "(", "-", "1", ")", ",", "3", ")", ")", ")", ",", "conv_matrix", ")", "rgb2", "=", "np", ".", "exp", "(", "logrgb2", ")", "return", "rescale_intensity", "(", "np", ".", "reshape", "(", "(", "rgb2", "-", "2", ")", ",", "stains", ".", "shape", ")", ",", "in_range", "=", "(", "(", "-", "1", ")", ",", "1", ")", ")" ]
stain to rgb color space conversion .
train
false
39,200
def getBooleanFromDictionary(dictionary, key): return getBooleanFromDictionaryDefault(True, dictionary, key)
[ "def", "getBooleanFromDictionary", "(", "dictionary", ",", "key", ")", ":", "return", "getBooleanFromDictionaryDefault", "(", "True", ",", "dictionary", ",", "key", ")" ]
get boolean from the dictionary and key .
train
false
39,201
def get_scalar_constant_value(v): if (('sparse' in globals()) and isinstance(v.type, sparse.SparseType)): if ((v.owner is not None) and isinstance(v.owner.op, sparse.CSM)): data = v.owner.inputs[0] return tensor.get_scalar_constant_value(data) return tensor.get_scalar_constant_value(v)
[ "def", "get_scalar_constant_value", "(", "v", ")", ":", "if", "(", "(", "'sparse'", "in", "globals", "(", ")", ")", "and", "isinstance", "(", "v", ".", "type", ",", "sparse", ".", "SparseType", ")", ")", ":", "if", "(", "(", "v", ".", "owner", "is", "not", "None", ")", "and", "isinstance", "(", "v", ".", "owner", ".", "op", ",", "sparse", ".", "CSM", ")", ")", ":", "data", "=", "v", ".", "owner", ".", "inputs", "[", "0", "]", "return", "tensor", ".", "get_scalar_constant_value", "(", "data", ")", "return", "tensor", ".", "get_scalar_constant_value", "(", "v", ")" ]
return the constant scalar value underlying variable v if v is the output of dimshuffles .
train
false
39,202
@snippet def client_run_sync_query_timeout(client, _): TIMEOUT_MS = 10 all_rows = [] def do_something_with(rows): all_rows.extend(rows) query = client.run_sync_query(QUERY) query.timeout_ms = TIMEOUT_MS query.use_query_cache = False query.run() assert (not query.complete) job = query.job job.reload() retry_count = 0 while ((retry_count < 10) and (job.state != u'DONE')): time.sleep((1.5 ** retry_count)) retry_count += 1 job.reload() assert (job.state == u'DONE') (rows, total_count, token) = query.fetch_data() while True: do_something_with(rows) if (token is None): break (rows, total_count, token) = query.fetch_data(page_token=token) assert (len(all_rows) == total_count)
[ "@", "snippet", "def", "client_run_sync_query_timeout", "(", "client", ",", "_", ")", ":", "TIMEOUT_MS", "=", "10", "all_rows", "=", "[", "]", "def", "do_something_with", "(", "rows", ")", ":", "all_rows", ".", "extend", "(", "rows", ")", "query", "=", "client", ".", "run_sync_query", "(", "QUERY", ")", "query", ".", "timeout_ms", "=", "TIMEOUT_MS", "query", ".", "use_query_cache", "=", "False", "query", ".", "run", "(", ")", "assert", "(", "not", "query", ".", "complete", ")", "job", "=", "query", ".", "job", "job", ".", "reload", "(", ")", "retry_count", "=", "0", "while", "(", "(", "retry_count", "<", "10", ")", "and", "(", "job", ".", "state", "!=", "u'DONE'", ")", ")", ":", "time", ".", "sleep", "(", "(", "1.5", "**", "retry_count", ")", ")", "retry_count", "+=", "1", "job", ".", "reload", "(", ")", "assert", "(", "job", ".", "state", "==", "u'DONE'", ")", "(", "rows", ",", "total_count", ",", "token", ")", "=", "query", ".", "fetch_data", "(", ")", "while", "True", ":", "do_something_with", "(", "rows", ")", "if", "(", "token", "is", "None", ")", ":", "break", "(", "rows", ",", "total_count", ",", "token", ")", "=", "query", ".", "fetch_data", "(", "page_token", "=", "token", ")", "assert", "(", "len", "(", "all_rows", ")", "==", "total_count", ")" ]
run a synchronous query w/ timeout .
train
false
39,204
def attachment_specs_update_or_create(context, attachment_id, specs): return IMPL.attachment_specs_update_or_create(context, attachment_id, specs)
[ "def", "attachment_specs_update_or_create", "(", "context", ",", "attachment_id", ",", "specs", ")", ":", "return", "IMPL", ".", "attachment_specs_update_or_create", "(", "context", ",", "attachment_id", ",", "specs", ")" ]
create or update attachment specs .
train
false
39,205
def rs_atanh(p, x, prec): if rs_is_puiseux(p, x): return rs_puiseux(rs_atanh, p, x, prec) R = p.ring const = 0 if _has_constant_term(p, x): zm = R.zero_monom c = p[zm] if (R.domain is EX): c_expr = c.as_expr() const = atanh(c_expr) elif isinstance(c, PolyElement): try: c_expr = c.as_expr() const = R(atanh(c_expr)) except ValueError: raise DomainError("The given series can't be expanded in this domain.") else: try: const = R(atanh(c)) except ValueError: raise DomainError("The given series can't be expanded in this domain.") dp = rs_diff(p, x) p1 = ((- rs_square(p, x, prec)) + 1) p1 = rs_series_inversion(p1, x, (prec - 1)) p1 = rs_mul(dp, p1, x, (prec - 1)) return (rs_integrate(p1, x) + const)
[ "def", "rs_atanh", "(", "p", ",", "x", ",", "prec", ")", ":", "if", "rs_is_puiseux", "(", "p", ",", "x", ")", ":", "return", "rs_puiseux", "(", "rs_atanh", ",", "p", ",", "x", ",", "prec", ")", "R", "=", "p", ".", "ring", "const", "=", "0", "if", "_has_constant_term", "(", "p", ",", "x", ")", ":", "zm", "=", "R", ".", "zero_monom", "c", "=", "p", "[", "zm", "]", "if", "(", "R", ".", "domain", "is", "EX", ")", ":", "c_expr", "=", "c", ".", "as_expr", "(", ")", "const", "=", "atanh", "(", "c_expr", ")", "elif", "isinstance", "(", "c", ",", "PolyElement", ")", ":", "try", ":", "c_expr", "=", "c", ".", "as_expr", "(", ")", "const", "=", "R", "(", "atanh", "(", "c_expr", ")", ")", "except", "ValueError", ":", "raise", "DomainError", "(", "\"The given series can't be expanded in this domain.\"", ")", "else", ":", "try", ":", "const", "=", "R", "(", "atanh", "(", "c", ")", ")", "except", "ValueError", ":", "raise", "DomainError", "(", "\"The given series can't be expanded in this domain.\"", ")", "dp", "=", "rs_diff", "(", "p", ",", "x", ")", "p1", "=", "(", "(", "-", "rs_square", "(", "p", ",", "x", ",", "prec", ")", ")", "+", "1", ")", "p1", "=", "rs_series_inversion", "(", "p1", ",", "x", ",", "(", "prec", "-", "1", ")", ")", "p1", "=", "rs_mul", "(", "dp", ",", "p1", ",", "x", ",", "(", "prec", "-", "1", ")", ")", "return", "(", "rs_integrate", "(", "p1", ",", "x", ")", "+", "const", ")" ]
hyperbolic arctangent of a series return the series expansion of the atanh of p .
train
false
39,207
def AddFieldsToDocumentPb(doc_id, fields, document): if (doc_id is not None): document.set_id(doc_id) for field_tuple in fields: name = field_tuple[0] value = field_tuple[1] field = document.add_field() field.set_name(name) field_value = field.mutable_value() field_value.set_string_value(value) if (len(field_tuple) > 2): field_value.set_type(field_tuple[2])
[ "def", "AddFieldsToDocumentPb", "(", "doc_id", ",", "fields", ",", "document", ")", ":", "if", "(", "doc_id", "is", "not", "None", ")", ":", "document", ".", "set_id", "(", "doc_id", ")", "for", "field_tuple", "in", "fields", ":", "name", "=", "field_tuple", "[", "0", "]", "value", "=", "field_tuple", "[", "1", "]", "field", "=", "document", ".", "add_field", "(", ")", "field", ".", "set_name", "(", "name", ")", "field_value", "=", "field", ".", "mutable_value", "(", ")", "field_value", ".", "set_string_value", "(", "value", ")", "if", "(", "len", "(", "field_tuple", ")", ">", "2", ")", ":", "field_value", ".", "set_type", "(", "field_tuple", "[", "2", "]", ")" ]
add the id and fields to document .
train
false
39,208
def _subs_tree(cls, tvars=None, args=None): if (cls.__origin__ is None): return cls current = cls.__origin__ orig_chain = [] while (current.__origin__ is not None): orig_chain.append(current) current = current.__origin__ tree_args = [] for arg in cls.__args__: tree_args.append(_replace_arg(arg, tvars, args)) for ocls in orig_chain: new_tree_args = [] for (i, arg) in enumerate(ocls.__args__): new_tree_args.append(_replace_arg(arg, ocls.__parameters__, tree_args)) tree_args = new_tree_args return tree_args
[ "def", "_subs_tree", "(", "cls", ",", "tvars", "=", "None", ",", "args", "=", "None", ")", ":", "if", "(", "cls", ".", "__origin__", "is", "None", ")", ":", "return", "cls", "current", "=", "cls", ".", "__origin__", "orig_chain", "=", "[", "]", "while", "(", "current", ".", "__origin__", "is", "not", "None", ")", ":", "orig_chain", ".", "append", "(", "current", ")", "current", "=", "current", ".", "__origin__", "tree_args", "=", "[", "]", "for", "arg", "in", "cls", ".", "__args__", ":", "tree_args", ".", "append", "(", "_replace_arg", "(", "arg", ",", "tvars", ",", "args", ")", ")", "for", "ocls", "in", "orig_chain", ":", "new_tree_args", "=", "[", "]", "for", "(", "i", ",", "arg", ")", "in", "enumerate", "(", "ocls", ".", "__args__", ")", ":", "new_tree_args", ".", "append", "(", "_replace_arg", "(", "arg", ",", "ocls", ".", "__parameters__", ",", "tree_args", ")", ")", "tree_args", "=", "new_tree_args", "return", "tree_args" ]
an internal helper function: calculate substitution tree for generic cls after replacing its type parameters with substitutions in tvars -> args .
train
true
39,209
def _buildTestClasses(_locals): for x in getModule('twisted').walkModules(): ignoredModules = ['twisted.test.reflect_helper', 'twisted.internet.test.process_', 'twisted.test.process_'] isIgnored = [x.name.startswith(ignored) for ignored in ignoredModules] if (True in isIgnored): continue class Test(NewStyleOnly, unittest.TestCase, ): '\n @see: L{NewStyleOnly}\n ' module = x.name acceptableName = x.name.replace('.', '_') Test.__name__ = acceptableName if hasattr(Test, '__qualname__'): Test.__qualname__ = acceptableName _locals.update({acceptableName: Test})
[ "def", "_buildTestClasses", "(", "_locals", ")", ":", "for", "x", "in", "getModule", "(", "'twisted'", ")", ".", "walkModules", "(", ")", ":", "ignoredModules", "=", "[", "'twisted.test.reflect_helper'", ",", "'twisted.internet.test.process_'", ",", "'twisted.test.process_'", "]", "isIgnored", "=", "[", "x", ".", "name", ".", "startswith", "(", "ignored", ")", "for", "ignored", "in", "ignoredModules", "]", "if", "(", "True", "in", "isIgnored", ")", ":", "continue", "class", "Test", "(", "NewStyleOnly", ",", "unittest", ".", "TestCase", ",", ")", ":", "module", "=", "x", ".", "name", "acceptableName", "=", "x", ".", "name", ".", "replace", "(", "'.'", ",", "'_'", ")", "Test", ".", "__name__", "=", "acceptableName", "if", "hasattr", "(", "Test", ",", "'__qualname__'", ")", ":", "Test", ".", "__qualname__", "=", "acceptableName", "_locals", ".", "update", "(", "{", "acceptableName", ":", "Test", "}", ")" ]
build the test classes that use l{newstyleonly} .
train
false
39,210
def sentinel_get_master_ip(master, host=None, port=None, password=None): server = _sconnect(host, port, password) ret = server.sentinel_get_master_addr_by_name(master) return dict(list(zip(('master_host', 'master_port'), ret)))
[ "def", "sentinel_get_master_ip", "(", "master", ",", "host", "=", "None", ",", "port", "=", "None", ",", "password", "=", "None", ")", ":", "server", "=", "_sconnect", "(", "host", ",", "port", ",", "password", ")", "ret", "=", "server", ".", "sentinel_get_master_addr_by_name", "(", "master", ")", "return", "dict", "(", "list", "(", "zip", "(", "(", "'master_host'", ",", "'master_port'", ")", ",", "ret", ")", ")", ")" ]
get ip for sentinel master .
train
true
39,211
def country_unalias(country): if (type(country) in [str, unicode]): return COUNTRY_ALIASES.get(country, country) if hasattr(country, '__iter__'): return [country_unalias(country_elt) for country_elt in country] return country
[ "def", "country_unalias", "(", "country", ")", ":", "if", "(", "type", "(", "country", ")", "in", "[", "str", ",", "unicode", "]", ")", ":", "return", "COUNTRY_ALIASES", ".", "get", "(", "country", ",", "country", ")", "if", "hasattr", "(", "country", ",", "'__iter__'", ")", ":", "return", "[", "country_unalias", "(", "country_elt", ")", "for", "country_elt", "in", "country", "]", "return", "country" ]
takes either a country code and returns either a country code or a list of country codes .
train
false
39,213
@shared_task(bind=True, base=DebugBasketTask, default_retry_delay=BASKET_TASK_RETRY_DELAY, max_retries=BASKET_TASK_MAX_RETRIES) def lookup_user_task(self, email): result = {} try: result = basket.lookup_user(email=email) except MaxRetriesExceededError as exc: raise exc except basket.BasketException as exc: if (not (exc[0] == u'User not found')): raise self.retry(exc=exc) result = exc.result return result
[ "@", "shared_task", "(", "bind", "=", "True", ",", "base", "=", "DebugBasketTask", ",", "default_retry_delay", "=", "BASKET_TASK_RETRY_DELAY", ",", "max_retries", "=", "BASKET_TASK_MAX_RETRIES", ")", "def", "lookup_user_task", "(", "self", ",", "email", ")", ":", "result", "=", "{", "}", "try", ":", "result", "=", "basket", ".", "lookup_user", "(", "email", "=", "email", ")", "except", "MaxRetriesExceededError", "as", "exc", ":", "raise", "exc", "except", "basket", ".", "BasketException", "as", "exc", ":", "if", "(", "not", "(", "exc", "[", "0", "]", "==", "u'User not found'", ")", ")", ":", "raise", "self", ".", "retry", "(", "exc", "=", "exc", ")", "result", "=", "exc", ".", "result", "return", "result" ]
task responsible for getting information about a user in basket .
train
false
39,215
def xl_range_abs(first_row, first_col, last_row, last_col): range1 = xl_rowcol_to_cell(first_row, first_col, True, True) range2 = xl_rowcol_to_cell(last_row, last_col, True, True) return ((range1 + ':') + range2)
[ "def", "xl_range_abs", "(", "first_row", ",", "first_col", ",", "last_row", ",", "last_col", ")", ":", "range1", "=", "xl_rowcol_to_cell", "(", "first_row", ",", "first_col", ",", "True", ",", "True", ")", "range2", "=", "xl_rowcol_to_cell", "(", "last_row", ",", "last_col", ",", "True", ",", "True", ")", "return", "(", "(", "range1", "+", "':'", ")", "+", "range2", ")" ]
convert zero indexed row and col cell references to a $a$1:$b$1 absolute range string .
train
false
39,217
def test_singleton_dim(): image = np.random.rand(1, 20) noisy = random_noise(image, mode='salt', amount=0.1, seed=42) assert (np.sum((noisy == 1)) == 2)
[ "def", "test_singleton_dim", "(", ")", ":", "image", "=", "np", ".", "random", ".", "rand", "(", "1", ",", "20", ")", "noisy", "=", "random_noise", "(", "image", ",", "mode", "=", "'salt'", ",", "amount", "=", "0.1", ",", "seed", "=", "42", ")", "assert", "(", "np", ".", "sum", "(", "(", "noisy", "==", "1", ")", ")", "==", "2", ")" ]
ensure images where size of a given dimension is 1 work correctly .
train
false
39,219
def isUndirected(G): for v in G: if (v in G[v]): return False for w in G[v]: if (v not in G[w]): return False return True
[ "def", "isUndirected", "(", "G", ")", ":", "for", "v", "in", "G", ":", "if", "(", "v", "in", "G", "[", "v", "]", ")", ":", "return", "False", "for", "w", "in", "G", "[", "v", "]", ":", "if", "(", "v", "not", "in", "G", "[", "w", "]", ")", ":", "return", "False", "return", "True" ]
check that g represents a simple undirected graph .
train
false
39,220
def check_exists(name): if (name in checks_list()): log.debug('[uptime] found {0} in checks'.format(name)) return True return False
[ "def", "check_exists", "(", "name", ")", ":", "if", "(", "name", "in", "checks_list", "(", ")", ")", ":", "log", ".", "debug", "(", "'[uptime] found {0} in checks'", ".", "format", "(", "name", ")", ")", "return", "True", "return", "False" ]
check if a given url is in being monitored by uptime cli example: .
train
false
39,221
def apply_item_metadata(item, track_info): item.artist = track_info.artist item.artist_sort = track_info.artist_sort item.artist_credit = track_info.artist_credit item.title = track_info.title item.mb_trackid = track_info.track_id if track_info.artist_id: item.mb_artistid = track_info.artist_id
[ "def", "apply_item_metadata", "(", "item", ",", "track_info", ")", ":", "item", ".", "artist", "=", "track_info", ".", "artist", "item", ".", "artist_sort", "=", "track_info", ".", "artist_sort", "item", ".", "artist_credit", "=", "track_info", ".", "artist_credit", "item", ".", "title", "=", "track_info", ".", "title", "item", ".", "mb_trackid", "=", "track_info", ".", "track_id", "if", "track_info", ".", "artist_id", ":", "item", ".", "mb_artistid", "=", "track_info", ".", "artist_id" ]
set an items metadata from its matched trackinfo object .
train
false
39,222
def quote_ps_string(s): s = s.replace('\\', '\\\\') s = s.replace('(', '\\(') s = s.replace(')', '\\)') s = s.replace("'", '\\251') s = s.replace('`', '\\301') s = re.sub('[^ -~\\n]', (lambda x: ('\\%03o' % ord(x.group()))), s) return s
[ "def", "quote_ps_string", "(", "s", ")", ":", "s", "=", "s", ".", "replace", "(", "'\\\\'", ",", "'\\\\\\\\'", ")", "s", "=", "s", ".", "replace", "(", "'('", ",", "'\\\\('", ")", "s", "=", "s", ".", "replace", "(", "')'", ",", "'\\\\)'", ")", "s", "=", "s", ".", "replace", "(", "\"'\"", ",", "'\\\\251'", ")", "s", "=", "s", ".", "replace", "(", "'`'", ",", "'\\\\301'", ")", "s", "=", "re", ".", "sub", "(", "'[^ -~\\\\n]'", ",", "(", "lambda", "x", ":", "(", "'\\\\%03o'", "%", "ord", "(", "x", ".", "group", "(", ")", ")", ")", ")", ",", "s", ")", "return", "s" ]
quote dangerous characters of s for use in a postscript string constant .
train
false
39,223
def load_endpoint_json(path): with open(path, 'r') as endpoints_file: return json.load(endpoints_file)
[ "def", "load_endpoint_json", "(", "path", ")", ":", "with", "open", "(", "path", ",", "'r'", ")", "as", "endpoints_file", ":", "return", "json", ".", "load", "(", "endpoints_file", ")" ]
loads a given json file & returns it .
train
false
39,224
def test_no_ready_python_process(noready_pyproc): with pytest.raises(testprocess.ProcessExited): with stopwatch(max_ms=5000): noready_pyproc.start()
[ "def", "test_no_ready_python_process", "(", "noready_pyproc", ")", ":", "with", "pytest", ".", "raises", "(", "testprocess", ".", "ProcessExited", ")", ":", "with", "stopwatch", "(", "max_ms", "=", "5000", ")", ":", "noready_pyproc", ".", "start", "(", ")" ]
when a process quits immediately .
train
false
39,227
def export_library_to_xml(modulestore, contentstore, library_key, root_dir, library_dir): LibraryExportManager(modulestore, contentstore, library_key, root_dir, library_dir).export()
[ "def", "export_library_to_xml", "(", "modulestore", ",", "contentstore", ",", "library_key", ",", "root_dir", ",", "library_dir", ")", ":", "LibraryExportManager", "(", "modulestore", ",", "contentstore", ",", "library_key", ",", "root_dir", ",", "library_dir", ")", ".", "export", "(", ")" ]
thin wrapper for the library export manager .
train
false
39,228
@image_comparison(baseline_images=[u'hatching_legend'], extensions=[u'pdf']) def test_hatching_legend(): fig = plt.figure(figsize=(1, 2)) a = plt.Rectangle([0, 0], 0, 0, facecolor=u'green', hatch=u'XXXX') b = plt.Rectangle([0, 0], 0, 0, facecolor=u'blue', hatch=u'XXXX') fig.legend([a, b, a, b], [u'', u'', u'', u''])
[ "@", "image_comparison", "(", "baseline_images", "=", "[", "u'hatching_legend'", "]", ",", "extensions", "=", "[", "u'pdf'", "]", ")", "def", "test_hatching_legend", "(", ")", ":", "fig", "=", "plt", ".", "figure", "(", "figsize", "=", "(", "1", ",", "2", ")", ")", "a", "=", "plt", ".", "Rectangle", "(", "[", "0", ",", "0", "]", ",", "0", ",", "0", ",", "facecolor", "=", "u'green'", ",", "hatch", "=", "u'XXXX'", ")", "b", "=", "plt", ".", "Rectangle", "(", "[", "0", ",", "0", "]", ",", "0", ",", "0", ",", "facecolor", "=", "u'blue'", ",", "hatch", "=", "u'XXXX'", ")", "fig", ".", "legend", "(", "[", "a", ",", "b", ",", "a", ",", "b", "]", ",", "[", "u''", ",", "u''", ",", "u''", ",", "u''", "]", ")" ]
test for correct hatching on patches in legend .
train
false
39,230
def abs(a): return _abs(a)
[ "def", "abs", "(", "a", ")", ":", "return", "_abs", "(", "a", ")" ]
returns the absolute url: .
train
false
39,231
def to_bundle_ingest_dirname(ts): return ts.isoformat().replace(':', ';')
[ "def", "to_bundle_ingest_dirname", "(", "ts", ")", ":", "return", "ts", ".", "isoformat", "(", ")", ".", "replace", "(", "':'", ",", "';'", ")" ]
convert a pandas timestamp into the name of the directory for the ingestion .
train
false
39,233
@handle_response_format @treeio_login_required def mlist_delete(request, mlist_id, response_format='html'): mlist = get_object_or_404(MailingList, pk=mlist_id) if (not request.user.profile.has_permission(mlist, mode='w')): return user_denied(request, message="You don't have access to this Mailing List", response_format=response_format) if request.POST: if ('delete' in request.POST): if ('trash' in request.POST): mlist.trash = True mlist.save() else: mlist.delete() return HttpResponseRedirect('/messaging/') elif ('cancel' in request.POST): return HttpResponseRedirect(reverse('messaging_mlist_view', args=[mlist.id])) context = _get_default_context(request) context.update({'mlist': mlist}) return render_to_response('messaging/mlist_delete', context, context_instance=RequestContext(request), response_format=response_format)
[ "@", "handle_response_format", "@", "treeio_login_required", "def", "mlist_delete", "(", "request", ",", "mlist_id", ",", "response_format", "=", "'html'", ")", ":", "mlist", "=", "get_object_or_404", "(", "MailingList", ",", "pk", "=", "mlist_id", ")", "if", "(", "not", "request", ".", "user", ".", "profile", ".", "has_permission", "(", "mlist", ",", "mode", "=", "'w'", ")", ")", ":", "return", "user_denied", "(", "request", ",", "message", "=", "\"You don't have access to this Mailing List\"", ",", "response_format", "=", "response_format", ")", "if", "request", ".", "POST", ":", "if", "(", "'delete'", "in", "request", ".", "POST", ")", ":", "if", "(", "'trash'", "in", "request", ".", "POST", ")", ":", "mlist", ".", "trash", "=", "True", "mlist", ".", "save", "(", ")", "else", ":", "mlist", ".", "delete", "(", ")", "return", "HttpResponseRedirect", "(", "'/messaging/'", ")", "elif", "(", "'cancel'", "in", "request", ".", "POST", ")", ":", "return", "HttpResponseRedirect", "(", "reverse", "(", "'messaging_mlist_view'", ",", "args", "=", "[", "mlist", ".", "id", "]", ")", ")", "context", "=", "_get_default_context", "(", "request", ")", "context", ".", "update", "(", "{", "'mlist'", ":", "mlist", "}", ")", "return", "render_to_response", "(", "'messaging/mlist_delete'", ",", "context", ",", "context_instance", "=", "RequestContext", "(", "request", ")", ",", "response_format", "=", "response_format", ")" ]
delete mlist page .
train
false
39,234
def revlookup(name): if (Base.defaults['server'] == []): Base.DiscoverNameServers() a = string.split(name, '.') a.reverse() b = (string.join(a, '.') + '.in-addr.arpa') return Base.DnsRequest(b, qtype='ptr').req().answers[0]['data']
[ "def", "revlookup", "(", "name", ")", ":", "if", "(", "Base", ".", "defaults", "[", "'server'", "]", "==", "[", "]", ")", ":", "Base", ".", "DiscoverNameServers", "(", ")", "a", "=", "string", ".", "split", "(", "name", ",", "'.'", ")", "a", ".", "reverse", "(", ")", "b", "=", "(", "string", ".", "join", "(", "a", ",", "'.'", ")", "+", "'.in-addr.arpa'", ")", "return", "Base", ".", "DnsRequest", "(", "b", ",", "qtype", "=", "'ptr'", ")", ".", "req", "(", ")", ".", "answers", "[", "0", "]", "[", "'data'", "]" ]
convenience routine for doing a reverse lookup of an address .
train
false
39,237
def get_hash_method(hash_method_name): if (hash_method_name not in HASH_METHODS): raise exception.SignatureVerificationError(reason=(_('Invalid signature hash method: %s') % hash_method_name)) return HASH_METHODS[hash_method_name]
[ "def", "get_hash_method", "(", "hash_method_name", ")", ":", "if", "(", "hash_method_name", "not", "in", "HASH_METHODS", ")", ":", "raise", "exception", ".", "SignatureVerificationError", "(", "reason", "=", "(", "_", "(", "'Invalid signature hash method: %s'", ")", "%", "hash_method_name", ")", ")", "return", "HASH_METHODS", "[", "hash_method_name", "]" ]
verify the hash method name and create the hash method .
train
false
39,238
def dmp_gff_list(f, u, K): if (not u): return dup_gff_list(f, K) else: raise MultivariatePolynomialError(f)
[ "def", "dmp_gff_list", "(", "f", ",", "u", ",", "K", ")", ":", "if", "(", "not", "u", ")", ":", "return", "dup_gff_list", "(", "f", ",", "K", ")", "else", ":", "raise", "MultivariatePolynomialError", "(", "f", ")" ]
compute greatest factorial factorization of f in k[x] .
train
false
39,239
def sonatype_clm(registry, xml_parent, data): clm = XML.SubElement(xml_parent, 'com.sonatype.insight.ci.hudson.PreBuildScan') clm.set('plugin', 'sonatype-clm-ci') SUPPORTED_VALUES = ['list', 'manual'] SUPPORTED_STAGES = ['build', 'stage-release', 'release', 'operate'] application_select = XML.SubElement(clm, 'applicationSelectType') application_mappings = [('value', 'value', 'list', SUPPORTED_VALUES), ('application-name', 'applicationId', None)] convert_mapping_to_xml(application_select, data, application_mappings, fail_required=True) path = XML.SubElement(clm, 'pathConfig') path_mappings = [('scan-targets', 'scanTargets', ''), ('module-excludes', 'moduleExcludes', ''), ('advanced-options', 'scanProperties', '')] convert_mapping_to_xml(path, data, path_mappings, fail_required=True) mappings = [('fail-on-clm-server-failure', 'failOnClmServerFailures', False), ('stage', 'stageId', 'build', SUPPORTED_STAGES), ('username', 'username', ''), ('password', 'password', '')] convert_mapping_to_xml(clm, data, mappings, fail_required=True)
[ "def", "sonatype_clm", "(", "registry", ",", "xml_parent", ",", "data", ")", ":", "clm", "=", "XML", ".", "SubElement", "(", "xml_parent", ",", "'com.sonatype.insight.ci.hudson.PreBuildScan'", ")", "clm", ".", "set", "(", "'plugin'", ",", "'sonatype-clm-ci'", ")", "SUPPORTED_VALUES", "=", "[", "'list'", ",", "'manual'", "]", "SUPPORTED_STAGES", "=", "[", "'build'", ",", "'stage-release'", ",", "'release'", ",", "'operate'", "]", "application_select", "=", "XML", ".", "SubElement", "(", "clm", ",", "'applicationSelectType'", ")", "application_mappings", "=", "[", "(", "'value'", ",", "'value'", ",", "'list'", ",", "SUPPORTED_VALUES", ")", ",", "(", "'application-name'", ",", "'applicationId'", ",", "None", ")", "]", "convert_mapping_to_xml", "(", "application_select", ",", "data", ",", "application_mappings", ",", "fail_required", "=", "True", ")", "path", "=", "XML", ".", "SubElement", "(", "clm", ",", "'pathConfig'", ")", "path_mappings", "=", "[", "(", "'scan-targets'", ",", "'scanTargets'", ",", "''", ")", ",", "(", "'module-excludes'", ",", "'moduleExcludes'", ",", "''", ")", ",", "(", "'advanced-options'", ",", "'scanProperties'", ",", "''", ")", "]", "convert_mapping_to_xml", "(", "path", ",", "data", ",", "path_mappings", ",", "fail_required", "=", "True", ")", "mappings", "=", "[", "(", "'fail-on-clm-server-failure'", ",", "'failOnClmServerFailures'", ",", "False", ")", ",", "(", "'stage'", ",", "'stageId'", ",", "'build'", ",", "SUPPORTED_STAGES", ")", ",", "(", "'username'", ",", "'username'", ",", "''", ")", ",", "(", "'password'", ",", "'password'", ",", "''", ")", "]", "convert_mapping_to_xml", "(", "clm", ",", "data", ",", "mappings", ",", "fail_required", "=", "True", ")" ]
yaml: sonatype-clm requires the jenkins :jenkins-wiki:sonatype clm plugin <sonatype+clm+%28formerly+insight+for+ci%29> .
train
false
39,240
def FixVCMacroSlashes(s): if ('$' in s): s = fix_vc_macro_slashes_regex.sub('\\1', s) return s
[ "def", "FixVCMacroSlashes", "(", "s", ")", ":", "if", "(", "'$'", "in", "s", ")", ":", "s", "=", "fix_vc_macro_slashes_regex", ".", "sub", "(", "'\\\\1'", ",", "s", ")", "return", "s" ]
replace macros which have excessive following slashes .
train
false
39,241
@utils.arg('node', metavar='<node>', help=_('ID of node')) def do_baremetal_node_show(cs, args): _emit_deprecation_warning('baremetal-node-show') node = _find_baremetal_node(cs, args.node) _print_baremetal_resource(node)
[ "@", "utils", ".", "arg", "(", "'node'", ",", "metavar", "=", "'<node>'", ",", "help", "=", "_", "(", "'ID of node'", ")", ")", "def", "do_baremetal_node_show", "(", "cs", ",", "args", ")", ":", "_emit_deprecation_warning", "(", "'baremetal-node-show'", ")", "node", "=", "_find_baremetal_node", "(", "cs", ",", "args", ".", "node", ")", "_print_baremetal_resource", "(", "node", ")" ]
deprecated: show information about a baremetal node .
train
false
39,243
def getSingleInfo(table, movieID, infoType, notAList=False): infoTypeID = InfoType.select((InfoType.q.info == infoType)) if (infoTypeID.count() == 0): return {} res = table.select(AND((table.q.movieID == movieID), (table.q.infoTypeID == infoTypeID[0].id))) retList = [] for r in res: info = r.info note = r.note if note: info += (u'::%s' % note) retList.append(info) if (not retList): return {} if (not notAList): return {infoType: retList} else: return {infoType: retList[0]}
[ "def", "getSingleInfo", "(", "table", ",", "movieID", ",", "infoType", ",", "notAList", "=", "False", ")", ":", "infoTypeID", "=", "InfoType", ".", "select", "(", "(", "InfoType", ".", "q", ".", "info", "==", "infoType", ")", ")", "if", "(", "infoTypeID", ".", "count", "(", ")", "==", "0", ")", ":", "return", "{", "}", "res", "=", "table", ".", "select", "(", "AND", "(", "(", "table", ".", "q", ".", "movieID", "==", "movieID", ")", ",", "(", "table", ".", "q", ".", "infoTypeID", "==", "infoTypeID", "[", "0", "]", ".", "id", ")", ")", ")", "retList", "=", "[", "]", "for", "r", "in", "res", ":", "info", "=", "r", ".", "info", "note", "=", "r", ".", "note", "if", "note", ":", "info", "+=", "(", "u'::%s'", "%", "note", ")", "retList", ".", "append", "(", "info", ")", "if", "(", "not", "retList", ")", ":", "return", "{", "}", "if", "(", "not", "notAList", ")", ":", "return", "{", "infoType", ":", "retList", "}", "else", ":", "return", "{", "infoType", ":", "retList", "[", "0", "]", "}" ]
return a dictionary in the form {infotype: infolistorstring} .
train
false
39,244
@pytest.mark.parametrize('encrypt', [True, False]) def test_token_inputs(db, config, encrypt, default_account): config['ENCRYPT_SECRETS'] = encrypt unicode_token = u'myunicodesecret' invalid_token = '\xff\x10' null_token = '\x1f\x00\xf1' default_account.refresh_token = unicode_token db.session.commit() secret_id = default_account.refresh_token_id secret = db.session.query(Secret).get(secret_id) assert (not isinstance(secret.secret, unicode)), 'secret cannot be unicode' assert (secret.secret == unicode_token), 'token not decrypted correctly' with pytest.raises(ValueError) as e: default_account.refresh_token = invalid_token assert (e.typename == 'ValueError'), 'token cannot be invalid UTF-8' with pytest.raises(ValueError) as f: default_account.refresh_token = null_token assert (f.typename == 'ValueError'), 'token cannot contain NULL byte' assert (default_account.refresh_token == unicode_token)
[ "@", "pytest", ".", "mark", ".", "parametrize", "(", "'encrypt'", ",", "[", "True", ",", "False", "]", ")", "def", "test_token_inputs", "(", "db", ",", "config", ",", "encrypt", ",", "default_account", ")", ":", "config", "[", "'ENCRYPT_SECRETS'", "]", "=", "encrypt", "unicode_token", "=", "u'myunicodesecret'", "invalid_token", "=", "'\\xff\\x10'", "null_token", "=", "'\\x1f\\x00\\xf1'", "default_account", ".", "refresh_token", "=", "unicode_token", "db", ".", "session", ".", "commit", "(", ")", "secret_id", "=", "default_account", ".", "refresh_token_id", "secret", "=", "db", ".", "session", ".", "query", "(", "Secret", ")", ".", "get", "(", "secret_id", ")", "assert", "(", "not", "isinstance", "(", "secret", ".", "secret", ",", "unicode", ")", ")", ",", "'secret cannot be unicode'", "assert", "(", "secret", ".", "secret", "==", "unicode_token", ")", ",", "'token not decrypted correctly'", "with", "pytest", ".", "raises", "(", "ValueError", ")", "as", "e", ":", "default_account", ".", "refresh_token", "=", "invalid_token", "assert", "(", "e", ".", "typename", "==", "'ValueError'", ")", ",", "'token cannot be invalid UTF-8'", "with", "pytest", ".", "raises", "(", "ValueError", ")", "as", "f", ":", "default_account", ".", "refresh_token", "=", "null_token", "assert", "(", "f", ".", "typename", "==", "'ValueError'", ")", ",", "'token cannot contain NULL byte'", "assert", "(", "default_account", ".", "refresh_token", "==", "unicode_token", ")" ]
ensure unicode tokens are converted to bytes .
train
false
39,245
def validate_settings(): if (settings.DEBUG or settings.IN_TEST_SUITE): return for (key, value) in [('SECRET_KEY', 'please change this'), ('SESSION_COOKIE_SECURE', False), ('APP_PURCHASE_SECRET', 'please change this')]: if (getattr(settings, key) == value): raise ImproperlyConfigured('{0} must be changed from default'.format(key)) for key in ('CSP_SCRIPT_SRC',): for url in getattr(settings, key): if (url.startswith('http://') and (url != settings.SITE_URL)): raise ImproperlyConfigured('{0} has a http URL: {1}'.format(key, url))
[ "def", "validate_settings", "(", ")", ":", "if", "(", "settings", ".", "DEBUG", "or", "settings", ".", "IN_TEST_SUITE", ")", ":", "return", "for", "(", "key", ",", "value", ")", "in", "[", "(", "'SECRET_KEY'", ",", "'please change this'", ")", ",", "(", "'SESSION_COOKIE_SECURE'", ",", "False", ")", ",", "(", "'APP_PURCHASE_SECRET'", ",", "'please change this'", ")", "]", ":", "if", "(", "getattr", "(", "settings", ",", "key", ")", "==", "value", ")", ":", "raise", "ImproperlyConfigured", "(", "'{0} must be changed from default'", ".", "format", "(", "key", ")", ")", "for", "key", "in", "(", "'CSP_SCRIPT_SRC'", ",", ")", ":", "for", "url", "in", "getattr", "(", "settings", ",", "key", ")", ":", "if", "(", "url", ".", "startswith", "(", "'http://'", ")", "and", "(", "url", "!=", "settings", ".", "SITE_URL", ")", ")", ":", "raise", "ImproperlyConfigured", "(", "'{0} has a http URL: {1}'", ".", "format", "(", "key", ",", "url", ")", ")" ]
validate that if not in debug mode .
train
false
39,246
@removals.remove(message='keystoneclient auth plugins are deprecated. Use keystoneauth.', version='2.1.0', removal_version='3.0.0') @positional() def register_argparse_arguments(parser, argv, default=None): in_parser = argparse.ArgumentParser(add_help=False) env_plugin = os.environ.get('OS_AUTH_PLUGIN', default) for p in (in_parser, parser): p.add_argument('--os-auth-plugin', metavar='<name>', default=env_plugin, help='The auth plugin to load') (options, _args) = in_parser.parse_known_args(argv) if (not options.os_auth_plugin): return None if isinstance(options.os_auth_plugin, type): msg = 'Default Authentication options' plugin = options.os_auth_plugin else: msg = ('Options specific to the %s plugin.' % options.os_auth_plugin) plugin = base.get_plugin_class(options.os_auth_plugin) group = parser.add_argument_group('Authentication Options', msg) plugin.register_argparse_arguments(group) return plugin
[ "@", "removals", ".", "remove", "(", "message", "=", "'keystoneclient auth plugins are deprecated. Use keystoneauth.'", ",", "version", "=", "'2.1.0'", ",", "removal_version", "=", "'3.0.0'", ")", "@", "positional", "(", ")", "def", "register_argparse_arguments", "(", "parser", ",", "argv", ",", "default", "=", "None", ")", ":", "in_parser", "=", "argparse", ".", "ArgumentParser", "(", "add_help", "=", "False", ")", "env_plugin", "=", "os", ".", "environ", ".", "get", "(", "'OS_AUTH_PLUGIN'", ",", "default", ")", "for", "p", "in", "(", "in_parser", ",", "parser", ")", ":", "p", ".", "add_argument", "(", "'--os-auth-plugin'", ",", "metavar", "=", "'<name>'", ",", "default", "=", "env_plugin", ",", "help", "=", "'The auth plugin to load'", ")", "(", "options", ",", "_args", ")", "=", "in_parser", ".", "parse_known_args", "(", "argv", ")", "if", "(", "not", "options", ".", "os_auth_plugin", ")", ":", "return", "None", "if", "isinstance", "(", "options", ".", "os_auth_plugin", ",", "type", ")", ":", "msg", "=", "'Default Authentication options'", "plugin", "=", "options", ".", "os_auth_plugin", "else", ":", "msg", "=", "(", "'Options specific to the %s plugin.'", "%", "options", ".", "os_auth_plugin", ")", "plugin", "=", "base", ".", "get_plugin_class", "(", "options", ".", "os_auth_plugin", ")", "group", "=", "parser", ".", "add_argument_group", "(", "'Authentication Options'", ",", "msg", ")", "plugin", ".", "register_argparse_arguments", "(", "group", ")", "return", "plugin" ]
register cli options needed to create a plugin .
train
false
39,247
def createProtocolMessages(data, flags=const.FLAG_PAYLOAD): messages = [] while (len(data) > const.MPU): messages.append(ProtocolMessage(data[:const.MPU], flags=flags)) data = data[const.MPU:] messages.append(ProtocolMessage(data, flags=flags)) return messages
[ "def", "createProtocolMessages", "(", "data", ",", "flags", "=", "const", ".", "FLAG_PAYLOAD", ")", ":", "messages", "=", "[", "]", "while", "(", "len", "(", "data", ")", ">", "const", ".", "MPU", ")", ":", "messages", ".", "append", "(", "ProtocolMessage", "(", "data", "[", ":", "const", ".", "MPU", "]", ",", "flags", "=", "flags", ")", ")", "data", "=", "data", "[", "const", ".", "MPU", ":", "]", "messages", ".", "append", "(", "ProtocolMessage", "(", "data", ",", "flags", "=", "flags", ")", ")", "return", "messages" ]
create protocol messages out of the given payload .
train
false
39,248
@require_POST @csrf_protect def generic_delete(request, what, obj_name=None): if (not test_user_authenticated(request)): return login(request, next=('/cobbler_web/%s/delete/%s' % (what, obj_name)), expired=True) if (obj_name is None): return error_page(request, ('You must specify a %s to delete' % what)) if (not remote.has_item(what, obj_name)): return error_page(request, ('Unknown %s specified' % what)) elif (not remote.check_access_no_fail(request.session['token'], ('remove_%s' % what), obj_name)): return error_page(request, ('You do not have permission to delete this %s' % what)) else: recursive = simplejson.loads(request.POST.get('recursive', 'false')) try: remote.xapi_object_edit(what, obj_name, 'remove', {'name': obj_name, 'recursive': recursive}, request.session['token']) except Exception as e: return error_page(request, str(e)) return HttpResponseRedirect(('/cobbler_web/%s/list' % what))
[ "@", "require_POST", "@", "csrf_protect", "def", "generic_delete", "(", "request", ",", "what", ",", "obj_name", "=", "None", ")", ":", "if", "(", "not", "test_user_authenticated", "(", "request", ")", ")", ":", "return", "login", "(", "request", ",", "next", "=", "(", "'/cobbler_web/%s/delete/%s'", "%", "(", "what", ",", "obj_name", ")", ")", ",", "expired", "=", "True", ")", "if", "(", "obj_name", "is", "None", ")", ":", "return", "error_page", "(", "request", ",", "(", "'You must specify a %s to delete'", "%", "what", ")", ")", "if", "(", "not", "remote", ".", "has_item", "(", "what", ",", "obj_name", ")", ")", ":", "return", "error_page", "(", "request", ",", "(", "'Unknown %s specified'", "%", "what", ")", ")", "elif", "(", "not", "remote", ".", "check_access_no_fail", "(", "request", ".", "session", "[", "'token'", "]", ",", "(", "'remove_%s'", "%", "what", ")", ",", "obj_name", ")", ")", ":", "return", "error_page", "(", "request", ",", "(", "'You do not have permission to delete this %s'", "%", "what", ")", ")", "else", ":", "recursive", "=", "simplejson", ".", "loads", "(", "request", ".", "POST", ".", "get", "(", "'recursive'", ",", "'false'", ")", ")", "try", ":", "remote", ".", "xapi_object_edit", "(", "what", ",", "obj_name", ",", "'remove'", ",", "{", "'name'", ":", "obj_name", ",", "'recursive'", ":", "recursive", "}", ",", "request", ".", "session", "[", "'token'", "]", ")", "except", "Exception", "as", "e", ":", "return", "error_page", "(", "request", ",", "str", "(", "e", ")", ")", "return", "HttpResponseRedirect", "(", "(", "'/cobbler_web/%s/list'", "%", "what", ")", ")" ]
deletes an object .
train
false
39,252
def bake_lazy_loaders(): BakedLazyLoader._strategy_keys[:] = [] properties.RelationshipProperty.strategy_for(lazy='select')(BakedLazyLoader) properties.RelationshipProperty.strategy_for(lazy=True)(BakedLazyLoader) properties.RelationshipProperty.strategy_for(lazy='baked_select')(BakedLazyLoader) strategies.LazyLoader._strategy_keys[:] = BakedLazyLoader._strategy_keys[:]
[ "def", "bake_lazy_loaders", "(", ")", ":", "BakedLazyLoader", ".", "_strategy_keys", "[", ":", "]", "=", "[", "]", "properties", ".", "RelationshipProperty", ".", "strategy_for", "(", "lazy", "=", "'select'", ")", "(", "BakedLazyLoader", ")", "properties", ".", "RelationshipProperty", ".", "strategy_for", "(", "lazy", "=", "True", ")", "(", "BakedLazyLoader", ")", "properties", ".", "RelationshipProperty", ".", "strategy_for", "(", "lazy", "=", "'baked_select'", ")", "(", "BakedLazyLoader", ")", "strategies", ".", "LazyLoader", ".", "_strategy_keys", "[", ":", "]", "=", "BakedLazyLoader", ".", "_strategy_keys", "[", ":", "]" ]
enable the use of baked queries for all lazyloaders systemwide .
train
false
39,253
def EMA(ds, count, timeperiod=(- (2 ** 31))): return call_talib_with_ds(ds, count, talib.EMA, timeperiod)
[ "def", "EMA", "(", "ds", ",", "count", ",", "timeperiod", "=", "(", "-", "(", "2", "**", "31", ")", ")", ")", ":", "return", "call_talib_with_ds", "(", "ds", ",", "count", ",", "talib", ".", "EMA", ",", "timeperiod", ")" ]
exponential moving average .
train
false
39,254
def parse_fuzzy(source, ch): saved_pos = source.pos if (ch != '{'): return None constraints = {} try: parse_fuzzy_item(source, constraints) while source.match(','): parse_fuzzy_item(source, constraints) except ParseError: source.pos = saved_pos return None if (not source.match('}')): raise error('expected }', source.string, source.pos) return constraints
[ "def", "parse_fuzzy", "(", "source", ",", "ch", ")", ":", "saved_pos", "=", "source", ".", "pos", "if", "(", "ch", "!=", "'{'", ")", ":", "return", "None", "constraints", "=", "{", "}", "try", ":", "parse_fuzzy_item", "(", "source", ",", "constraints", ")", "while", "source", ".", "match", "(", "','", ")", ":", "parse_fuzzy_item", "(", "source", ",", "constraints", ")", "except", "ParseError", ":", "source", ".", "pos", "=", "saved_pos", "return", "None", "if", "(", "not", "source", ".", "match", "(", "'}'", ")", ")", ":", "raise", "error", "(", "'expected }'", ",", "source", ".", "string", ",", "source", ".", "pos", ")", "return", "constraints" ]
parses a fuzzy setting .
train
false
39,256
def user_full_name(strategy, details, user=None, **kwargs): if user: full_name = details.get(u'fullname', u'').strip() if ((not full_name) and ((u'first_name' in details) or (u'last_name' in details))): first_name = details.get(u'first_name', u'') last_name = details.get(u'last_name', u'') if (first_name and (first_name not in last_name)): full_name = u'{0} {1}'.format(first_name, last_name) elif first_name: full_name = first_name else: full_name = last_name full_name = full_name.strip() if (len(full_name) > 30): full_name = full_name[:30] if (full_name and (full_name != user.first_name)): user.first_name = full_name strategy.storage.user.changed(user)
[ "def", "user_full_name", "(", "strategy", ",", "details", ",", "user", "=", "None", ",", "**", "kwargs", ")", ":", "if", "user", ":", "full_name", "=", "details", ".", "get", "(", "u'fullname'", ",", "u''", ")", ".", "strip", "(", ")", "if", "(", "(", "not", "full_name", ")", "and", "(", "(", "u'first_name'", "in", "details", ")", "or", "(", "u'last_name'", "in", "details", ")", ")", ")", ":", "first_name", "=", "details", ".", "get", "(", "u'first_name'", ",", "u''", ")", "last_name", "=", "details", ".", "get", "(", "u'last_name'", ",", "u''", ")", "if", "(", "first_name", "and", "(", "first_name", "not", "in", "last_name", ")", ")", ":", "full_name", "=", "u'{0} {1}'", ".", "format", "(", "first_name", ",", "last_name", ")", "elif", "first_name", ":", "full_name", "=", "first_name", "else", ":", "full_name", "=", "last_name", "full_name", "=", "full_name", ".", "strip", "(", ")", "if", "(", "len", "(", "full_name", ")", ">", "30", ")", ":", "full_name", "=", "full_name", "[", ":", "30", "]", "if", "(", "full_name", "and", "(", "full_name", "!=", "user", ".", "first_name", ")", ")", ":", "user", ".", "first_name", "=", "full_name", "strategy", ".", "storage", ".", "user", ".", "changed", "(", "user", ")" ]
update user full name using data from provider .
train
false
39,257
def register_schema(path, schema): schema_paths[path] = schema
[ "def", "register_schema", "(", "path", ",", "schema", ")", ":", "schema_paths", "[", "path", "]", "=", "schema" ]
register schema to be available at path for $refs .
train
false
39,258
def serialize_json(obj, pretty=False, indent=None, **kwargs): for name in ['allow_nan', 'separators', 'sort_keys']: if (name in kwargs): raise ValueError(('The value of %r is computed internally, overriding is not permissable.' % name)) pretty = settings.pretty(pretty) if pretty: separators = (',', ': ') else: separators = (',', ':') if (pretty and (indent is None)): indent = 2 return json.dumps(obj, cls=BokehJSONEncoder, allow_nan=False, indent=indent, separators=separators, sort_keys=True, **kwargs)
[ "def", "serialize_json", "(", "obj", ",", "pretty", "=", "False", ",", "indent", "=", "None", ",", "**", "kwargs", ")", ":", "for", "name", "in", "[", "'allow_nan'", ",", "'separators'", ",", "'sort_keys'", "]", ":", "if", "(", "name", "in", "kwargs", ")", ":", "raise", "ValueError", "(", "(", "'The value of %r is computed internally, overriding is not permissable.'", "%", "name", ")", ")", "pretty", "=", "settings", ".", "pretty", "(", "pretty", ")", "if", "pretty", ":", "separators", "=", "(", "','", ",", "': '", ")", "else", ":", "separators", "=", "(", "','", ",", "':'", ")", "if", "(", "pretty", "and", "(", "indent", "is", "None", ")", ")", ":", "indent", "=", "2", "return", "json", ".", "dumps", "(", "obj", ",", "cls", "=", "BokehJSONEncoder", ",", "allow_nan", "=", "False", ",", "indent", "=", "indent", ",", "separators", "=", "separators", ",", "sort_keys", "=", "True", ",", "**", "kwargs", ")" ]
return a serialized json representation of objects .
train
true
39,259
def logFactorial(x): return lgamma((x + 1.0))
[ "def", "logFactorial", "(", "x", ")", ":", "return", "lgamma", "(", "(", "x", "+", "1.0", ")", ")" ]
approximation to the log of the factorial function .
train
false
39,261
def _shell_command(command): repo_dir = os.path.dirname(os.path.abspath(__file__)) command_subprocess = subprocess.Popen(command, stdout=subprocess.PIPE, stderr=subprocess.PIPE, cwd=repo_dir, universal_newlines=True) return command_subprocess.communicate()[0]
[ "def", "_shell_command", "(", "command", ")", ":", "repo_dir", "=", "os", ".", "path", ".", "dirname", "(", "os", ".", "path", ".", "abspath", "(", "__file__", ")", ")", "command_subprocess", "=", "subprocess", ".", "Popen", "(", "command", ",", "stdout", "=", "subprocess", ".", "PIPE", ",", "stderr", "=", "subprocess", ".", "PIPE", ",", "cwd", "=", "repo_dir", ",", "universal_newlines", "=", "True", ")", "return", "command_subprocess", ".", "communicate", "(", ")", "[", "0", "]" ]
return the first result of a shell command .
train
false
39,262
@ensure_csrf_cookie @cache_control(no_cache=True, no_store=True, must_revalidate=True) @require_POST @login_required def set_course_mode_price(request, course_id): try: course_price = int(request.POST['course_price']) except ValueError: return JsonResponse({'message': _('Please Enter the numeric value for the course price')}, status=400) currency = request.POST['currency'] course_key = SlashSeparatedCourseKey.from_deprecated_string(course_id) course_honor_mode = CourseMode.objects.filter(mode_slug='honor', course_id=course_key) if (not course_honor_mode): return JsonResponse({'message': _('CourseMode with the mode slug({mode_slug}) DoesNotExist').format(mode_slug='honor')}, status=400) CourseModesArchive.objects.create(course_id=course_id, mode_slug='honor', mode_display_name='Honor Code Certificate', min_price=course_honor_mode[0].min_price, currency=course_honor_mode[0].currency, expiration_datetime=datetime.datetime.now(pytz.utc), expiration_date=datetime.date.today()) course_honor_mode.update(min_price=course_price, currency=currency) return JsonResponse({'message': _('CourseMode price updated successfully')})
[ "@", "ensure_csrf_cookie", "@", "cache_control", "(", "no_cache", "=", "True", ",", "no_store", "=", "True", ",", "must_revalidate", "=", "True", ")", "@", "require_POST", "@", "login_required", "def", "set_course_mode_price", "(", "request", ",", "course_id", ")", ":", "try", ":", "course_price", "=", "int", "(", "request", ".", "POST", "[", "'course_price'", "]", ")", "except", "ValueError", ":", "return", "JsonResponse", "(", "{", "'message'", ":", "_", "(", "'Please Enter the numeric value for the course price'", ")", "}", ",", "status", "=", "400", ")", "currency", "=", "request", ".", "POST", "[", "'currency'", "]", "course_key", "=", "SlashSeparatedCourseKey", ".", "from_deprecated_string", "(", "course_id", ")", "course_honor_mode", "=", "CourseMode", ".", "objects", ".", "filter", "(", "mode_slug", "=", "'honor'", ",", "course_id", "=", "course_key", ")", "if", "(", "not", "course_honor_mode", ")", ":", "return", "JsonResponse", "(", "{", "'message'", ":", "_", "(", "'CourseMode with the mode slug({mode_slug}) DoesNotExist'", ")", ".", "format", "(", "mode_slug", "=", "'honor'", ")", "}", ",", "status", "=", "400", ")", "CourseModesArchive", ".", "objects", ".", "create", "(", "course_id", "=", "course_id", ",", "mode_slug", "=", "'honor'", ",", "mode_display_name", "=", "'Honor Code Certificate'", ",", "min_price", "=", "course_honor_mode", "[", "0", "]", ".", "min_price", ",", "currency", "=", "course_honor_mode", "[", "0", "]", ".", "currency", ",", "expiration_datetime", "=", "datetime", ".", "datetime", ".", "now", "(", "pytz", ".", "utc", ")", ",", "expiration_date", "=", "datetime", ".", "date", ".", "today", "(", ")", ")", "course_honor_mode", ".", "update", "(", "min_price", "=", "course_price", ",", "currency", "=", "currency", ")", "return", "JsonResponse", "(", "{", "'message'", ":", "_", "(", "'CourseMode price updated successfully'", ")", "}", ")" ]
set the new course price and add new entry in the coursemodesarchive table .
train
false
39,263
def backupConfigZip(fileList, archive, arcname=None): try: with zipfile.ZipFile(archive, u'w', zipfile.ZIP_DEFLATED, allowZip64=True) as z: for f in list(set(fileList)): z.write(f, os.path.relpath(f, arcname)) return True except Exception as e: sickrage.srCore.srLogger.error(u'Zip creation error: {} '.format(e.message)) return False
[ "def", "backupConfigZip", "(", "fileList", ",", "archive", ",", "arcname", "=", "None", ")", ":", "try", ":", "with", "zipfile", ".", "ZipFile", "(", "archive", ",", "u'w'", ",", "zipfile", ".", "ZIP_DEFLATED", ",", "allowZip64", "=", "True", ")", "as", "z", ":", "for", "f", "in", "list", "(", "set", "(", "fileList", ")", ")", ":", "z", ".", "write", "(", "f", ",", "os", ".", "path", ".", "relpath", "(", "f", ",", "arcname", ")", ")", "return", "True", "except", "Exception", "as", "e", ":", "sickrage", ".", "srCore", ".", "srLogger", ".", "error", "(", "u'Zip creation error: {} '", ".", "format", "(", "e", ".", "message", ")", ")", "return", "False" ]
store the config file as a zip .
train
false
39,264
def _get_course_block_counts(auth_token, block_url): headers = {'Authorization': 'Bearer {}'.format(auth_token)} response = requests.get(block_url, headers=headers) if (response.status_code != 200): print 'url {} returned status code {}'.format(block_url, response.status_code) return {} response_json = response.json() if ((BLOCK_ROOT_KEY in response_json) and (BLOCKS_KEY in response_json)): root_val = response_json[BLOCK_ROOT_KEY] counts = response_json[BLOCKS_KEY][root_val][BLOCK_COUNTS_KEY] return counts return {}
[ "def", "_get_course_block_counts", "(", "auth_token", ",", "block_url", ")", ":", "headers", "=", "{", "'Authorization'", ":", "'Bearer {}'", ".", "format", "(", "auth_token", ")", "}", "response", "=", "requests", ".", "get", "(", "block_url", ",", "headers", "=", "headers", ")", "if", "(", "response", ".", "status_code", "!=", "200", ")", ":", "print", "'url {} returned status code {}'", ".", "format", "(", "block_url", ",", "response", ".", "status_code", ")", "return", "{", "}", "response_json", "=", "response", ".", "json", "(", ")", "if", "(", "(", "BLOCK_ROOT_KEY", "in", "response_json", ")", "and", "(", "BLOCKS_KEY", "in", "response_json", ")", ")", ":", "root_val", "=", "response_json", "[", "BLOCK_ROOT_KEY", "]", "counts", "=", "response_json", "[", "BLOCKS_KEY", "]", "[", "root_val", "]", "[", "BLOCK_COUNTS_KEY", "]", "return", "counts", "return", "{", "}" ]
get the block counts for a given block_url arguments: auth_token : the authentication token to access the api block_url : the respective url for a courses xblock data returns: dict: a dictionary containing the block counts .
train
false
39,267
def Overlay(child, parent): for arg in (child, parent): if (not isinstance(arg, collections.Mapping)): raise DefinitionError(('Trying to merge badly defined hints. Child: %s, Parent: %s' % (type(child), type(parent)))) for attr in ['fix', 'format', 'problem', 'summary']: if (not child.get(attr)): child[attr] = parent.get(attr, '').strip() return child
[ "def", "Overlay", "(", "child", ",", "parent", ")", ":", "for", "arg", "in", "(", "child", ",", "parent", ")", ":", "if", "(", "not", "isinstance", "(", "arg", ",", "collections", ".", "Mapping", ")", ")", ":", "raise", "DefinitionError", "(", "(", "'Trying to merge badly defined hints. Child: %s, Parent: %s'", "%", "(", "type", "(", "child", ")", ",", "type", "(", "parent", ")", ")", ")", ")", "for", "attr", "in", "[", "'fix'", ",", "'format'", ",", "'problem'", ",", "'summary'", "]", ":", "if", "(", "not", "child", ".", "get", "(", "attr", ")", ")", ":", "child", "[", "attr", "]", "=", "parent", ".", "get", "(", "attr", ",", "''", ")", ".", "strip", "(", ")", "return", "child" ]
adds hint attributes to a child hint if they are not defined .
train
true
39,268
def make_archive(base_name, format, root_dir=None, base_dir=None, verbose=0, dry_run=0, owner=None, group=None, logger=None): save_cwd = os.getcwd() if (root_dir is not None): if (logger is not None): logger.debug("changing into '%s'", root_dir) base_name = os.path.abspath(base_name) if (not dry_run): os.chdir(root_dir) if (base_dir is None): base_dir = os.curdir kwargs = {'dry_run': dry_run, 'logger': logger} try: format_info = _ARCHIVE_FORMATS[format] except KeyError: raise ValueError(("unknown archive format '%s'" % format)) func = format_info[0] for (arg, val) in format_info[1]: kwargs[arg] = val if (format != 'zip'): kwargs['owner'] = owner kwargs['group'] = group try: filename = func(base_name, base_dir, **kwargs) finally: if (root_dir is not None): if (logger is not None): logger.debug("changing back to '%s'", save_cwd) os.chdir(save_cwd) return filename
[ "def", "make_archive", "(", "base_name", ",", "format", ",", "root_dir", "=", "None", ",", "base_dir", "=", "None", ",", "verbose", "=", "0", ",", "dry_run", "=", "0", ",", "owner", "=", "None", ",", "group", "=", "None", ",", "logger", "=", "None", ")", ":", "save_cwd", "=", "os", ".", "getcwd", "(", ")", "if", "(", "root_dir", "is", "not", "None", ")", ":", "if", "(", "logger", "is", "not", "None", ")", ":", "logger", ".", "debug", "(", "\"changing into '%s'\"", ",", "root_dir", ")", "base_name", "=", "os", ".", "path", ".", "abspath", "(", "base_name", ")", "if", "(", "not", "dry_run", ")", ":", "os", ".", "chdir", "(", "root_dir", ")", "if", "(", "base_dir", "is", "None", ")", ":", "base_dir", "=", "os", ".", "curdir", "kwargs", "=", "{", "'dry_run'", ":", "dry_run", ",", "'logger'", ":", "logger", "}", "try", ":", "format_info", "=", "_ARCHIVE_FORMATS", "[", "format", "]", "except", "KeyError", ":", "raise", "ValueError", "(", "(", "\"unknown archive format '%s'\"", "%", "format", ")", ")", "func", "=", "format_info", "[", "0", "]", "for", "(", "arg", ",", "val", ")", "in", "format_info", "[", "1", "]", ":", "kwargs", "[", "arg", "]", "=", "val", "if", "(", "format", "!=", "'zip'", ")", ":", "kwargs", "[", "'owner'", "]", "=", "owner", "kwargs", "[", "'group'", "]", "=", "group", "try", ":", "filename", "=", "func", "(", "base_name", ",", "base_dir", ",", "**", "kwargs", ")", "finally", ":", "if", "(", "root_dir", "is", "not", "None", ")", ":", "if", "(", "logger", "is", "not", "None", ")", ":", "logger", ".", "debug", "(", "\"changing back to '%s'\"", ",", "save_cwd", ")", "os", ".", "chdir", "(", "save_cwd", ")", "return", "filename" ]
create an archive file .
train
true
39,269
def writeXMLElement(fileNames, target, xmlElement): object = target.object if (object == None): print 'Warning, writeTarget in write could not get object for:' print xmlElement return fileNameRoot = evaluate.getEvaluatedStringDefault('', 'name', target) fileNameRoot = evaluate.getEvaluatedStringDefault(fileNameRoot, 'id', target) fileNameRoot = evaluate.getEvaluatedStringDefault(fileNameRoot, 'file', xmlElement) fileNameRoot += evaluate.getEvaluatedStringDefault('', 'suffix', xmlElement) extension = evaluate.getEvaluatedStringDefault(object.getFabricationExtension(), 'extension', xmlElement) fileName = ('%s.%s' % (fileNameRoot, extension)) suffixIndex = 1 while (fileName in fileNames): fileName = ('%s_%s.%s' % (fileNameRoot, suffixIndex, extension)) suffixIndex += 1 fileNames.append(fileName) folderName = evaluate.getEvaluatedStringDefault('', 'folder', xmlElement) absoluteFolderDirectory = os.path.join(os.path.dirname(xmlElement.getRoot().parser.fileName), folderName) archive.makeDirectory(absoluteFolderDirectory) archive.writeFileText(os.path.join(absoluteFolderDirectory, fileName), object.getFabricationText())
[ "def", "writeXMLElement", "(", "fileNames", ",", "target", ",", "xmlElement", ")", ":", "object", "=", "target", ".", "object", "if", "(", "object", "==", "None", ")", ":", "print", "'Warning, writeTarget in write could not get object for:'", "print", "xmlElement", "return", "fileNameRoot", "=", "evaluate", ".", "getEvaluatedStringDefault", "(", "''", ",", "'name'", ",", "target", ")", "fileNameRoot", "=", "evaluate", ".", "getEvaluatedStringDefault", "(", "fileNameRoot", ",", "'id'", ",", "target", ")", "fileNameRoot", "=", "evaluate", ".", "getEvaluatedStringDefault", "(", "fileNameRoot", ",", "'file'", ",", "xmlElement", ")", "fileNameRoot", "+=", "evaluate", ".", "getEvaluatedStringDefault", "(", "''", ",", "'suffix'", ",", "xmlElement", ")", "extension", "=", "evaluate", ".", "getEvaluatedStringDefault", "(", "object", ".", "getFabricationExtension", "(", ")", ",", "'extension'", ",", "xmlElement", ")", "fileName", "=", "(", "'%s.%s'", "%", "(", "fileNameRoot", ",", "extension", ")", ")", "suffixIndex", "=", "1", "while", "(", "fileName", "in", "fileNames", ")", ":", "fileName", "=", "(", "'%s_%s.%s'", "%", "(", "fileNameRoot", ",", "suffixIndex", ",", "extension", ")", ")", "suffixIndex", "+=", "1", "fileNames", ".", "append", "(", "fileName", ")", "folderName", "=", "evaluate", ".", "getEvaluatedStringDefault", "(", "''", ",", "'folder'", ",", "xmlElement", ")", "absoluteFolderDirectory", "=", "os", ".", "path", ".", "join", "(", "os", ".", "path", ".", "dirname", "(", "xmlElement", ".", "getRoot", "(", ")", ".", "parser", ".", "fileName", ")", ",", "folderName", ")", "archive", ".", "makeDirectory", "(", "absoluteFolderDirectory", ")", "archive", ".", "writeFileText", "(", "os", ".", "path", ".", "join", "(", "absoluteFolderDirectory", ",", "fileName", ")", ",", "object", ".", "getFabricationText", "(", ")", ")" ]
write target .
train
false
39,272
def iterdecode(value): if (not value): return tuple() result = [] accumulator = u'' escaped = False for c in value: if (not escaped): if (c == CHAR_ESCAPE): escaped = True continue elif (c == CHAR_SEPARATOR): result.append(accumulator) accumulator = u'' continue else: escaped = False accumulator += c result.append(accumulator) return tuple(result)
[ "def", "iterdecode", "(", "value", ")", ":", "if", "(", "not", "value", ")", ":", "return", "tuple", "(", ")", "result", "=", "[", "]", "accumulator", "=", "u''", "escaped", "=", "False", "for", "c", "in", "value", ":", "if", "(", "not", "escaped", ")", ":", "if", "(", "c", "==", "CHAR_ESCAPE", ")", ":", "escaped", "=", "True", "continue", "elif", "(", "c", "==", "CHAR_SEPARATOR", ")", ":", "result", ".", "append", "(", "accumulator", ")", "accumulator", "=", "u''", "continue", "else", ":", "escaped", "=", "False", "accumulator", "+=", "c", "result", ".", "append", "(", "accumulator", ")", "return", "tuple", "(", "result", ")" ]
decode enumerable from string presentation as a tuple .
train
false
39,273
def normalize_indices(context, builder, index_types, indices): if len(indices): (index_types, indices) = zip(*[normalize_index(context, builder, idxty, idx) for (idxty, idx) in zip(index_types, indices)]) return (index_types, indices)
[ "def", "normalize_indices", "(", "context", ",", "builder", ",", "index_types", ",", "indices", ")", ":", "if", "len", "(", "indices", ")", ":", "(", "index_types", ",", "indices", ")", "=", "zip", "(", "*", "[", "normalize_index", "(", "context", ",", "builder", ",", "idxty", ",", "idx", ")", "for", "(", "idxty", ",", "idx", ")", "in", "zip", "(", "index_types", ",", "indices", ")", "]", ")", "return", "(", "index_types", ",", "indices", ")" ]
same as normalize_index() .
train
false
39,274
@verbose def compute_proj_epochs(epochs, n_grad=2, n_mag=2, n_eeg=2, n_jobs=1, desc_prefix=None, verbose=None): data = _compute_cov_epochs(epochs, n_jobs) event_id = epochs.event_id if ((event_id is None) or (len(list(event_id.keys())) == 0)): event_id = '0' elif (len(event_id.keys()) == 1): event_id = str(list(event_id.values())[0]) else: event_id = 'Multiple-events' if (desc_prefix is None): desc_prefix = ('%s-%-.3f-%-.3f' % (event_id, epochs.tmin, epochs.tmax)) return _compute_proj(data, epochs.info, n_grad, n_mag, n_eeg, desc_prefix)
[ "@", "verbose", "def", "compute_proj_epochs", "(", "epochs", ",", "n_grad", "=", "2", ",", "n_mag", "=", "2", ",", "n_eeg", "=", "2", ",", "n_jobs", "=", "1", ",", "desc_prefix", "=", "None", ",", "verbose", "=", "None", ")", ":", "data", "=", "_compute_cov_epochs", "(", "epochs", ",", "n_jobs", ")", "event_id", "=", "epochs", ".", "event_id", "if", "(", "(", "event_id", "is", "None", ")", "or", "(", "len", "(", "list", "(", "event_id", ".", "keys", "(", ")", ")", ")", "==", "0", ")", ")", ":", "event_id", "=", "'0'", "elif", "(", "len", "(", "event_id", ".", "keys", "(", ")", ")", "==", "1", ")", ":", "event_id", "=", "str", "(", "list", "(", "event_id", ".", "values", "(", ")", ")", "[", "0", "]", ")", "else", ":", "event_id", "=", "'Multiple-events'", "if", "(", "desc_prefix", "is", "None", ")", ":", "desc_prefix", "=", "(", "'%s-%-.3f-%-.3f'", "%", "(", "event_id", ",", "epochs", ".", "tmin", ",", "epochs", ".", "tmax", ")", ")", "return", "_compute_proj", "(", "data", ",", "epochs", ".", "info", ",", "n_grad", ",", "n_mag", ",", "n_eeg", ",", "desc_prefix", ")" ]
compute ssp vectors on epochs .
train
false
39,276
def revoke_grant(key_id, grant_id, region=None, key=None, keyid=None, profile=None): conn = _get_conn(region=region, key=key, keyid=keyid, profile=profile) if key_id.startswith('alias/'): key_id = _get_key_id(key_id) r = {} try: conn.revoke_grant(key_id, grant_id) r['result'] = True except boto.exception.BotoServerError as e: r['result'] = False r['error'] = __utils__['boto.get_error'](e) return r
[ "def", "revoke_grant", "(", "key_id", ",", "grant_id", ",", "region", "=", "None", ",", "key", "=", "None", ",", "keyid", "=", "None", ",", "profile", "=", "None", ")", ":", "conn", "=", "_get_conn", "(", "region", "=", "region", ",", "key", "=", "key", ",", "keyid", "=", "keyid", ",", "profile", "=", "profile", ")", "if", "key_id", ".", "startswith", "(", "'alias/'", ")", ":", "key_id", "=", "_get_key_id", "(", "key_id", ")", "r", "=", "{", "}", "try", ":", "conn", ".", "revoke_grant", "(", "key_id", ",", "grant_id", ")", "r", "[", "'result'", "]", "=", "True", "except", "boto", ".", "exception", ".", "BotoServerError", "as", "e", ":", "r", "[", "'result'", "]", "=", "False", "r", "[", "'error'", "]", "=", "__utils__", "[", "'boto.get_error'", "]", "(", "e", ")", "return", "r" ]
revoke a grant from a key .
train
true
39,277
def addElementToListDictionary(element, key, listDictionary): if (key in listDictionary): listDictionary[key].append(element) else: listDictionary[key] = [element]
[ "def", "addElementToListDictionary", "(", "element", ",", "key", ",", "listDictionary", ")", ":", "if", "(", "key", "in", "listDictionary", ")", ":", "listDictionary", "[", "key", "]", ".", "append", "(", "element", ")", "else", ":", "listDictionary", "[", "key", "]", "=", "[", "element", "]" ]
add an element to the list table .
train
false
39,278
def _bit_length(num): return len(bin(num).lstrip(u'-0b'))
[ "def", "_bit_length", "(", "num", ")", ":", "return", "len", "(", "bin", "(", "num", ")", ".", "lstrip", "(", "u'-0b'", ")", ")" ]
return number of bits needed to encode given number .
train
false
39,279
def dorogovtsev_goltsev_mendes_graph(n, create_using=None): if (create_using is not None): if create_using.is_directed(): raise nx.NetworkXError('Directed Graph not supported') if create_using.is_multigraph(): raise nx.NetworkXError('Multigraph not supported') G = empty_graph(0, create_using) G.name = 'Dorogovtsev-Goltsev-Mendes Graph' G.add_edge(0, 1) if (n == 0): return G new_node = 2 for i in range(1, (n + 1)): last_generation_edges = list(G.edges()) number_of_edges_in_last_generation = len(last_generation_edges) for j in range(0, number_of_edges_in_last_generation): G.add_edge(new_node, last_generation_edges[j][0]) G.add_edge(new_node, last_generation_edges[j][1]) new_node += 1 return G
[ "def", "dorogovtsev_goltsev_mendes_graph", "(", "n", ",", "create_using", "=", "None", ")", ":", "if", "(", "create_using", "is", "not", "None", ")", ":", "if", "create_using", ".", "is_directed", "(", ")", ":", "raise", "nx", ".", "NetworkXError", "(", "'Directed Graph not supported'", ")", "if", "create_using", ".", "is_multigraph", "(", ")", ":", "raise", "nx", ".", "NetworkXError", "(", "'Multigraph not supported'", ")", "G", "=", "empty_graph", "(", "0", ",", "create_using", ")", "G", ".", "name", "=", "'Dorogovtsev-Goltsev-Mendes Graph'", "G", ".", "add_edge", "(", "0", ",", "1", ")", "if", "(", "n", "==", "0", ")", ":", "return", "G", "new_node", "=", "2", "for", "i", "in", "range", "(", "1", ",", "(", "n", "+", "1", ")", ")", ":", "last_generation_edges", "=", "list", "(", "G", ".", "edges", "(", ")", ")", "number_of_edges_in_last_generation", "=", "len", "(", "last_generation_edges", ")", "for", "j", "in", "range", "(", "0", ",", "number_of_edges_in_last_generation", ")", ":", "G", ".", "add_edge", "(", "new_node", ",", "last_generation_edges", "[", "j", "]", "[", "0", "]", ")", "G", ".", "add_edge", "(", "new_node", ",", "last_generation_edges", "[", "j", "]", "[", "1", "]", ")", "new_node", "+=", "1", "return", "G" ]
return the hierarchically constructed dorogovtsev-goltsev-mendes graph .
train
false
39,283
def course_discovery_api_client(user): scopes = ['email', 'profile'] expires_in = settings.OAUTH_ID_TOKEN_EXPIRATION jwt = JwtBuilder(user).build_token(scopes, expires_in) return EdxRestApiClient(settings.COURSE_CATALOG_API_URL, jwt=jwt)
[ "def", "course_discovery_api_client", "(", "user", ")", ":", "scopes", "=", "[", "'email'", ",", "'profile'", "]", "expires_in", "=", "settings", ".", "OAUTH_ID_TOKEN_EXPIRATION", "jwt", "=", "JwtBuilder", "(", "user", ")", ".", "build_token", "(", "scopes", ",", "expires_in", ")", "return", "EdxRestApiClient", "(", "settings", ".", "COURSE_CATALOG_API_URL", ",", "jwt", "=", "jwt", ")" ]
returns a course discovery api client setup with authentication for the specified user .
train
false
39,284
def _get_gecos(name): try: gecos_field = pwd.getpwnam(name).pw_gecos.split(',', 3) except KeyError: raise CommandExecutionError("User '{0}' does not exist".format(name)) if (not gecos_field): return {} else: while (len(gecos_field) < 4): gecos_field.append('') return {'fullname': locales.sdecode(gecos_field[0]), 'roomnumber': locales.sdecode(gecos_field[1]), 'workphone': locales.sdecode(gecos_field[2]), 'homephone': locales.sdecode(gecos_field[3])}
[ "def", "_get_gecos", "(", "name", ")", ":", "try", ":", "gecos_field", "=", "pwd", ".", "getpwnam", "(", "name", ")", ".", "pw_gecos", ".", "split", "(", "','", ",", "3", ")", "except", "KeyError", ":", "raise", "CommandExecutionError", "(", "\"User '{0}' does not exist\"", ".", "format", "(", "name", ")", ")", "if", "(", "not", "gecos_field", ")", ":", "return", "{", "}", "else", ":", "while", "(", "len", "(", "gecos_field", ")", "<", "4", ")", ":", "gecos_field", ".", "append", "(", "''", ")", "return", "{", "'fullname'", ":", "locales", ".", "sdecode", "(", "gecos_field", "[", "0", "]", ")", ",", "'roomnumber'", ":", "locales", ".", "sdecode", "(", "gecos_field", "[", "1", "]", ")", ",", "'workphone'", ":", "locales", ".", "sdecode", "(", "gecos_field", "[", "2", "]", ")", ",", "'homephone'", ":", "locales", ".", "sdecode", "(", "gecos_field", "[", "3", "]", ")", "}" ]
retrieve gecos field info and return it in dictionary form .
train
false
39,285
def trace_value(value, msg='Value: '): print msg, value return value
[ "def", "trace_value", "(", "value", ",", "msg", "=", "'Value: '", ")", ":", "print", "msg", ",", "value", "return", "value" ]
prints value and returns value .
train
false
39,286
def check_headers_valid(headers): reason = invalid_headers_reason(headers) if (reason is not None): raise InvalidEmailError(reason)
[ "def", "check_headers_valid", "(", "headers", ")", ":", "reason", "=", "invalid_headers_reason", "(", "headers", ")", "if", "(", "reason", "is", "not", "None", ")", ":", "raise", "InvalidEmailError", "(", "reason", ")" ]
check that headers is a valid dictionary for headers .
train
false
39,288
def filter_eliot_main(args, base_path, top_level, stdin=None, stdout=None): stdin = (sys.stdin if (stdin is None) else stdin) stdout = (sys.stdout if (stdout is None) else stdout) if args: lines = chain_files(args) else: lines = stdin for line in lines: for output in process_line(line): stdout.write(output)
[ "def", "filter_eliot_main", "(", "args", ",", "base_path", ",", "top_level", ",", "stdin", "=", "None", ",", "stdout", "=", "None", ")", ":", "stdin", "=", "(", "sys", ".", "stdin", "if", "(", "stdin", "is", "None", ")", "else", "stdin", ")", "stdout", "=", "(", "sys", ".", "stdout", "if", "(", "stdout", "is", "None", ")", "else", "stdout", ")", "if", "args", ":", "lines", "=", "chain_files", "(", "args", ")", "else", ":", "lines", "=", "stdin", "for", "line", "in", "lines", ":", "for", "output", "in", "process_line", "(", "line", ")", ":", "stdout", ".", "write", "(", "output", ")" ]
filter logs .
train
false
39,290
def MakeTokenRegex(meta_left, meta_right): key = (meta_left, meta_right) if (key not in _token_re_cache): _token_re_cache[key] = re.compile((((('(' + re.escape(meta_left)) + '.+?') + re.escape(meta_right)) + ')')) return _token_re_cache[key]
[ "def", "MakeTokenRegex", "(", "meta_left", ",", "meta_right", ")", ":", "key", "=", "(", "meta_left", ",", "meta_right", ")", "if", "(", "key", "not", "in", "_token_re_cache", ")", ":", "_token_re_cache", "[", "key", "]", "=", "re", ".", "compile", "(", "(", "(", "(", "(", "'('", "+", "re", ".", "escape", "(", "meta_left", ")", ")", "+", "'.+?'", ")", "+", "re", ".", "escape", "(", "meta_right", ")", ")", "+", "')'", ")", ")", "return", "_token_re_cache", "[", "key", "]" ]
return a regular expression for tokenization .
train
false
39,291
def determine_run_start_sync_state(run_change_sysuptime, start_save_sysuptime): if (start_save_sysuptime == 0): return False elif (start_save_sysuptime >= run_change_sysuptime): return True return False
[ "def", "determine_run_start_sync_state", "(", "run_change_sysuptime", ",", "start_save_sysuptime", ")", ":", "if", "(", "start_save_sysuptime", "==", "0", ")", ":", "return", "False", "elif", "(", "start_save_sysuptime", ">=", "run_change_sysuptime", ")", ":", "return", "True", "return", "False" ]
return true if run/start are in sync return false if run/start are out of sync three cases: 1 .
train
false
39,292
def extract_url_path_and_query(full_url=None, no_query=False): if (full_url is None): full_url = request.url split = urlsplit(full_url) result = (split.path or '/') if ((not no_query) and split.query): result += ('?' + split.query) return result
[ "def", "extract_url_path_and_query", "(", "full_url", "=", "None", ",", "no_query", "=", "False", ")", ":", "if", "(", "full_url", "is", "None", ")", ":", "full_url", "=", "request", ".", "url", "split", "=", "urlsplit", "(", "full_url", ")", "result", "=", "(", "split", ".", "path", "or", "'/'", ")", "if", "(", "(", "not", "no_query", ")", "and", "split", ".", "query", ")", ":", "result", "+=", "(", "'?'", "+", "split", ".", "query", ")", "return", "result" ]
convert URL to /aaa/p .
train
false
39,293
def _setup_server(hass, config): conf = config.get(DOMAIN, {}) if ((CONF_EMBEDDED not in conf) and (CONF_BROKER in conf)): return None server = prepare_setup_platform(hass, config, DOMAIN, 'server') if (server is None): _LOGGER.error('Unable to load embedded server') return None (success, broker_config) = server.start(hass, conf.get(CONF_EMBEDDED)) return (success and broker_config)
[ "def", "_setup_server", "(", "hass", ",", "config", ")", ":", "conf", "=", "config", ".", "get", "(", "DOMAIN", ",", "{", "}", ")", "if", "(", "(", "CONF_EMBEDDED", "not", "in", "conf", ")", "and", "(", "CONF_BROKER", "in", "conf", ")", ")", ":", "return", "None", "server", "=", "prepare_setup_platform", "(", "hass", ",", "config", ",", "DOMAIN", ",", "'server'", ")", "if", "(", "server", "is", "None", ")", ":", "_LOGGER", ".", "error", "(", "'Unable to load embedded server'", ")", "return", "None", "(", "success", ",", "broker_config", ")", "=", "server", ".", "start", "(", "hass", ",", "conf", ".", "get", "(", "CONF_EMBEDDED", ")", ")", "return", "(", "success", "and", "broker_config", ")" ]
try to start embedded mqtt broker .
train
false
39,294
def testtable(): testtable = table([['A1', 'A2'], ['B1', 'B2'], ['C1', 'C2']]) assert (testtable.xpath('/ns0:tbl/ns0:tr[2]/ns0:tc[2]/ns0:p/ns0:r/ns0:t', namespaces={'ns0': 'http://schemas.openxmlformats.org/wordprocessingml/2006/main'})[0].text == 'B2')
[ "def", "testtable", "(", ")", ":", "testtable", "=", "table", "(", "[", "[", "'A1'", ",", "'A2'", "]", ",", "[", "'B1'", ",", "'B2'", "]", ",", "[", "'C1'", ",", "'C2'", "]", "]", ")", "assert", "(", "testtable", ".", "xpath", "(", "'/ns0:tbl/ns0:tr[2]/ns0:tc[2]/ns0:p/ns0:r/ns0:t'", ",", "namespaces", "=", "{", "'ns0'", ":", "'http://schemas.openxmlformats.org/wordprocessingml/2006/main'", "}", ")", "[", "0", "]", ".", "text", "==", "'B2'", ")" ]
ensure tables make sense .
train
false
39,296
def test_takes_all_arguments(): assert (not hug.introspect.takes_all_arguments(function_with_kwargs, 'argument1', 'argument2', 'argument3')) assert (not hug.introspect.takes_all_arguments(function_with_args, 'argument1', 'argument2', 'argument3')) assert (not hug.introspect.takes_all_arguments(function_with_neither, 'argument1', 'argument2', 'argument3')) assert hug.introspect.takes_all_arguments(function_with_both, 'argument1', 'argument2', 'argument3')
[ "def", "test_takes_all_arguments", "(", ")", ":", "assert", "(", "not", "hug", ".", "introspect", ".", "takes_all_arguments", "(", "function_with_kwargs", ",", "'argument1'", ",", "'argument2'", ",", "'argument3'", ")", ")", "assert", "(", "not", "hug", ".", "introspect", ".", "takes_all_arguments", "(", "function_with_args", ",", "'argument1'", ",", "'argument2'", ",", "'argument3'", ")", ")", "assert", "(", "not", "hug", ".", "introspect", ".", "takes_all_arguments", "(", "function_with_neither", ",", "'argument1'", ",", "'argument2'", ",", "'argument3'", ")", ")", "assert", "hug", ".", "introspect", ".", "takes_all_arguments", "(", "function_with_both", ",", "'argument1'", ",", "'argument2'", ",", "'argument3'", ")" ]
test to ensure hug introspection can correctly identify if a function takes all specified arguments .
train
false
39,297
def matrix_IQR(x): num_cols = x.shape[1] min_vals = zeros(num_cols) max_vals = zeros(num_cols) for i in range(x.shape[1]): col = x[:, i] (min_vals[i], max_vals[i]) = IQR(col) return (min_vals, max_vals)
[ "def", "matrix_IQR", "(", "x", ")", ":", "num_cols", "=", "x", ".", "shape", "[", "1", "]", "min_vals", "=", "zeros", "(", "num_cols", ")", "max_vals", "=", "zeros", "(", "num_cols", ")", "for", "i", "in", "range", "(", "x", ".", "shape", "[", "1", "]", ")", ":", "col", "=", "x", "[", ":", ",", "i", "]", "(", "min_vals", "[", "i", "]", ",", "max_vals", "[", "i", "]", ")", "=", "IQR", "(", "col", ")", "return", "(", "min_vals", ",", "max_vals", ")" ]
calculates the iqr for each column in an array .
train
false
39,298
def get_course_creator_status(user): user = CourseCreator.objects.filter(user=user) if (user.count() == 0): return None else: return user[0].state
[ "def", "get_course_creator_status", "(", "user", ")", ":", "user", "=", "CourseCreator", ".", "objects", ".", "filter", "(", "user", "=", "user", ")", "if", "(", "user", ".", "count", "(", ")", "==", "0", ")", ":", "return", "None", "else", ":", "return", "user", "[", "0", "]", ".", "state" ]
returns the status for a particular user .
train
false
39,299
@pytest.mark.django_db def test_user_has_manager_permissions(no_perms_user, administrate, tp0): language0 = tp0.language project0 = tp0.project assert (not no_perms_user.has_manager_permissions()) criteria = {'user': no_perms_user, 'directory': tp0.directory} ps = PermissionSet.objects.get_or_create(**criteria)[0] ps.positive_permissions.set([administrate]) ps.save() assert no_perms_user.has_manager_permissions() ps.positive_permissions.clear() assert (not no_perms_user.has_manager_permissions()) criteria['directory'] = language0.directory ps = PermissionSet.objects.get_or_create(**criteria)[0] ps.positive_permissions.set([administrate]) ps.save() assert no_perms_user.has_manager_permissions() ps.positive_permissions.clear() assert (not no_perms_user.has_manager_permissions()) criteria['directory'] = project0.directory ps = PermissionSet.objects.get_or_create(**criteria)[0] ps.positive_permissions.set([administrate]) ps.save() assert no_perms_user.has_manager_permissions() ps.positive_permissions.clear() assert (not no_perms_user.has_manager_permissions())
[ "@", "pytest", ".", "mark", ".", "django_db", "def", "test_user_has_manager_permissions", "(", "no_perms_user", ",", "administrate", ",", "tp0", ")", ":", "language0", "=", "tp0", ".", "language", "project0", "=", "tp0", ".", "project", "assert", "(", "not", "no_perms_user", ".", "has_manager_permissions", "(", ")", ")", "criteria", "=", "{", "'user'", ":", "no_perms_user", ",", "'directory'", ":", "tp0", ".", "directory", "}", "ps", "=", "PermissionSet", ".", "objects", ".", "get_or_create", "(", "**", "criteria", ")", "[", "0", "]", "ps", ".", "positive_permissions", ".", "set", "(", "[", "administrate", "]", ")", "ps", ".", "save", "(", ")", "assert", "no_perms_user", ".", "has_manager_permissions", "(", ")", "ps", ".", "positive_permissions", ".", "clear", "(", ")", "assert", "(", "not", "no_perms_user", ".", "has_manager_permissions", "(", ")", ")", "criteria", "[", "'directory'", "]", "=", "language0", ".", "directory", "ps", "=", "PermissionSet", ".", "objects", ".", "get_or_create", "(", "**", "criteria", ")", "[", "0", "]", "ps", ".", "positive_permissions", ".", "set", "(", "[", "administrate", "]", ")", "ps", ".", "save", "(", ")", "assert", "no_perms_user", ".", "has_manager_permissions", "(", ")", "ps", ".", "positive_permissions", ".", "clear", "(", ")", "assert", "(", "not", "no_perms_user", ".", "has_manager_permissions", "(", ")", ")", "criteria", "[", "'directory'", "]", "=", "project0", ".", "directory", "ps", "=", "PermissionSet", ".", "objects", ".", "get_or_create", "(", "**", "criteria", ")", "[", "0", "]", "ps", ".", "positive_permissions", ".", "set", "(", "[", "administrate", "]", ")", "ps", ".", "save", "(", ")", "assert", "no_perms_user", ".", "has_manager_permissions", "(", ")", "ps", ".", "positive_permissions", ".", "clear", "(", ")", "assert", "(", "not", "no_perms_user", ".", "has_manager_permissions", "(", ")", ")" ]
test user has_manager_permissions method .
train
false
39,300
def unicode_list(_list): return [unicode_obj(x) for x in _list]
[ "def", "unicode_list", "(", "_list", ")", ":", "return", "[", "unicode_obj", "(", "x", ")", "for", "x", "in", "_list", "]" ]
make sure every element in list is unicode .
train
false
39,301
def func2(a, b): return (a, b)
[ "def", "func2", "(", "a", ",", "b", ")", ":", "return", "(", "a", ",", "b", ")" ]
test function for issue 9936 .
train
false
39,303
def div_by_zero(x): return (x / 0)
[ "def", "div_by_zero", "(", "x", ")", ":", "return", "(", "x", "/", "0", ")" ]
prepare for a division-by-zero exception .
train
false
39,307
def test_temperature_unit(): schema = vol.Schema(cv.temperature_unit) with pytest.raises(vol.MultipleInvalid): schema('K') schema('C') schema('F')
[ "def", "test_temperature_unit", "(", ")", ":", "schema", "=", "vol", ".", "Schema", "(", "cv", ".", "temperature_unit", ")", "with", "pytest", ".", "raises", "(", "vol", ".", "MultipleInvalid", ")", ":", "schema", "(", "'K'", ")", "schema", "(", "'C'", ")", "schema", "(", "'F'", ")" ]
test temperature unit validation .
train
false
39,308
def write_blob_diff(f, old_file, new_file): (old_path, old_mode, old_blob) = old_file (new_path, new_mode, new_blob) = new_file old_path = patch_filename(old_path, 'a') new_path = patch_filename(new_path, 'b') def lines(blob): if (blob is not None): return blob.splitlines() else: return [] f.writelines(gen_diff_header((old_path, new_path), (old_mode, new_mode), (getattr(old_blob, 'id', None), getattr(new_blob, 'id', None)))) old_contents = lines(old_blob) new_contents = lines(new_blob) f.writelines(unified_diff(old_contents, new_contents, old_path, new_path))
[ "def", "write_blob_diff", "(", "f", ",", "old_file", ",", "new_file", ")", ":", "(", "old_path", ",", "old_mode", ",", "old_blob", ")", "=", "old_file", "(", "new_path", ",", "new_mode", ",", "new_blob", ")", "=", "new_file", "old_path", "=", "patch_filename", "(", "old_path", ",", "'a'", ")", "new_path", "=", "patch_filename", "(", "new_path", ",", "'b'", ")", "def", "lines", "(", "blob", ")", ":", "if", "(", "blob", "is", "not", "None", ")", ":", "return", "blob", ".", "splitlines", "(", ")", "else", ":", "return", "[", "]", "f", ".", "writelines", "(", "gen_diff_header", "(", "(", "old_path", ",", "new_path", ")", ",", "(", "old_mode", ",", "new_mode", ")", ",", "(", "getattr", "(", "old_blob", ",", "'id'", ",", "None", ")", ",", "getattr", "(", "new_blob", ",", "'id'", ",", "None", ")", ")", ")", ")", "old_contents", "=", "lines", "(", "old_blob", ")", "new_contents", "=", "lines", "(", "new_blob", ")", "f", ".", "writelines", "(", "unified_diff", "(", "old_contents", ",", "new_contents", ",", "old_path", ",", "new_path", ")", ")" ]
write blob diff .
train
false
39,309
def executable(path): return os.access(path, os.X_OK)
[ "def", "executable", "(", "path", ")", ":", "return", "os", ".", "access", "(", "path", ",", "os", ".", "X_OK", ")" ]
test for executable file .
train
false