id_within_dataset
int64
1
55.5k
snippet
stringlengths
19
14.2k
tokens
listlengths
6
1.63k
nl
stringlengths
6
352
split_within_dataset
stringclasses
1 value
is_duplicated
bool
2 classes
18,347
def fork_waitfor_timed(tmp, pid, timeout): timer_expired = True poll_time = 2 time_passed = 0 while (time_passed < timeout): time.sleep(poll_time) (child_pid, status) = os.waitpid(pid, os.WNOHANG) if ((child_pid, status) == (0, 0)): time_passed = (time_passed + poll_time) else: timer_expired = False break if timer_expired: logging.info('Timer expired (%d sec.), nuking pid %d', timeout, pid) utils.nuke_pid(pid) (child_pid, status) = os.waitpid(pid, 0) raise error.TestError(('Test timeout expired, rc=%d' % status)) else: _check_for_subprocess_exception(tmp, pid) if status: raise error.TestError(('Test subprocess failed rc=%d' % status))
[ "def", "fork_waitfor_timed", "(", "tmp", ",", "pid", ",", "timeout", ")", ":", "timer_expired", "=", "True", "poll_time", "=", "2", "time_passed", "=", "0", "while", "(", "time_passed", "<", "timeout", ")", ":", "time", ".", "sleep", "(", "poll_time", ")", "(", "child_pid", ",", "status", ")", "=", "os", ".", "waitpid", "(", "pid", ",", "os", ".", "WNOHANG", ")", "if", "(", "(", "child_pid", ",", "status", ")", "==", "(", "0", ",", "0", ")", ")", ":", "time_passed", "=", "(", "time_passed", "+", "poll_time", ")", "else", ":", "timer_expired", "=", "False", "break", "if", "timer_expired", ":", "logging", ".", "info", "(", "'Timer expired (%d sec.), nuking pid %d'", ",", "timeout", ",", "pid", ")", "utils", ".", "nuke_pid", "(", "pid", ")", "(", "child_pid", ",", "status", ")", "=", "os", ".", "waitpid", "(", "pid", ",", "0", ")", "raise", "error", ".", "TestError", "(", "(", "'Test timeout expired, rc=%d'", "%", "status", ")", ")", "else", ":", "_check_for_subprocess_exception", "(", "tmp", ",", "pid", ")", "if", "status", ":", "raise", "error", ".", "TestError", "(", "(", "'Test subprocess failed rc=%d'", "%", "status", ")", ")" ]
waits for pid until it terminates or timeout expires .
train
false
18,348
def findLinksRel(link_attrs_list, target_rel): matchesTarget = (lambda attrs: linkHasRel(attrs, target_rel)) return filter(matchesTarget, link_attrs_list)
[ "def", "findLinksRel", "(", "link_attrs_list", ",", "target_rel", ")", ":", "matchesTarget", "=", "(", "lambda", "attrs", ":", "linkHasRel", "(", "attrs", ",", "target_rel", ")", ")", "return", "filter", "(", "matchesTarget", ",", "link_attrs_list", ")" ]
filter the list of link attributes on whether it has target_rel as a relationship .
train
true
18,350
def magic_memit(self, line=''): (opts, stmt) = self.parse_options(line, 'r:t', posix=False, strict=False) repeat = int(getattr(opts, 'r', 1)) if (repeat < 1): repeat = 1 timeout = int(getattr(opts, 't', 0)) if (timeout <= 0): timeout = None mem_usage = [] for _ in range(repeat): tmp = memory_usage((_func_exec, (stmt, self.shell.user_ns)), timeout=timeout) mem_usage.extend(tmp) if mem_usage: print ('maximum of %d: %f MB per loop' % (repeat, max(mem_usage))) else: print 'ERROR: could not read memory usage, try with a lower interval or more iterations'
[ "def", "magic_memit", "(", "self", ",", "line", "=", "''", ")", ":", "(", "opts", ",", "stmt", ")", "=", "self", ".", "parse_options", "(", "line", ",", "'r:t'", ",", "posix", "=", "False", ",", "strict", "=", "False", ")", "repeat", "=", "int", "(", "getattr", "(", "opts", ",", "'r'", ",", "1", ")", ")", "if", "(", "repeat", "<", "1", ")", ":", "repeat", "=", "1", "timeout", "=", "int", "(", "getattr", "(", "opts", ",", "'t'", ",", "0", ")", ")", "if", "(", "timeout", "<=", "0", ")", ":", "timeout", "=", "None", "mem_usage", "=", "[", "]", "for", "_", "in", "range", "(", "repeat", ")", ":", "tmp", "=", "memory_usage", "(", "(", "_func_exec", ",", "(", "stmt", ",", "self", ".", "shell", ".", "user_ns", ")", ")", ",", "timeout", "=", "timeout", ")", "mem_usage", ".", "extend", "(", "tmp", ")", "if", "mem_usage", ":", "print", "(", "'maximum of %d: %f MB per loop'", "%", "(", "repeat", ",", "max", "(", "mem_usage", ")", ")", ")", "else", ":", "print", "'ERROR: could not read memory usage, try with a lower interval or more iterations'" ]
measure memory usage of a python statement usage .
train
false
18,351
def _api_change_opts(name, output, kwargs): value = kwargs.get('value') value2 = kwargs.get('value2') if (value and value2 and value2.isdigit()): result = NzbQueue.do.change_opts(value, int(value2)) return report(output, keyword='status', data=bool((result > 0)))
[ "def", "_api_change_opts", "(", "name", ",", "output", ",", "kwargs", ")", ":", "value", "=", "kwargs", ".", "get", "(", "'value'", ")", "value2", "=", "kwargs", ".", "get", "(", "'value2'", ")", "if", "(", "value", "and", "value2", "and", "value2", ".", "isdigit", "(", ")", ")", ":", "result", "=", "NzbQueue", ".", "do", ".", "change_opts", "(", "value", ",", "int", "(", "value2", ")", ")", "return", "report", "(", "output", ",", "keyword", "=", "'status'", ",", "data", "=", "bool", "(", "(", "result", ">", "0", ")", ")", ")" ]
api: accepts output .
train
false
18,352
def sequences_to_file(results, outfile_name): f = open(outfile_name, 'w+') for (label, seq) in results: output_lines = [] output_lines.append(('>%s\n' % label)) output_lines.append(('%s\n' % seq)) f.writelines(output_lines) f.close()
[ "def", "sequences_to_file", "(", "results", ",", "outfile_name", ")", ":", "f", "=", "open", "(", "outfile_name", ",", "'w+'", ")", "for", "(", "label", ",", "seq", ")", "in", "results", ":", "output_lines", "=", "[", "]", "output_lines", ".", "append", "(", "(", "'>%s\\n'", "%", "label", ")", ")", "output_lines", ".", "append", "(", "(", "'%s\\n'", "%", "seq", ")", ")", "f", ".", "writelines", "(", "output_lines", ")", "f", ".", "close", "(", ")" ]
translate a generator of label .
train
false
18,354
def create_oebbook(log, path_or_stream, opts, reader=None, encoding='utf-8', populate=True, for_regex_wizard=False, specialize=None): from calibre.ebooks.oeb.base import OEBBook html_preprocessor = HTMLPreProcessor(log, opts, regex_wizard_callback=regex_wizard_callback) if (not encoding): encoding = None oeb = OEBBook(log, html_preprocessor, pretty_print=opts.pretty_print, input_encoding=encoding) if (not populate): return oeb if (specialize is not None): oeb = (specialize(oeb) or oeb) log('Parsing all content...') if (reader is None): from calibre.ebooks.oeb.reader import OEBReader reader = OEBReader reader()(oeb, path_or_stream) return oeb
[ "def", "create_oebbook", "(", "log", ",", "path_or_stream", ",", "opts", ",", "reader", "=", "None", ",", "encoding", "=", "'utf-8'", ",", "populate", "=", "True", ",", "for_regex_wizard", "=", "False", ",", "specialize", "=", "None", ")", ":", "from", "calibre", ".", "ebooks", ".", "oeb", ".", "base", "import", "OEBBook", "html_preprocessor", "=", "HTMLPreProcessor", "(", "log", ",", "opts", ",", "regex_wizard_callback", "=", "regex_wizard_callback", ")", "if", "(", "not", "encoding", ")", ":", "encoding", "=", "None", "oeb", "=", "OEBBook", "(", "log", ",", "html_preprocessor", ",", "pretty_print", "=", "opts", ".", "pretty_print", ",", "input_encoding", "=", "encoding", ")", "if", "(", "not", "populate", ")", ":", "return", "oeb", "if", "(", "specialize", "is", "not", "None", ")", ":", "oeb", "=", "(", "specialize", "(", "oeb", ")", "or", "oeb", ")", "log", "(", "'Parsing all content...'", ")", "if", "(", "reader", "is", "None", ")", ":", "from", "calibre", ".", "ebooks", ".", "oeb", ".", "reader", "import", "OEBReader", "reader", "=", "OEBReader", "reader", "(", ")", "(", "oeb", ",", "path_or_stream", ")", "return", "oeb" ]
create an oebbook .
train
false
18,356
def patched(f): def wrapped(*args, **kwargs): kwargs['return_response'] = False kwargs['prefetch'] = True config = kwargs.get('config', {}) config.update(safe_mode=True) kwargs['config'] = config return f(*args, **kwargs) return wrapped
[ "def", "patched", "(", "f", ")", ":", "def", "wrapped", "(", "*", "args", ",", "**", "kwargs", ")", ":", "kwargs", "[", "'return_response'", "]", "=", "False", "kwargs", "[", "'prefetch'", "]", "=", "True", "config", "=", "kwargs", ".", "get", "(", "'config'", ",", "{", "}", ")", "config", ".", "update", "(", "safe_mode", "=", "True", ")", "kwargs", "[", "'config'", "]", "=", "config", "return", "f", "(", "*", "args", ",", "**", "kwargs", ")", "return", "wrapped" ]
patches calls to run_on_app_servers on instance methods .
train
false
18,357
def format_quote(tweet): screen_name = tweet['user']['screen_name'] text = tweet['text'] tid = str(tweet['id']) if ('#owner' not in c['QUOTE_FORMAT']): printNicely(light_magenta('Quote should contains #owner')) return False if ('#comment' not in c['QUOTE_FORMAT']): printNicely(light_magenta('Quote format should have #comment')) return False formater = '' try: formater = c['QUOTE_FORMAT'] formater = formater.replace('#owner', screen_name) formater = formater.replace('#tweet', text) formater = formater.replace('#tid', tid) formater = emojize(formater) except: pass notice = formater.split() notice = lmap((lambda x: (light_green(x) if (x == '#comment') else x)), notice) notice = lmap((lambda x: (color_func(c['TWEET']['rt'])(x) if (x == 'RT') else x)), notice) notice = lmap((lambda x: (cycle_color(x) if (x[0] == '@') else x)), notice) notice = lmap((lambda x: (color_func(c['TWEET']['link'])(x) if (x[0:4] == 'http') else x)), notice) notice = lmap((lambda x: (color_func(c['TWEET']['hashtag'])(x) if x.startswith('#') else x)), notice) notice = ' '.join(notice) notice = ((light_magenta('Quoting: "') + notice) + light_magenta('"')) printNicely(notice) return formater
[ "def", "format_quote", "(", "tweet", ")", ":", "screen_name", "=", "tweet", "[", "'user'", "]", "[", "'screen_name'", "]", "text", "=", "tweet", "[", "'text'", "]", "tid", "=", "str", "(", "tweet", "[", "'id'", "]", ")", "if", "(", "'#owner'", "not", "in", "c", "[", "'QUOTE_FORMAT'", "]", ")", ":", "printNicely", "(", "light_magenta", "(", "'Quote should contains #owner'", ")", ")", "return", "False", "if", "(", "'#comment'", "not", "in", "c", "[", "'QUOTE_FORMAT'", "]", ")", ":", "printNicely", "(", "light_magenta", "(", "'Quote format should have #comment'", ")", ")", "return", "False", "formater", "=", "''", "try", ":", "formater", "=", "c", "[", "'QUOTE_FORMAT'", "]", "formater", "=", "formater", ".", "replace", "(", "'#owner'", ",", "screen_name", ")", "formater", "=", "formater", ".", "replace", "(", "'#tweet'", ",", "text", ")", "formater", "=", "formater", ".", "replace", "(", "'#tid'", ",", "tid", ")", "formater", "=", "emojize", "(", "formater", ")", "except", ":", "pass", "notice", "=", "formater", ".", "split", "(", ")", "notice", "=", "lmap", "(", "(", "lambda", "x", ":", "(", "light_green", "(", "x", ")", "if", "(", "x", "==", "'#comment'", ")", "else", "x", ")", ")", ",", "notice", ")", "notice", "=", "lmap", "(", "(", "lambda", "x", ":", "(", "color_func", "(", "c", "[", "'TWEET'", "]", "[", "'rt'", "]", ")", "(", "x", ")", "if", "(", "x", "==", "'RT'", ")", "else", "x", ")", ")", ",", "notice", ")", "notice", "=", "lmap", "(", "(", "lambda", "x", ":", "(", "cycle_color", "(", "x", ")", "if", "(", "x", "[", "0", "]", "==", "'@'", ")", "else", "x", ")", ")", ",", "notice", ")", "notice", "=", "lmap", "(", "(", "lambda", "x", ":", "(", "color_func", "(", "c", "[", "'TWEET'", "]", "[", "'link'", "]", ")", "(", "x", ")", "if", "(", "x", "[", "0", ":", "4", "]", "==", "'http'", ")", "else", "x", ")", ")", ",", "notice", ")", "notice", "=", "lmap", "(", "(", "lambda", "x", ":", "(", "color_func", "(", "c", "[", "'TWEET'", "]", "[", "'hashtag'", "]", ")", "(", "x", ")", "if", "x", ".", "startswith", "(", "'#'", ")", "else", "x", ")", ")", ",", "notice", ")", "notice", "=", "' '", ".", "join", "(", "notice", ")", "notice", "=", "(", "(", "light_magenta", "(", "'Quoting: \"'", ")", "+", "notice", ")", "+", "light_magenta", "(", "'\"'", ")", ")", "printNicely", "(", "notice", ")", "return", "formater" ]
returns a formatted string of a quote .
train
false
18,358
def sort_ids(ids, mapping): def _lookup(id): try: return len(mapping[id]) except KeyError: return 0 deco = [(_lookup(id), id) for id in ids] deco.sort(reverse=True) return [id for (_, id) in deco]
[ "def", "sort_ids", "(", "ids", ",", "mapping", ")", ":", "def", "_lookup", "(", "id", ")", ":", "try", ":", "return", "len", "(", "mapping", "[", "id", "]", ")", "except", "KeyError", ":", "return", "0", "deco", "=", "[", "(", "_lookup", "(", "id", ")", ",", "id", ")", "for", "id", "in", "ids", "]", "deco", ".", "sort", "(", "reverse", "=", "True", ")", "return", "[", "id", "for", "(", "_", ",", "id", ")", "in", "deco", "]" ]
sorts ids based on their cluster_size .
train
false
18,360
def test_resource_url(): data = jinja.render('test2.html') print data url = QUrl(data) assert url.isValid() assert (url.scheme() == 'file') path = url.path() if (os.name == 'nt'): path = path.lstrip('/') path = path.replace('/', os.sep) with open(path, 'r', encoding='utf-8') as f: assert (f.read().splitlines()[0] == 'Hello World!')
[ "def", "test_resource_url", "(", ")", ":", "data", "=", "jinja", ".", "render", "(", "'test2.html'", ")", "print", "data", "url", "=", "QUrl", "(", "data", ")", "assert", "url", ".", "isValid", "(", ")", "assert", "(", "url", ".", "scheme", "(", ")", "==", "'file'", ")", "path", "=", "url", ".", "path", "(", ")", "if", "(", "os", ".", "name", "==", "'nt'", ")", ":", "path", "=", "path", ".", "lstrip", "(", "'/'", ")", "path", "=", "path", ".", "replace", "(", "'/'", ",", "os", ".", "sep", ")", "with", "open", "(", "path", ",", "'r'", ",", "encoding", "=", "'utf-8'", ")", "as", "f", ":", "assert", "(", "f", ".", "read", "(", ")", ".", "splitlines", "(", ")", "[", "0", "]", "==", "'Hello World!'", ")" ]
test resource_url() which can be used from templates .
train
false
18,362
def directed_connected_components(digr): if (not digr.DIRECTED): raise Exception(('%s is not a directed graph' % digr)) finishing_times = DFS_loop(digr.get_transpose()) (nodes_explored, connected_components) = ([], []) for node in finishing_times[::(-1)]: component = [] outer_dfs(digr, node, nodes_explored, component) if component: nodes_explored += component connected_components.append(component) return connected_components
[ "def", "directed_connected_components", "(", "digr", ")", ":", "if", "(", "not", "digr", ".", "DIRECTED", ")", ":", "raise", "Exception", "(", "(", "'%s is not a directed graph'", "%", "digr", ")", ")", "finishing_times", "=", "DFS_loop", "(", "digr", ".", "get_transpose", "(", ")", ")", "(", "nodes_explored", ",", "connected_components", ")", "=", "(", "[", "]", ",", "[", "]", ")", "for", "node", "in", "finishing_times", "[", ":", ":", "(", "-", "1", ")", "]", ":", "component", "=", "[", "]", "outer_dfs", "(", "digr", ",", "node", ",", "nodes_explored", ",", "component", ")", "if", "component", ":", "nodes_explored", "+=", "component", "connected_components", ".", "append", "(", "component", ")", "return", "connected_components" ]
returns a list of strongly connected components in a directed graph using kosarajus two pass algorithm .
train
false
18,363
def read_partition_file(filename): for (record_index, record) in enumerate(screed.open(filename)): (_, partition_id) = record.name.rsplit(' DCTB ', 1) (yield (record_index, record, int(partition_id)))
[ "def", "read_partition_file", "(", "filename", ")", ":", "for", "(", "record_index", ",", "record", ")", "in", "enumerate", "(", "screed", ".", "open", "(", "filename", ")", ")", ":", "(", "_", ",", "partition_id", ")", "=", "record", ".", "name", ".", "rsplit", "(", "' DCTB '", ",", "1", ")", "(", "yield", "(", "record_index", ",", "record", ",", "int", "(", "partition_id", ")", ")", ")" ]
utility function to get partitioned reads from file .
train
false
18,366
def setup_ecobee(hass, network, config): if (network.pin is not None): request_configuration(network, hass, config) return if ('ecobee' in _CONFIGURING): configurator = get_component('configurator') configurator.request_done(_CONFIGURING.pop('ecobee')) hold_temp = config[DOMAIN].get(CONF_HOLD_TEMP) discovery.load_platform(hass, 'climate', DOMAIN, {'hold_temp': hold_temp}, config) discovery.load_platform(hass, 'sensor', DOMAIN, {}, config) discovery.load_platform(hass, 'binary_sensor', DOMAIN, {}, config)
[ "def", "setup_ecobee", "(", "hass", ",", "network", ",", "config", ")", ":", "if", "(", "network", ".", "pin", "is", "not", "None", ")", ":", "request_configuration", "(", "network", ",", "hass", ",", "config", ")", "return", "if", "(", "'ecobee'", "in", "_CONFIGURING", ")", ":", "configurator", "=", "get_component", "(", "'configurator'", ")", "configurator", ".", "request_done", "(", "_CONFIGURING", ".", "pop", "(", "'ecobee'", ")", ")", "hold_temp", "=", "config", "[", "DOMAIN", "]", ".", "get", "(", "CONF_HOLD_TEMP", ")", "discovery", ".", "load_platform", "(", "hass", ",", "'climate'", ",", "DOMAIN", ",", "{", "'hold_temp'", ":", "hold_temp", "}", ",", "config", ")", "discovery", ".", "load_platform", "(", "hass", ",", "'sensor'", ",", "DOMAIN", ",", "{", "}", ",", "config", ")", "discovery", ".", "load_platform", "(", "hass", ",", "'binary_sensor'", ",", "DOMAIN", ",", "{", "}", ",", "config", ")" ]
setup ecobee thermostat .
train
false
18,368
def set_default(ruby, runas=None): return _rvm(['alias', 'create', 'default', ruby], runas=runas)
[ "def", "set_default", "(", "ruby", ",", "runas", "=", "None", ")", ":", "return", "_rvm", "(", "[", "'alias'", ",", "'create'", ",", "'default'", ",", "ruby", "]", ",", "runas", "=", "runas", ")" ]
set / add a default value to defaults .
train
false
18,370
@LocalContext def run_shellcode(bytes, **kw): return ELF.from_bytes(bytes, **kw).process()
[ "@", "LocalContext", "def", "run_shellcode", "(", "bytes", ",", "**", "kw", ")", ":", "return", "ELF", ".", "from_bytes", "(", "bytes", ",", "**", "kw", ")", ".", "process", "(", ")" ]
given assembled machine code bytes .
train
false
18,371
@contextmanager def assertNoFDsLeaked(test_case): gc.collect() def process_fds(): path = FilePath('/dev/fd') if (not path.exists()): raise SkipTest('/dev/fd is not available.') return set([child.basename() for child in path.children()]) fds = process_fds() try: (yield) finally: test_case.assertEqual(process_fds(), fds)
[ "@", "contextmanager", "def", "assertNoFDsLeaked", "(", "test_case", ")", ":", "gc", ".", "collect", "(", ")", "def", "process_fds", "(", ")", ":", "path", "=", "FilePath", "(", "'/dev/fd'", ")", "if", "(", "not", "path", ".", "exists", "(", ")", ")", ":", "raise", "SkipTest", "(", "'/dev/fd is not available.'", ")", "return", "set", "(", "[", "child", ".", "basename", "(", ")", "for", "child", "in", "path", ".", "children", "(", ")", "]", ")", "fds", "=", "process_fds", "(", ")", "try", ":", "(", "yield", ")", "finally", ":", "test_case", ".", "assertEqual", "(", "process_fds", "(", ")", ",", "fds", ")" ]
context manager that asserts no file descriptors are leaked .
train
false
18,372
def decode_long(data): nbytes = len(data) if (nbytes == 0): return 0L ashex = _binascii.hexlify(data[::(-1)]) n = long(ashex, 16) if (data[(-1)] >= '\x80'): n -= (1L << (nbytes * 8)) return n
[ "def", "decode_long", "(", "data", ")", ":", "nbytes", "=", "len", "(", "data", ")", "if", "(", "nbytes", "==", "0", ")", ":", "return", "0", "L", "ashex", "=", "_binascii", ".", "hexlify", "(", "data", "[", ":", ":", "(", "-", "1", ")", "]", ")", "n", "=", "long", "(", "ashex", ",", "16", ")", "if", "(", "data", "[", "(", "-", "1", ")", "]", ">=", "'\\x80'", ")", ":", "n", "-=", "(", "1", "L", "<<", "(", "nbytes", "*", "8", ")", ")", "return", "n" ]
decode a long from a twos complement little-endian binary string .
train
false
18,373
@mock_ec2 def test_modify_attribute_blockDeviceMapping(): conn = boto.ec2.connect_to_region(u'us-east-1') reservation = conn.run_instances(u'ami-1234abcd') instance = reservation.instances[0] with assert_raises(JSONResponseError) as ex: instance.modify_attribute(u'blockDeviceMapping', {u'/dev/sda1': True}, dry_run=True) ex.exception.reason.should.equal(u'DryRunOperation') ex.exception.status.should.equal(400) ex.exception.message.should.equal(u'An error occurred (DryRunOperation) when calling the ModifyInstanceAttribute operation: Request would have succeeded, but DryRun flag is set') instance.modify_attribute(u'blockDeviceMapping', {u'/dev/sda1': True}) instance = ec2_backends[conn.region.name].get_instance(instance.id) instance.block_device_mapping.should.have.key(u'/dev/sda1') instance.block_device_mapping[u'/dev/sda1'].delete_on_termination.should.be(True)
[ "@", "mock_ec2", "def", "test_modify_attribute_blockDeviceMapping", "(", ")", ":", "conn", "=", "boto", ".", "ec2", ".", "connect_to_region", "(", "u'us-east-1'", ")", "reservation", "=", "conn", ".", "run_instances", "(", "u'ami-1234abcd'", ")", "instance", "=", "reservation", ".", "instances", "[", "0", "]", "with", "assert_raises", "(", "JSONResponseError", ")", "as", "ex", ":", "instance", ".", "modify_attribute", "(", "u'blockDeviceMapping'", ",", "{", "u'/dev/sda1'", ":", "True", "}", ",", "dry_run", "=", "True", ")", "ex", ".", "exception", ".", "reason", ".", "should", ".", "equal", "(", "u'DryRunOperation'", ")", "ex", ".", "exception", ".", "status", ".", "should", ".", "equal", "(", "400", ")", "ex", ".", "exception", ".", "message", ".", "should", ".", "equal", "(", "u'An error occurred (DryRunOperation) when calling the ModifyInstanceAttribute operation: Request would have succeeded, but DryRun flag is set'", ")", "instance", ".", "modify_attribute", "(", "u'blockDeviceMapping'", ",", "{", "u'/dev/sda1'", ":", "True", "}", ")", "instance", "=", "ec2_backends", "[", "conn", ".", "region", ".", "name", "]", ".", "get_instance", "(", "instance", ".", "id", ")", "instance", ".", "block_device_mapping", ".", "should", ".", "have", ".", "key", "(", "u'/dev/sda1'", ")", "instance", ".", "block_device_mapping", "[", "u'/dev/sda1'", "]", ".", "delete_on_termination", ".", "should", ".", "be", "(", "True", ")" ]
reproduces the missing feature explained at [0] .
train
false
18,374
def _repack_pkgs(pkgs, normalize=True): if (normalize and ('pkg.normalize_name' in __salt__)): _normalize_name = __salt__['pkg.normalize_name'] else: _normalize_name = (lambda pkgname: pkgname) return dict([(_normalize_name(str(x)), (str(y) if (y is not None) else y)) for (x, y) in six.iteritems(salt.utils.repack_dictlist(pkgs))])
[ "def", "_repack_pkgs", "(", "pkgs", ",", "normalize", "=", "True", ")", ":", "if", "(", "normalize", "and", "(", "'pkg.normalize_name'", "in", "__salt__", ")", ")", ":", "_normalize_name", "=", "__salt__", "[", "'pkg.normalize_name'", "]", "else", ":", "_normalize_name", "=", "(", "lambda", "pkgname", ":", "pkgname", ")", "return", "dict", "(", "[", "(", "_normalize_name", "(", "str", "(", "x", ")", ")", ",", "(", "str", "(", "y", ")", "if", "(", "y", "is", "not", "None", ")", "else", "y", ")", ")", "for", "(", "x", ",", "y", ")", "in", "six", ".", "iteritems", "(", "salt", ".", "utils", ".", "repack_dictlist", "(", "pkgs", ")", ")", "]", ")" ]
repack packages specified using "pkgs" argument to pkg states into a single dictionary .
train
true
18,375
def get_pkg_names(args): print('\n'.join((pkg_name for (pkg_name, pkg_tests) in get_affected_packages(args))))
[ "def", "get_pkg_names", "(", "args", ")", ":", "print", "(", "'\\n'", ".", "join", "(", "(", "pkg_name", "for", "(", "pkg_name", ",", "pkg_tests", ")", "in", "get_affected_packages", "(", "args", ")", ")", ")", ")" ]
print package names that would be affected .
train
false
18,376
@fixture def db(): global _db if (_db is None): _db = orm.new_session_factory('sqlite:///:memory:', echo=True)() user = orm.User(name=getuser()) user.servers.append(orm.Server()) hub = orm.Hub(server=orm.Server()) _db.add(user) _db.add(hub) _db.commit() return _db
[ "@", "fixture", "def", "db", "(", ")", ":", "global", "_db", "if", "(", "_db", "is", "None", ")", ":", "_db", "=", "orm", ".", "new_session_factory", "(", "'sqlite:///:memory:'", ",", "echo", "=", "True", ")", "(", ")", "user", "=", "orm", ".", "User", "(", "name", "=", "getuser", "(", ")", ")", "user", ".", "servers", ".", "append", "(", "orm", ".", "Server", "(", ")", ")", "hub", "=", "orm", ".", "Hub", "(", "server", "=", "orm", ".", "Server", "(", ")", ")", "_db", ".", "add", "(", "user", ")", "_db", ".", "add", "(", "hub", ")", "_db", ".", "commit", "(", ")", "return", "_db" ]
get a db session .
train
false
18,377
@register.function @jinja2.contextfunction def mkt_breadcrumbs(context, product=None, items=None, crumb_size=40, add_default=True, cls=None): if add_default: crumbs = [(reverse('home'), _('Home'))] else: crumbs = [] if product: if items: url_ = product.get_detail_url() else: url_ = None crumbs += [(None, _('Apps')), (url_, product.name)] if items: crumbs.extend(items) if (len(crumbs) == 1): crumbs = [] crumbs = [(u, truncate(label, crumb_size)) for (u, label) in crumbs] t = env.get_template('site/helpers/breadcrumbs.html').render({'breadcrumbs': crumbs, 'cls': cls}) return jinja2.Markup(t)
[ "@", "register", ".", "function", "@", "jinja2", ".", "contextfunction", "def", "mkt_breadcrumbs", "(", "context", ",", "product", "=", "None", ",", "items", "=", "None", ",", "crumb_size", "=", "40", ",", "add_default", "=", "True", ",", "cls", "=", "None", ")", ":", "if", "add_default", ":", "crumbs", "=", "[", "(", "reverse", "(", "'home'", ")", ",", "_", "(", "'Home'", ")", ")", "]", "else", ":", "crumbs", "=", "[", "]", "if", "product", ":", "if", "items", ":", "url_", "=", "product", ".", "get_detail_url", "(", ")", "else", ":", "url_", "=", "None", "crumbs", "+=", "[", "(", "None", ",", "_", "(", "'Apps'", ")", ")", ",", "(", "url_", ",", "product", ".", "name", ")", "]", "if", "items", ":", "crumbs", ".", "extend", "(", "items", ")", "if", "(", "len", "(", "crumbs", ")", "==", "1", ")", ":", "crumbs", "=", "[", "]", "crumbs", "=", "[", "(", "u", ",", "truncate", "(", "label", ",", "crumb_size", ")", ")", "for", "(", "u", ",", "label", ")", "in", "crumbs", "]", "t", "=", "env", ".", "get_template", "(", "'site/helpers/breadcrumbs.html'", ")", ".", "render", "(", "{", "'breadcrumbs'", ":", "crumbs", ",", "'cls'", ":", "cls", "}", ")", "return", "jinja2", ".", "Markup", "(", "t", ")" ]
wrapper function for breadcrumbs .
train
false
18,379
def _to_report(name): if (name in _REPORTS): return name raise UsageError('Unrecognized report {}. {}'.format(name, _to_report.coerceDoc))
[ "def", "_to_report", "(", "name", ")", ":", "if", "(", "name", "in", "_REPORTS", ")", ":", "return", "name", "raise", "UsageError", "(", "'Unrecognized report {}. {}'", ".", "format", "(", "name", ",", "_to_report", ".", "coerceDoc", ")", ")" ]
ensure that name is a valid report .
train
false
18,380
def model_from_yaml(yaml_string, custom_objects=None): import yaml from keras.utils.layer_utils import layer_from_config config = yaml.load(yaml_string) return layer_from_config(config, custom_objects=custom_objects)
[ "def", "model_from_yaml", "(", "yaml_string", ",", "custom_objects", "=", "None", ")", ":", "import", "yaml", "from", "keras", ".", "utils", ".", "layer_utils", "import", "layer_from_config", "config", "=", "yaml", ".", "load", "(", "yaml_string", ")", "return", "layer_from_config", "(", "config", ",", "custom_objects", "=", "custom_objects", ")" ]
parses a yaml model configuration file and returns a model instance .
train
false
18,381
@login_required def project_version_delete_html(request, project_slug, version_slug): project = get_object_or_404(Project.objects.for_admin_user(request.user), slug=project_slug) version = get_object_or_404(Version.objects.public(user=request.user, project=project, only_active=False), slug=version_slug) if (not version.active): version.built = False version.save() broadcast(type='app', task=tasks.clear_artifacts, args=[version.pk]) else: return HttpResponseBadRequest("Can't delete HTML for an active version.") return HttpResponseRedirect(reverse('project_version_list', kwargs={'project_slug': project_slug}))
[ "@", "login_required", "def", "project_version_delete_html", "(", "request", ",", "project_slug", ",", "version_slug", ")", ":", "project", "=", "get_object_or_404", "(", "Project", ".", "objects", ".", "for_admin_user", "(", "request", ".", "user", ")", ",", "slug", "=", "project_slug", ")", "version", "=", "get_object_or_404", "(", "Version", ".", "objects", ".", "public", "(", "user", "=", "request", ".", "user", ",", "project", "=", "project", ",", "only_active", "=", "False", ")", ",", "slug", "=", "version_slug", ")", "if", "(", "not", "version", ".", "active", ")", ":", "version", ".", "built", "=", "False", "version", ".", "save", "(", ")", "broadcast", "(", "type", "=", "'app'", ",", "task", "=", "tasks", ".", "clear_artifacts", ",", "args", "=", "[", "version", ".", "pk", "]", ")", "else", ":", "return", "HttpResponseBadRequest", "(", "\"Can't delete HTML for an active version.\"", ")", "return", "HttpResponseRedirect", "(", "reverse", "(", "'project_version_list'", ",", "kwargs", "=", "{", "'project_slug'", ":", "project_slug", "}", ")", ")" ]
project version delete html this marks a version as not built .
train
false
18,382
def bump_cache_for_item(item): cache.bump_version(_get_namespace_for_item(item))
[ "def", "bump_cache_for_item", "(", "item", ")", ":", "cache", ".", "bump_version", "(", "_get_namespace_for_item", "(", "item", ")", ")" ]
bump cache for given item use this only for non product items .
train
false
18,383
def marketing_link_context_processor(request): marketing_urls = configuration_helpers.get_value('MKTG_URLS', settings.MKTG_URLS) return dict([(('MKTG_URL_' + k), marketing_link(k)) for k in (settings.MKTG_URL_LINK_MAP.viewkeys() | marketing_urls.viewkeys())])
[ "def", "marketing_link_context_processor", "(", "request", ")", ":", "marketing_urls", "=", "configuration_helpers", ".", "get_value", "(", "'MKTG_URLS'", ",", "settings", ".", "MKTG_URLS", ")", "return", "dict", "(", "[", "(", "(", "'MKTG_URL_'", "+", "k", ")", ",", "marketing_link", "(", "k", ")", ")", "for", "k", "in", "(", "settings", ".", "MKTG_URL_LINK_MAP", ".", "viewkeys", "(", ")", "|", "marketing_urls", ".", "viewkeys", "(", ")", ")", "]", ")" ]
a django context processor to give templates access to marketing urls returns a dict whose keys are the marketing link names usable with the marketing_link method prefixed with mktg_url_ and whose values are the corresponding urls as computed by the marketing_link method .
train
false
18,384
def test_invalid_upgrade_strategy_causes_error(script): result = script.pip_install_local('--upgrade', '--upgrade-strategy=bazinga', 'simple', expect_error=True) assert result.returncode assert ('invalid choice' in result.stderr)
[ "def", "test_invalid_upgrade_strategy_causes_error", "(", "script", ")", ":", "result", "=", "script", ".", "pip_install_local", "(", "'--upgrade'", ",", "'--upgrade-strategy=bazinga'", ",", "'simple'", ",", "expect_error", "=", "True", ")", "assert", "result", ".", "returncode", "assert", "(", "'invalid choice'", "in", "result", ".", "stderr", ")" ]
it errors out when the upgrade-strategy is an invalid/unrecognised one .
train
false
18,385
def pivot_simple(index, columns, values): if ((len(index) != len(columns)) or (len(columns) != len(values))): raise AssertionError('Length of index, columns, and values must be the same') if (len(index) == 0): return DataFrame(index=[]) hindex = MultiIndex.from_arrays([index, columns]) series = Series(values.ravel(), index=hindex) series = series.sort_index(level=0) return series.unstack()
[ "def", "pivot_simple", "(", "index", ",", "columns", ",", "values", ")", ":", "if", "(", "(", "len", "(", "index", ")", "!=", "len", "(", "columns", ")", ")", "or", "(", "len", "(", "columns", ")", "!=", "len", "(", "values", ")", ")", ")", ":", "raise", "AssertionError", "(", "'Length of index, columns, and values must be the same'", ")", "if", "(", "len", "(", "index", ")", "==", "0", ")", ":", "return", "DataFrame", "(", "index", "=", "[", "]", ")", "hindex", "=", "MultiIndex", ".", "from_arrays", "(", "[", "index", ",", "columns", "]", ")", "series", "=", "Series", "(", "values", ".", "ravel", "(", ")", ",", "index", "=", "hindex", ")", "series", "=", "series", ".", "sort_index", "(", "level", "=", "0", ")", "return", "series", ".", "unstack", "(", ")" ]
produce pivot table based on 3 columns of this dataframe .
train
false
18,387
def blog_post_feed(request, format, **kwargs): try: return {u'rss': PostsRSS, u'atom': PostsAtom}[format](**kwargs)(request) except KeyError: raise Http404()
[ "def", "blog_post_feed", "(", "request", ",", "format", ",", "**", "kwargs", ")", ":", "try", ":", "return", "{", "u'rss'", ":", "PostsRSS", ",", "u'atom'", ":", "PostsAtom", "}", "[", "format", "]", "(", "**", "kwargs", ")", "(", "request", ")", "except", "KeyError", ":", "raise", "Http404", "(", ")" ]
blog posts feeds - maps format to the correct feed view .
train
true
18,388
def getuser(): import os for name in ('LOGNAME', 'USER', 'LNAME', 'USERNAME'): user = os.environ.get(name) if user: return user import pwd return pwd.getpwuid(os.getuid())[0]
[ "def", "getuser", "(", ")", ":", "import", "os", "for", "name", "in", "(", "'LOGNAME'", ",", "'USER'", ",", "'LNAME'", ",", "'USERNAME'", ")", ":", "user", "=", "os", ".", "environ", ".", "get", "(", "name", ")", "if", "user", ":", "return", "user", "import", "pwd", "return", "pwd", ".", "getpwuid", "(", "os", ".", "getuid", "(", ")", ")", "[", "0", "]" ]
get the username from the environment or password database .
train
false
18,389
def decorator_from_middleware_with_args(middleware_class): return make_middleware_decorator(middleware_class)
[ "def", "decorator_from_middleware_with_args", "(", "middleware_class", ")", ":", "return", "make_middleware_decorator", "(", "middleware_class", ")" ]
like decorator_from_middleware .
train
false
18,390
def _wrap_urlopen(url, timeout=None): try: raw = urlopen(url, timeout=timeout) except IOError as e: msg = (u'Error getting %s: %s' % (url, e)) log.error(msg) raise RequestException(msg) resp = requests.Response() resp.raw = raw orig_read = raw.read resp.raw.read = (lambda size, **kwargs: orig_read(size)) resp.status_code = (raw.code or 200) resp.headers = requests.structures.CaseInsensitiveDict(raw.headers) return resp
[ "def", "_wrap_urlopen", "(", "url", ",", "timeout", "=", "None", ")", ":", "try", ":", "raw", "=", "urlopen", "(", "url", ",", "timeout", "=", "timeout", ")", "except", "IOError", "as", "e", ":", "msg", "=", "(", "u'Error getting %s: %s'", "%", "(", "url", ",", "e", ")", ")", "log", ".", "error", "(", "msg", ")", "raise", "RequestException", "(", "msg", ")", "resp", "=", "requests", ".", "Response", "(", ")", "resp", ".", "raw", "=", "raw", "orig_read", "=", "raw", ".", "read", "resp", ".", "raw", ".", "read", "=", "(", "lambda", "size", ",", "**", "kwargs", ":", "orig_read", "(", "size", ")", ")", "resp", ".", "status_code", "=", "(", "raw", ".", "code", "or", "200", ")", "resp", ".", "headers", "=", "requests", ".", "structures", ".", "CaseInsensitiveDict", "(", "raw", ".", "headers", ")", "return", "resp" ]
handles alternate schemes using urllib .
train
false
18,391
def _convert_seconds(seconds): return u'{0:.0f}:{1:02.0f}'.format(*divmod(float(seconds), 60))
[ "def", "_convert_seconds", "(", "seconds", ")", ":", "return", "u'{0:.0f}:{1:02.0f}'", ".", "format", "(", "*", "divmod", "(", "float", "(", "seconds", ")", ",", "60", ")", ")" ]
convert seconds to minutes:seconds format .
train
false
18,392
def prefix_to_attr(attr_id): attr_by_prefix = {'fsmt-': 'MountTargetId', 'subnet-': 'SubnetId', 'eni-': 'NetworkInterfaceId', 'sg-': 'SecurityGroups'} prefix = first_or_default(filter((lambda pref: str(attr_id).startswith(pref)), attr_by_prefix.keys())) if prefix: return attr_by_prefix[prefix] return 'IpAddress'
[ "def", "prefix_to_attr", "(", "attr_id", ")", ":", "attr_by_prefix", "=", "{", "'fsmt-'", ":", "'MountTargetId'", ",", "'subnet-'", ":", "'SubnetId'", ",", "'eni-'", ":", "'NetworkInterfaceId'", ",", "'sg-'", ":", "'SecurityGroups'", "}", "prefix", "=", "first_or_default", "(", "filter", "(", "(", "lambda", "pref", ":", "str", "(", "attr_id", ")", ".", "startswith", "(", "pref", ")", ")", ",", "attr_by_prefix", ".", "keys", "(", ")", ")", ")", "if", "prefix", ":", "return", "attr_by_prefix", "[", "prefix", "]", "return", "'IpAddress'" ]
helper method to convert id prefix to mount target attribute .
train
false
18,393
def new_plot(parent=None, subplot_kw=None, **fig_kw): if (subplot_kw is None): subplot_kw = {} canvas = new_canvas(**fig_kw) canvas.setParent(parent) fig = canvas.figure ax = fig.add_subplot(1, 1, 1, **subplot_kw) return (fig, ax)
[ "def", "new_plot", "(", "parent", "=", "None", ",", "subplot_kw", "=", "None", ",", "**", "fig_kw", ")", ":", "if", "(", "subplot_kw", "is", "None", ")", ":", "subplot_kw", "=", "{", "}", "canvas", "=", "new_canvas", "(", "**", "fig_kw", ")", "canvas", ".", "setParent", "(", "parent", ")", "fig", "=", "canvas", ".", "figure", "ax", "=", "fig", ".", "add_subplot", "(", "1", ",", "1", ",", "1", ",", "**", "subplot_kw", ")", "return", "(", "fig", ",", "ax", ")" ]
return new figure and axes .
train
false
18,394
def _set_model_dict(resource_type_name, properties_target, prefix, created_at, updated_at): model_dict = {'name': resource_type_name, 'properties_target': properties_target, 'prefix': prefix, 'created_at': created_at, 'updated_at': updated_at} return model_dict
[ "def", "_set_model_dict", "(", "resource_type_name", ",", "properties_target", ",", "prefix", ",", "created_at", ",", "updated_at", ")", ":", "model_dict", "=", "{", "'name'", ":", "resource_type_name", ",", "'properties_target'", ":", "properties_target", ",", "'prefix'", ":", "prefix", ",", "'created_at'", ":", "created_at", ",", "'updated_at'", ":", "updated_at", "}", "return", "model_dict" ]
return a model dict set with the passed in key values .
train
false
18,395
def onInit(isReload): DEBUG_MSG(('onInit::isReload:%s' % isReload))
[ "def", "onInit", "(", "isReload", ")", ":", "DEBUG_MSG", "(", "(", "'onInit::isReload:%s'", "%", "isReload", ")", ")" ]
kbengine method .
train
false
18,396
def _modify_eni_properties(eni_id, properties=None, vm_=None): if (not isinstance(properties, dict)): raise SaltCloudException('ENI properties must be a dictionary') params = {'Action': 'ModifyNetworkInterfaceAttribute', 'NetworkInterfaceId': eni_id} for (k, v) in six.iteritems(properties): params[k] = v retries = 5 while (retries > 0): retries = (retries - 1) result = aws.query(params, return_root=True, location=get_location(vm_), provider=get_provider(), opts=__opts__, sigver='4') if (isinstance(result, dict) and result.get('error')): time.sleep(1) continue return result raise SaltCloudException("Could not change interface <{0}> attributes <'{1}'> after 5 retries".format(eni_id, properties))
[ "def", "_modify_eni_properties", "(", "eni_id", ",", "properties", "=", "None", ",", "vm_", "=", "None", ")", ":", "if", "(", "not", "isinstance", "(", "properties", ",", "dict", ")", ")", ":", "raise", "SaltCloudException", "(", "'ENI properties must be a dictionary'", ")", "params", "=", "{", "'Action'", ":", "'ModifyNetworkInterfaceAttribute'", ",", "'NetworkInterfaceId'", ":", "eni_id", "}", "for", "(", "k", ",", "v", ")", "in", "six", ".", "iteritems", "(", "properties", ")", ":", "params", "[", "k", "]", "=", "v", "retries", "=", "5", "while", "(", "retries", ">", "0", ")", ":", "retries", "=", "(", "retries", "-", "1", ")", "result", "=", "aws", ".", "query", "(", "params", ",", "return_root", "=", "True", ",", "location", "=", "get_location", "(", "vm_", ")", ",", "provider", "=", "get_provider", "(", ")", ",", "opts", "=", "__opts__", ",", "sigver", "=", "'4'", ")", "if", "(", "isinstance", "(", "result", ",", "dict", ")", "and", "result", ".", "get", "(", "'error'", ")", ")", ":", "time", ".", "sleep", "(", "1", ")", "continue", "return", "result", "raise", "SaltCloudException", "(", "\"Could not change interface <{0}> attributes <'{1}'> after 5 retries\"", ".", "format", "(", "eni_id", ",", "properties", ")", ")" ]
change properties of the interface with id eni_id to the values in properties dict .
train
true
18,398
def _fallback_next(it, default=_unspecified_next_default): if (default is _unspecified_next_default): return it.next() else: try: return it.next() except StopIteration: return default
[ "def", "_fallback_next", "(", "it", ",", "default", "=", "_unspecified_next_default", ")", ":", "if", "(", "default", "is", "_unspecified_next_default", ")", ":", "return", "it", ".", "next", "(", ")", "else", ":", "try", ":", "return", "it", ".", "next", "(", ")", "except", "StopIteration", ":", "return", "default" ]
retrieve the next item from the iterator by calling its next() method .
train
false
18,399
def add_all_srs(): q = Subreddit._query(sort=asc('_date')) for sr in fetch_things2(q): for q in all_queries(get_links, sr, ('hot', 'new'), ['all']): q.update() for q in all_queries(get_links, sr, time_filtered_sorts, db_times.keys()): q.update() get_spam_links(sr).update() get_spam_comments(sr).update() get_reported_links(sr).update() get_reported_comments(sr).update()
[ "def", "add_all_srs", "(", ")", ":", "q", "=", "Subreddit", ".", "_query", "(", "sort", "=", "asc", "(", "'_date'", ")", ")", "for", "sr", "in", "fetch_things2", "(", "q", ")", ":", "for", "q", "in", "all_queries", "(", "get_links", ",", "sr", ",", "(", "'hot'", ",", "'new'", ")", ",", "[", "'all'", "]", ")", ":", "q", ".", "update", "(", ")", "for", "q", "in", "all_queries", "(", "get_links", ",", "sr", ",", "time_filtered_sorts", ",", "db_times", ".", "keys", "(", ")", ")", ":", "q", ".", "update", "(", ")", "get_spam_links", "(", "sr", ")", ".", "update", "(", ")", "get_spam_comments", "(", "sr", ")", ".", "update", "(", ")", "get_reported_links", "(", "sr", ")", ".", "update", "(", ")", "get_reported_comments", "(", "sr", ")", ".", "update", "(", ")" ]
recalculates every listing query for every subreddit .
train
false
18,401
def strict_parse_args(parser, raw_args): args = parser.parse_args() unexpected_params = (set(raw_args) - {allowed_arg.name for allowed_arg in parser.args}) if unexpected_params: raise InputError('Unexpected query parameters {}'.format(unexpected_params)) return args
[ "def", "strict_parse_args", "(", "parser", ",", "raw_args", ")", ":", "args", "=", "parser", ".", "parse_args", "(", ")", "unexpected_params", "=", "(", "set", "(", "raw_args", ")", "-", "{", "allowed_arg", ".", "name", "for", "allowed_arg", "in", "parser", ".", "args", "}", ")", "if", "unexpected_params", ":", "raise", "InputError", "(", "'Unexpected query parameters {}'", ".", "format", "(", "unexpected_params", ")", ")", "return", "args" ]
wrapper around parser .
train
false
18,402
def getEpoch(): return str((int(time.time()) / const.EPOCH_GRANULARITY))
[ "def", "getEpoch", "(", ")", ":", "return", "str", "(", "(", "int", "(", "time", ".", "time", "(", ")", ")", "/", "const", ".", "EPOCH_GRANULARITY", ")", ")" ]
return the unix epoch divided by a constant as string .
train
false
18,404
def genFull(pset, min_, max_, type_=None): def condition(height, depth): 'Expression generation stops when the depth is equal to height.' return (depth == height) return generate(pset, min_, max_, condition, type_)
[ "def", "genFull", "(", "pset", ",", "min_", ",", "max_", ",", "type_", "=", "None", ")", ":", "def", "condition", "(", "height", ",", "depth", ")", ":", "return", "(", "depth", "==", "height", ")", "return", "generate", "(", "pset", ",", "min_", ",", "max_", ",", "condition", ",", "type_", ")" ]
generate an expression where each leaf has a the same depth between *min* and *max* .
train
false
18,405
def beaker(registry, xml_parent, data): beaker = XML.SubElement(xml_parent, 'org.jenkinsci.plugins.beakerbuilder.BeakerBuilder') jobSource = XML.SubElement(beaker, 'jobSource') if (('content' in data) and ('path' in data)): raise JenkinsJobsException("Use just one of 'content' or 'path'") elif ('content' in data): jobSourceClass = 'org.jenkinsci.plugins.beakerbuilder.StringJobSource' jobSource.set('class', jobSourceClass) XML.SubElement(jobSource, 'jobContent').text = data['content'] elif ('path' in data): jobSourceClass = 'org.jenkinsci.plugins.beakerbuilder.FileJobSource' jobSource.set('class', jobSourceClass) XML.SubElement(jobSource, 'jobPath').text = data['path'] else: raise JenkinsJobsException("Use one of 'content' or 'path'") XML.SubElement(beaker, 'downloadFiles').text = str(data.get('download-logs', False)).lower()
[ "def", "beaker", "(", "registry", ",", "xml_parent", ",", "data", ")", ":", "beaker", "=", "XML", ".", "SubElement", "(", "xml_parent", ",", "'org.jenkinsci.plugins.beakerbuilder.BeakerBuilder'", ")", "jobSource", "=", "XML", ".", "SubElement", "(", "beaker", ",", "'jobSource'", ")", "if", "(", "(", "'content'", "in", "data", ")", "and", "(", "'path'", "in", "data", ")", ")", ":", "raise", "JenkinsJobsException", "(", "\"Use just one of 'content' or 'path'\"", ")", "elif", "(", "'content'", "in", "data", ")", ":", "jobSourceClass", "=", "'org.jenkinsci.plugins.beakerbuilder.StringJobSource'", "jobSource", ".", "set", "(", "'class'", ",", "jobSourceClass", ")", "XML", ".", "SubElement", "(", "jobSource", ",", "'jobContent'", ")", ".", "text", "=", "data", "[", "'content'", "]", "elif", "(", "'path'", "in", "data", ")", ":", "jobSourceClass", "=", "'org.jenkinsci.plugins.beakerbuilder.FileJobSource'", "jobSource", ".", "set", "(", "'class'", ",", "jobSourceClass", ")", "XML", ".", "SubElement", "(", "jobSource", ",", "'jobPath'", ")", ".", "text", "=", "data", "[", "'path'", "]", "else", ":", "raise", "JenkinsJobsException", "(", "\"Use one of 'content' or 'path'\"", ")", "XML", ".", "SubElement", "(", "beaker", ",", "'downloadFiles'", ")", ".", "text", "=", "str", "(", "data", ".", "get", "(", "'download-logs'", ",", "False", ")", ")", ".", "lower", "(", ")" ]
yaml: beaker execute a beaker build step .
train
false
18,406
def parse_gff_attributes(attr_str): attributes_list = attr_str.split(';') attributes = {} for name_value_pair in attributes_list: pair = name_value_pair.strip().split('=') if (len(pair) == 1): pair = name_value_pair.strip().split('"') if (len(pair) == 1): continue if (pair == ''): continue name = pair[0].strip() if (name == ''): continue value = pair[1].strip(' "') attributes[name] = value if (len(attributes) == 0): attributes['group'] = attr_str return attributes
[ "def", "parse_gff_attributes", "(", "attr_str", ")", ":", "attributes_list", "=", "attr_str", ".", "split", "(", "';'", ")", "attributes", "=", "{", "}", "for", "name_value_pair", "in", "attributes_list", ":", "pair", "=", "name_value_pair", ".", "strip", "(", ")", ".", "split", "(", "'='", ")", "if", "(", "len", "(", "pair", ")", "==", "1", ")", ":", "pair", "=", "name_value_pair", ".", "strip", "(", ")", ".", "split", "(", "'\"'", ")", "if", "(", "len", "(", "pair", ")", "==", "1", ")", ":", "continue", "if", "(", "pair", "==", "''", ")", ":", "continue", "name", "=", "pair", "[", "0", "]", ".", "strip", "(", ")", "if", "(", "name", "==", "''", ")", ":", "continue", "value", "=", "pair", "[", "1", "]", ".", "strip", "(", "' \"'", ")", "attributes", "[", "name", "]", "=", "value", "if", "(", "len", "(", "attributes", ")", "==", "0", ")", ":", "attributes", "[", "'group'", "]", "=", "attr_str", "return", "attributes" ]
parses a gff/gtf attribute string and returns a dictionary of name-value pairs .
train
false
18,409
def _find_vector_rotation(a, b): R = np.eye(3) v = np.cross(a, b) if np.allclose(v, 0.0): return R s = np.dot(v, v) c = np.dot(a, b) vx = _skew_symmetric_cross(v) R += (vx + ((np.dot(vx, vx) * (1 - c)) / s)) return R
[ "def", "_find_vector_rotation", "(", "a", ",", "b", ")", ":", "R", "=", "np", ".", "eye", "(", "3", ")", "v", "=", "np", ".", "cross", "(", "a", ",", "b", ")", "if", "np", ".", "allclose", "(", "v", ",", "0.0", ")", ":", "return", "R", "s", "=", "np", ".", "dot", "(", "v", ",", "v", ")", "c", "=", "np", ".", "dot", "(", "a", ",", "b", ")", "vx", "=", "_skew_symmetric_cross", "(", "v", ")", "R", "+=", "(", "vx", "+", "(", "(", "np", ".", "dot", "(", "vx", ",", "vx", ")", "*", "(", "1", "-", "c", ")", ")", "/", "s", ")", ")", "return", "R" ]
find the rotation matrix that maps unit vector a to b .
train
false
18,410
def ReadFemPreg(dct_file='2002FemPreg.dct', dat_file='2002FemPreg.dat.gz'): dct = thinkstats2.ReadStataDct(dct_file) df = dct.ReadFixedWidth(dat_file, compression='gzip') CleanFemPreg(df) return df
[ "def", "ReadFemPreg", "(", "dct_file", "=", "'2002FemPreg.dct'", ",", "dat_file", "=", "'2002FemPreg.dat.gz'", ")", ":", "dct", "=", "thinkstats2", ".", "ReadStataDct", "(", "dct_file", ")", "df", "=", "dct", ".", "ReadFixedWidth", "(", "dat_file", ",", "compression", "=", "'gzip'", ")", "CleanFemPreg", "(", "df", ")", "return", "df" ]
reads the nsfg pregnancy data .
train
false
18,412
def schema_serializer_class(serializer_class, **kwargs): def decorator(func): func.schema_serializer_class = serializer_class func.kwargs = kwargs return func return decorator
[ "def", "schema_serializer_class", "(", "serializer_class", ",", "**", "kwargs", ")", ":", "def", "decorator", "(", "func", ")", ":", "func", ".", "schema_serializer_class", "=", "serializer_class", "func", ".", "kwargs", "=", "kwargs", "return", "func", "return", "decorator" ]
a decorator to set a serializer class in detail or list method of viewsets making it possible to extract the right serializer to generate the proper documentation .
train
false
18,413
def seq3(seq, custom_map=None, undef_code='Xaa'): if (custom_map is None): custom_map = {'*': 'Ter'} threecode = dict((list(IUPACData.protein_letters_1to3_extended.items()) + list(custom_map.items()))) return ''.join((threecode.get(aa, undef_code) for aa in seq))
[ "def", "seq3", "(", "seq", ",", "custom_map", "=", "None", ",", "undef_code", "=", "'Xaa'", ")", ":", "if", "(", "custom_map", "is", "None", ")", ":", "custom_map", "=", "{", "'*'", ":", "'Ter'", "}", "threecode", "=", "dict", "(", "(", "list", "(", "IUPACData", ".", "protein_letters_1to3_extended", ".", "items", "(", ")", ")", "+", "list", "(", "custom_map", ".", "items", "(", ")", ")", ")", ")", "return", "''", ".", "join", "(", "(", "threecode", ".", "get", "(", "aa", ",", "undef_code", ")", "for", "aa", "in", "seq", ")", ")" ]
turn a one letter code protein sequence into one with three letter codes .
train
false
18,414
def AppIdWithDefaultPartition(app_id, default_partition): if (not default_partition): return app_id if ('~' in app_id): return app_id return ((default_partition + '~') + app_id)
[ "def", "AppIdWithDefaultPartition", "(", "app_id", ",", "default_partition", ")", ":", "if", "(", "not", "default_partition", ")", ":", "return", "app_id", "if", "(", "'~'", "in", "app_id", ")", ":", "return", "app_id", "return", "(", "(", "default_partition", "+", "'~'", ")", "+", "app_id", ")" ]
add a partition to an application id if necessary .
train
false
18,416
def GetValidHostsForCert(cert): if ('subjectAltName' in cert): return [x[1] for x in cert['subjectAltName'] if (x[0].lower() == 'dns')] else: return [x[0][1] for x in cert['subject'] if (x[0][0].lower() == 'commonname')]
[ "def", "GetValidHostsForCert", "(", "cert", ")", ":", "if", "(", "'subjectAltName'", "in", "cert", ")", ":", "return", "[", "x", "[", "1", "]", "for", "x", "in", "cert", "[", "'subjectAltName'", "]", "if", "(", "x", "[", "0", "]", ".", "lower", "(", ")", "==", "'dns'", ")", "]", "else", ":", "return", "[", "x", "[", "0", "]", "[", "1", "]", "for", "x", "in", "cert", "[", "'subject'", "]", "if", "(", "x", "[", "0", "]", "[", "0", "]", ".", "lower", "(", ")", "==", "'commonname'", ")", "]" ]
returns a list of valid host globs for an ssl certificate .
train
true
18,420
def colon_separated_user_group(arg): try: parts = arg.split(':', 1) if (len(parts) == 1): uid = name_to_uid(parts[0]) gid = (-1) else: uid = name_to_uid(parts[0]) gid = name_to_gid(parts[1]) return (uid, gid) except: raise ValueError(('Invalid user:group definition %s' % arg))
[ "def", "colon_separated_user_group", "(", "arg", ")", ":", "try", ":", "parts", "=", "arg", ".", "split", "(", "':'", ",", "1", ")", "if", "(", "len", "(", "parts", ")", "==", "1", ")", ":", "uid", "=", "name_to_uid", "(", "parts", "[", "0", "]", ")", "gid", "=", "(", "-", "1", ")", "else", ":", "uid", "=", "name_to_uid", "(", "parts", "[", "0", "]", ")", "gid", "=", "name_to_gid", "(", "parts", "[", "1", "]", ")", "return", "(", "uid", ",", "gid", ")", "except", ":", "raise", "ValueError", "(", "(", "'Invalid user:group definition %s'", "%", "arg", ")", ")" ]
find a user id and group id from a string like user:group .
train
false
18,421
def _TestQueryFollowed(tester, user_cookie, request_dict): validator = tester.validator (user_id, device_id) = tester.GetIdsFromCookie(user_cookie) def _MakeViewpointDict(followed): 'Create a viewpoint dict from the followed object plus its\n referenced viewpoint object.\n ' viewpoint = validator.GetModelObject(Viewpoint, followed.viewpoint_id) follower = validator.GetModelObject(Follower, DBKey(followed.user_id, followed.viewpoint_id)) metadata_dict = viewpoint.MakeMetadataDict(follower) if (follower.CanViewContent() and ('cover_photo' in metadata_dict)): photo_dict = metadata_dict['cover_photo'] obj_store = ObjectStore.GetInstance(ObjectStore.PHOTO) _AddPhotoUrls(obj_store, photo_dict) return metadata_dict actual_dict = tester.SendRequest('query_followed', user_cookie, request_dict) followed = validator.QueryModelObjects(Followed, user_id, limit=request_dict.get('limit', None), start_key=request_dict.get('start_key', None)) expected_dict = {'viewpoints': [_MakeViewpointDict(f) for f in followed]} if (len(followed) > 0): expected_dict['last_key'] = followed[(-1)].sort_key tester._CompareResponseDicts('query_followed', user_id, request_dict, expected_dict, actual_dict) return actual_dict
[ "def", "_TestQueryFollowed", "(", "tester", ",", "user_cookie", ",", "request_dict", ")", ":", "validator", "=", "tester", ".", "validator", "(", "user_id", ",", "device_id", ")", "=", "tester", ".", "GetIdsFromCookie", "(", "user_cookie", ")", "def", "_MakeViewpointDict", "(", "followed", ")", ":", "viewpoint", "=", "validator", ".", "GetModelObject", "(", "Viewpoint", ",", "followed", ".", "viewpoint_id", ")", "follower", "=", "validator", ".", "GetModelObject", "(", "Follower", ",", "DBKey", "(", "followed", ".", "user_id", ",", "followed", ".", "viewpoint_id", ")", ")", "metadata_dict", "=", "viewpoint", ".", "MakeMetadataDict", "(", "follower", ")", "if", "(", "follower", ".", "CanViewContent", "(", ")", "and", "(", "'cover_photo'", "in", "metadata_dict", ")", ")", ":", "photo_dict", "=", "metadata_dict", "[", "'cover_photo'", "]", "obj_store", "=", "ObjectStore", ".", "GetInstance", "(", "ObjectStore", ".", "PHOTO", ")", "_AddPhotoUrls", "(", "obj_store", ",", "photo_dict", ")", "return", "metadata_dict", "actual_dict", "=", "tester", ".", "SendRequest", "(", "'query_followed'", ",", "user_cookie", ",", "request_dict", ")", "followed", "=", "validator", ".", "QueryModelObjects", "(", "Followed", ",", "user_id", ",", "limit", "=", "request_dict", ".", "get", "(", "'limit'", ",", "None", ")", ",", "start_key", "=", "request_dict", ".", "get", "(", "'start_key'", ",", "None", ")", ")", "expected_dict", "=", "{", "'viewpoints'", ":", "[", "_MakeViewpointDict", "(", "f", ")", "for", "f", "in", "followed", "]", "}", "if", "(", "len", "(", "followed", ")", ">", "0", ")", ":", "expected_dict", "[", "'last_key'", "]", "=", "followed", "[", "(", "-", "1", ")", "]", ".", "sort_key", "tester", ".", "_CompareResponseDicts", "(", "'query_followed'", ",", "user_id", ",", "request_dict", ",", "expected_dict", ",", "actual_dict", ")", "return", "actual_dict" ]
called by the servicetester in order to test query_followed service api call .
train
false
18,422
def FindPreviousMatchingAngleBracket(clean_lines, linenum, init_prefix): line = init_prefix nesting_stack = ['>'] while True: match = Search('^(.*)([<>(),;\\[\\]])[^<>(),;\\[\\]]*$', line) if match: operator = match.group(2) line = match.group(1) if (nesting_stack[(-1)] == '>'): if (operator in ('>', ')', ']')): nesting_stack.append(operator) elif (operator == '<'): nesting_stack.pop() if (not nesting_stack): return True elif (operator == ','): return True else: return False elif (operator in ('>', ')', ']')): nesting_stack.append(operator) elif (operator in ('(', '[')): nesting_stack.pop() else: linenum -= 1 if (linenum < 0): break line = clean_lines.elided[linenum] return False
[ "def", "FindPreviousMatchingAngleBracket", "(", "clean_lines", ",", "linenum", ",", "init_prefix", ")", ":", "line", "=", "init_prefix", "nesting_stack", "=", "[", "'>'", "]", "while", "True", ":", "match", "=", "Search", "(", "'^(.*)([<>(),;\\\\[\\\\]])[^<>(),;\\\\[\\\\]]*$'", ",", "line", ")", "if", "match", ":", "operator", "=", "match", ".", "group", "(", "2", ")", "line", "=", "match", ".", "group", "(", "1", ")", "if", "(", "nesting_stack", "[", "(", "-", "1", ")", "]", "==", "'>'", ")", ":", "if", "(", "operator", "in", "(", "'>'", ",", "')'", ",", "']'", ")", ")", ":", "nesting_stack", ".", "append", "(", "operator", ")", "elif", "(", "operator", "==", "'<'", ")", ":", "nesting_stack", ".", "pop", "(", ")", "if", "(", "not", "nesting_stack", ")", ":", "return", "True", "elif", "(", "operator", "==", "','", ")", ":", "return", "True", "else", ":", "return", "False", "elif", "(", "operator", "in", "(", "'>'", ",", "')'", ",", "']'", ")", ")", ":", "nesting_stack", ".", "append", "(", "operator", ")", "elif", "(", "operator", "in", "(", "'('", ",", "'['", ")", ")", ":", "nesting_stack", ".", "pop", "(", ")", "else", ":", "linenum", "-=", "1", "if", "(", "linenum", "<", "0", ")", ":", "break", "line", "=", "clean_lines", ".", "elided", "[", "linenum", "]", "return", "False" ]
find the corresponding < that started a template .
train
false
18,423
def whathdr(filename): f = open(filename, 'rb') h = f.read(512) for tf in tests: res = tf(h, f) if res: return res return None
[ "def", "whathdr", "(", "filename", ")", ":", "f", "=", "open", "(", "filename", ",", "'rb'", ")", "h", "=", "f", ".", "read", "(", "512", ")", "for", "tf", "in", "tests", ":", "res", "=", "tf", "(", "h", ",", "f", ")", "if", "res", ":", "return", "res", "return", "None" ]
recognize sound headers .
train
false
18,424
def send_password_changed_notice(user): if config_value('SEND_PASSWORD_CHANGE_EMAIL'): subject = config_value('EMAIL_SUBJECT_PASSWORD_CHANGE_NOTICE') send_mail(subject, user.email, 'change_notice', user=user)
[ "def", "send_password_changed_notice", "(", "user", ")", ":", "if", "config_value", "(", "'SEND_PASSWORD_CHANGE_EMAIL'", ")", ":", "subject", "=", "config_value", "(", "'EMAIL_SUBJECT_PASSWORD_CHANGE_NOTICE'", ")", "send_mail", "(", "subject", ",", "user", ".", "email", ",", "'change_notice'", ",", "user", "=", "user", ")" ]
sends the password changed notice email for the specified user .
train
true
18,425
def checksum(*paths): ret = dict() if (not paths): raise CommandExecutionError('No package files has been specified.') for package_file in paths: ret[package_file] = (bool(__salt__['file.file_exists'](package_file)) and (not __salt__['cmd.retcode'](['rpm', '-K', '--quiet', package_file], ignore_retcode=True, output_loglevel='trace', python_shell=False))) return ret
[ "def", "checksum", "(", "*", "paths", ")", ":", "ret", "=", "dict", "(", ")", "if", "(", "not", "paths", ")", ":", "raise", "CommandExecutionError", "(", "'No package files has been specified.'", ")", "for", "package_file", "in", "paths", ":", "ret", "[", "package_file", "]", "=", "(", "bool", "(", "__salt__", "[", "'file.file_exists'", "]", "(", "package_file", ")", ")", "and", "(", "not", "__salt__", "[", "'cmd.retcode'", "]", "(", "[", "'rpm'", ",", "'-K'", ",", "'--quiet'", ",", "package_file", "]", ",", "ignore_retcode", "=", "True", ",", "output_loglevel", "=", "'trace'", ",", "python_shell", "=", "False", ")", ")", ")", "return", "ret" ]
calculate standard internet checksum over data starting at startth byte skip_word: if specified .
train
false
18,427
@endpoint(u'/interface-data/tag-browser') def tag_browser(ctx, rd): (db, library_id) = get_library_data(ctx, rd)[:2] etag = (u'%s||%s||%s' % (db.last_modified(), rd.username, library_id)) etag = hashlib.sha1(etag.encode(u'utf-8')).hexdigest() def generate(): (db, library_id) = get_library_data(ctx, rd)[:2] return json(ctx, rd, tag_browser, categories_as_json(ctx, rd, db)) return rd.etagged_dynamic_response(etag, generate)
[ "@", "endpoint", "(", "u'/interface-data/tag-browser'", ")", "def", "tag_browser", "(", "ctx", ",", "rd", ")", ":", "(", "db", ",", "library_id", ")", "=", "get_library_data", "(", "ctx", ",", "rd", ")", "[", ":", "2", "]", "etag", "=", "(", "u'%s||%s||%s'", "%", "(", "db", ".", "last_modified", "(", ")", ",", "rd", ".", "username", ",", "library_id", ")", ")", "etag", "=", "hashlib", ".", "sha1", "(", "etag", ".", "encode", "(", "u'utf-8'", ")", ")", ".", "hexdigest", "(", ")", "def", "generate", "(", ")", ":", "(", "db", ",", "library_id", ")", "=", "get_library_data", "(", "ctx", ",", "rd", ")", "[", ":", "2", "]", "return", "json", "(", "ctx", ",", "rd", ",", "tag_browser", ",", "categories_as_json", "(", "ctx", ",", "rd", ",", "db", ")", ")", "return", "rd", ".", "etagged_dynamic_response", "(", "etag", ",", "generate", ")" ]
get the tag browser serialized as json optional: ?library_id=<default library>&sort_tags_by=name&partition_method=first letter &collapse_at=25&dont_collapse=&hide_empty_categories= .
train
false
18,428
def wait_on_interfaces(interfaces, timeout=10): result = defaultdict(list) timeout = (time.time() + timeout) while ((len(result) < len(interfaces)) and (time.time() < timeout)): rin = [i for i in interfaces.values()] win = [i for i in interfaces.values() if i.unsent_requests] (rout, wout, xout) = select.select(rin, win, [], 1) for interface in wout: interface.send_requests() for interface in rout: responses = interface.get_responses() if responses: result[interface.server].extend(responses) return result
[ "def", "wait_on_interfaces", "(", "interfaces", ",", "timeout", "=", "10", ")", ":", "result", "=", "defaultdict", "(", "list", ")", "timeout", "=", "(", "time", ".", "time", "(", ")", "+", "timeout", ")", "while", "(", "(", "len", "(", "result", ")", "<", "len", "(", "interfaces", ")", ")", "and", "(", "time", ".", "time", "(", ")", "<", "timeout", ")", ")", ":", "rin", "=", "[", "i", "for", "i", "in", "interfaces", ".", "values", "(", ")", "]", "win", "=", "[", "i", "for", "i", "in", "interfaces", ".", "values", "(", ")", "if", "i", ".", "unsent_requests", "]", "(", "rout", ",", "wout", ",", "xout", ")", "=", "select", ".", "select", "(", "rin", ",", "win", ",", "[", "]", ",", "1", ")", "for", "interface", "in", "wout", ":", "interface", ".", "send_requests", "(", ")", "for", "interface", "in", "rout", ":", "responses", "=", "interface", ".", "get_responses", "(", ")", "if", "responses", ":", "result", "[", "interface", ".", "server", "]", ".", "extend", "(", "responses", ")", "return", "result" ]
return a map of servers to a list of tuples .
train
false
18,429
def xml_findall(xpath): def xpath_findall(value): validate(ET.iselement, value) return value.findall(xpath) return transform(xpath_findall)
[ "def", "xml_findall", "(", "xpath", ")", ":", "def", "xpath_findall", "(", "value", ")", ":", "validate", "(", "ET", ".", "iselement", ",", "value", ")", "return", "value", ".", "findall", "(", "xpath", ")", "return", "transform", "(", "xpath_findall", ")" ]
find a list of xml elements via xpath .
train
true
18,430
def _merge_into_reversed(*iterables): return sorted(itertools.chain(*iterables), reverse=True)
[ "def", "_merge_into_reversed", "(", "*", "iterables", ")", ":", "return", "sorted", "(", "itertools", ".", "chain", "(", "*", "iterables", ")", ",", "reverse", "=", "True", ")" ]
merge multiple sorted inputs into a single output in reverse order .
train
false
18,431
def check_user_tags(dt): try: frappe.db.sql((u'select `_user_tags` from `tab%s` limit 1' % dt)) except Exception as e: if (e.args[0] == 1054): DocTags(dt).setup()
[ "def", "check_user_tags", "(", "dt", ")", ":", "try", ":", "frappe", ".", "db", ".", "sql", "(", "(", "u'select `_user_tags` from `tab%s` limit 1'", "%", "dt", ")", ")", "except", "Exception", "as", "e", ":", "if", "(", "e", ".", "args", "[", "0", "]", "==", "1054", ")", ":", "DocTags", "(", "dt", ")", ".", "setup", "(", ")" ]
if the user does not have a tags column .
train
false
18,433
def _run_server(local_addr, local_addr_family, local_socket_type, local_linger_args, remote_addr, remote_addr_family, remote_socket_type, queue): class _ThreadedTCPServer(SocketServer.ThreadingMixIn, SocketServer.TCPServer, object, ): 'Threaded streaming server for forwarding' address_family = local_addr_family socket_type = local_socket_type allow_reuse_address = True def __init__(self): handler_class_factory = partial(_TCPHandler, local_linger_args=local_linger_args, remote_addr=remote_addr, remote_addr_family=remote_addr_family, remote_socket_type=remote_socket_type) super(_ThreadedTCPServer, self).__init__(local_addr, handler_class_factory, bind_and_activate=True) server = _ThreadedTCPServer() queue.put([server.socket.family, server.server_address]) queue.close() server.serve_forever()
[ "def", "_run_server", "(", "local_addr", ",", "local_addr_family", ",", "local_socket_type", ",", "local_linger_args", ",", "remote_addr", ",", "remote_addr_family", ",", "remote_socket_type", ",", "queue", ")", ":", "class", "_ThreadedTCPServer", "(", "SocketServer", ".", "ThreadingMixIn", ",", "SocketServer", ".", "TCPServer", ",", "object", ",", ")", ":", "address_family", "=", "local_addr_family", "socket_type", "=", "local_socket_type", "allow_reuse_address", "=", "True", "def", "__init__", "(", "self", ")", ":", "handler_class_factory", "=", "partial", "(", "_TCPHandler", ",", "local_linger_args", "=", "local_linger_args", ",", "remote_addr", "=", "remote_addr", ",", "remote_addr_family", "=", "remote_addr_family", ",", "remote_socket_type", "=", "remote_socket_type", ")", "super", "(", "_ThreadedTCPServer", ",", "self", ")", ".", "__init__", "(", "local_addr", ",", "handler_class_factory", ",", "bind_and_activate", "=", "True", ")", "server", "=", "_ThreadedTCPServer", "(", ")", "queue", ".", "put", "(", "[", "server", ".", "socket", ".", "family", ",", "server", ".", "server_address", "]", ")", "queue", ".", "close", "(", ")", "server", ".", "serve_forever", "(", ")" ]
run the server; executed in the subprocess .
train
false
18,434
def text_to_segments(text): lines = text.replace(' ', ' \xa0').splitlines() if (not lines): return [] segments = [] for line in lines[:(-1)]: if line: segments.append(hangups.ChatMessageSegment(line)) segments.append(hangups.ChatMessageSegment('\n', hangups.SegmentType.LINE_BREAK)) if lines[(-1)]: segments.append(hangups.ChatMessageSegment(lines[(-1)])) return segments
[ "def", "text_to_segments", "(", "text", ")", ":", "lines", "=", "text", ".", "replace", "(", "' '", ",", "' \\xa0'", ")", ".", "splitlines", "(", ")", "if", "(", "not", "lines", ")", ":", "return", "[", "]", "segments", "=", "[", "]", "for", "line", "in", "lines", "[", ":", "(", "-", "1", ")", "]", ":", "if", "line", ":", "segments", ".", "append", "(", "hangups", ".", "ChatMessageSegment", "(", "line", ")", ")", "segments", ".", "append", "(", "hangups", ".", "ChatMessageSegment", "(", "'\\n'", ",", "hangups", ".", "SegmentType", ".", "LINE_BREAK", ")", ")", "if", "lines", "[", "(", "-", "1", ")", "]", ":", "segments", ".", "append", "(", "hangups", ".", "ChatMessageSegment", "(", "lines", "[", "(", "-", "1", ")", "]", ")", ")", "return", "segments" ]
create list of message segments from text .
train
false
18,435
def averageOnTime(vectors, numSamples=None): if (vectors.ndim == 1): vectors.shape = ((-1), 1) numTimeSteps = len(vectors) numElements = len(vectors[0]) if (numSamples is None): numSamples = numElements countOn = range(numElements) else: countOn = numpy.random.randint(0, numElements, numSamples) sumOfLengths = 0.0 onTimeFreqCounts = None n = 0 for i in countOn: (onTime, segments, durations) = _listOfOnTimesInVec(vectors[:, i]) if (onTime != 0.0): sumOfLengths += onTime n += segments onTimeFreqCounts = _accumulateFrequencyCounts(durations, onTimeFreqCounts) if (n > 0): return ((sumOfLengths / n), onTimeFreqCounts) else: return (0.0, onTimeFreqCounts)
[ "def", "averageOnTime", "(", "vectors", ",", "numSamples", "=", "None", ")", ":", "if", "(", "vectors", ".", "ndim", "==", "1", ")", ":", "vectors", ".", "shape", "=", "(", "(", "-", "1", ")", ",", "1", ")", "numTimeSteps", "=", "len", "(", "vectors", ")", "numElements", "=", "len", "(", "vectors", "[", "0", "]", ")", "if", "(", "numSamples", "is", "None", ")", ":", "numSamples", "=", "numElements", "countOn", "=", "range", "(", "numElements", ")", "else", ":", "countOn", "=", "numpy", ".", "random", ".", "randint", "(", "0", ",", "numElements", ",", "numSamples", ")", "sumOfLengths", "=", "0.0", "onTimeFreqCounts", "=", "None", "n", "=", "0", "for", "i", "in", "countOn", ":", "(", "onTime", ",", "segments", ",", "durations", ")", "=", "_listOfOnTimesInVec", "(", "vectors", "[", ":", ",", "i", "]", ")", "if", "(", "onTime", "!=", "0.0", ")", ":", "sumOfLengths", "+=", "onTime", "n", "+=", "segments", "onTimeFreqCounts", "=", "_accumulateFrequencyCounts", "(", "durations", ",", "onTimeFreqCounts", ")", "if", "(", "n", ">", "0", ")", ":", "return", "(", "(", "sumOfLengths", "/", "n", ")", ",", "onTimeFreqCounts", ")", "else", ":", "return", "(", "0.0", ",", "onTimeFreqCounts", ")" ]
returns the average on-time .
train
true
18,436
def _mkanchors(ws, hs, x_ctr, y_ctr): ws = ws[:, np.newaxis] hs = hs[:, np.newaxis] anchors = np.hstack(((x_ctr - (0.5 * (ws - 1))), (y_ctr - (0.5 * (hs - 1))), (x_ctr + (0.5 * (ws - 1))), (y_ctr + (0.5 * (hs - 1))))) return anchors
[ "def", "_mkanchors", "(", "ws", ",", "hs", ",", "x_ctr", ",", "y_ctr", ")", ":", "ws", "=", "ws", "[", ":", ",", "np", ".", "newaxis", "]", "hs", "=", "hs", "[", ":", ",", "np", ".", "newaxis", "]", "anchors", "=", "np", ".", "hstack", "(", "(", "(", "x_ctr", "-", "(", "0.5", "*", "(", "ws", "-", "1", ")", ")", ")", ",", "(", "y_ctr", "-", "(", "0.5", "*", "(", "hs", "-", "1", ")", ")", ")", ",", "(", "x_ctr", "+", "(", "0.5", "*", "(", "ws", "-", "1", ")", ")", ")", ",", "(", "y_ctr", "+", "(", "0.5", "*", "(", "hs", "-", "1", ")", ")", ")", ")", ")", "return", "anchors" ]
given a vector of widths and heights around a center .
train
true
18,437
def _decode_string_escape_py2(str_): return str_.decode('string_escape')
[ "def", "_decode_string_escape_py2", "(", "str_", ")", ":", "return", "str_", ".", "decode", "(", "'string_escape'", ")" ]
python2 string escape do not use directly .
train
false
18,438
def setConnection(uri, tables, encoding='utf8', debug=False): kw = {} _uriLower = uri.lower() if _uriLower.startswith('mysql'): kw['use_unicode'] = 1 kw['charset'] = encoding kw['local_infile'] = 1 conn = connectionForURI(uri, **kw) conn.debug = debug for table in tables: table.setConnection(conn) table._cacheValue = False conn.paramstyle = conn.module.paramstyle return conn
[ "def", "setConnection", "(", "uri", ",", "tables", ",", "encoding", "=", "'utf8'", ",", "debug", "=", "False", ")", ":", "kw", "=", "{", "}", "_uriLower", "=", "uri", ".", "lower", "(", ")", "if", "_uriLower", ".", "startswith", "(", "'mysql'", ")", ":", "kw", "[", "'use_unicode'", "]", "=", "1", "kw", "[", "'charset'", "]", "=", "encoding", "kw", "[", "'local_infile'", "]", "=", "1", "conn", "=", "connectionForURI", "(", "uri", ",", "**", "kw", ")", "conn", ".", "debug", "=", "debug", "for", "table", "in", "tables", ":", "table", ".", "setConnection", "(", "conn", ")", "table", ".", "_cacheValue", "=", "False", "conn", ".", "paramstyle", "=", "conn", ".", "module", ".", "paramstyle", "return", "conn" ]
set connection for every table .
train
false
18,440
def libvlc_audio_equalizer_set_amp_at_index(p_equalizer, f_amp, u_band): f = (_Cfunctions.get('libvlc_audio_equalizer_set_amp_at_index', None) or _Cfunction('libvlc_audio_equalizer_set_amp_at_index', ((1,), (1,), (1,)), None, ctypes.c_int, ctypes.c_void_p, ctypes.c_float, ctypes.c_uint)) return f(p_equalizer, f_amp, u_band)
[ "def", "libvlc_audio_equalizer_set_amp_at_index", "(", "p_equalizer", ",", "f_amp", ",", "u_band", ")", ":", "f", "=", "(", "_Cfunctions", ".", "get", "(", "'libvlc_audio_equalizer_set_amp_at_index'", ",", "None", ")", "or", "_Cfunction", "(", "'libvlc_audio_equalizer_set_amp_at_index'", ",", "(", "(", "1", ",", ")", ",", "(", "1", ",", ")", ",", "(", "1", ",", ")", ")", ",", "None", ",", "ctypes", ".", "c_int", ",", "ctypes", ".", "c_void_p", ",", "ctypes", ".", "c_float", ",", "ctypes", ".", "c_uint", ")", ")", "return", "f", "(", "p_equalizer", ",", "f_amp", ",", "u_band", ")" ]
set a new amplification value for a particular equalizer frequency band .
train
true
18,441
def createCertRequest(pkey, digest='md5', **name): req = crypto.X509Req() subj = req.get_subject() for (key, value) in name.items(): setattr(subj, key, value) req.set_pubkey(pkey) req.sign(pkey, digest) return req
[ "def", "createCertRequest", "(", "pkey", ",", "digest", "=", "'md5'", ",", "**", "name", ")", ":", "req", "=", "crypto", ".", "X509Req", "(", ")", "subj", "=", "req", ".", "get_subject", "(", ")", "for", "(", "key", ",", "value", ")", "in", "name", ".", "items", "(", ")", ":", "setattr", "(", "subj", ",", "key", ",", "value", ")", "req", ".", "set_pubkey", "(", "pkey", ")", "req", ".", "sign", "(", "pkey", ",", "digest", ")", "return", "req" ]
create a certificate request .
train
true
18,443
@csrf_exempt def openid_login_complete(request, redirect_field_name=REDIRECT_FIELD_NAME, render_failure=None): render_failure = (render_failure or default_render_failure) openid_response = openid_views.parse_openid_response(request) if (not openid_response): return render_failure(request, 'This is an OpenID relying party endpoint.') if (openid_response.status == SUCCESS): external_id = openid_response.identity_url oid_backend = openid_auth.OpenIDBackend() details = oid_backend._extract_user_details(openid_response) log.debug('openid success, details=%s', details) url = getattr(settings, 'OPENID_SSO_SERVER_URL', None) external_domain = '{0}{1}'.format(OPENID_DOMAIN_PREFIX, url) fullname = ('%s %s' % (details.get('first_name', ''), details.get('last_name', ''))) return _external_login_or_signup(request, external_id, external_domain, details, details.get('email', ''), fullname, retfun=functools.partial(redirect, get_next_url_for_login_page(request))) return render_failure(request, 'Openid failure')
[ "@", "csrf_exempt", "def", "openid_login_complete", "(", "request", ",", "redirect_field_name", "=", "REDIRECT_FIELD_NAME", ",", "render_failure", "=", "None", ")", ":", "render_failure", "=", "(", "render_failure", "or", "default_render_failure", ")", "openid_response", "=", "openid_views", ".", "parse_openid_response", "(", "request", ")", "if", "(", "not", "openid_response", ")", ":", "return", "render_failure", "(", "request", ",", "'This is an OpenID relying party endpoint.'", ")", "if", "(", "openid_response", ".", "status", "==", "SUCCESS", ")", ":", "external_id", "=", "openid_response", ".", "identity_url", "oid_backend", "=", "openid_auth", ".", "OpenIDBackend", "(", ")", "details", "=", "oid_backend", ".", "_extract_user_details", "(", "openid_response", ")", "log", ".", "debug", "(", "'openid success, details=%s'", ",", "details", ")", "url", "=", "getattr", "(", "settings", ",", "'OPENID_SSO_SERVER_URL'", ",", "None", ")", "external_domain", "=", "'{0}{1}'", ".", "format", "(", "OPENID_DOMAIN_PREFIX", ",", "url", ")", "fullname", "=", "(", "'%s %s'", "%", "(", "details", ".", "get", "(", "'first_name'", ",", "''", ")", ",", "details", ".", "get", "(", "'last_name'", ",", "''", ")", ")", ")", "return", "_external_login_or_signup", "(", "request", ",", "external_id", ",", "external_domain", ",", "details", ",", "details", ".", "get", "(", "'email'", ",", "''", ")", ",", "fullname", ",", "retfun", "=", "functools", ".", "partial", "(", "redirect", ",", "get_next_url_for_login_page", "(", "request", ")", ")", ")", "return", "render_failure", "(", "request", ",", "'Openid failure'", ")" ]
complete the openid login process .
train
false
18,444
def from_key_val_list(value): if (value is None): return None if isinstance(value, (str, bytes, bool, int)): raise ValueError('cannot encode objects that are not 2-tuples') return OrderedDict(value)
[ "def", "from_key_val_list", "(", "value", ")", ":", "if", "(", "value", "is", "None", ")", ":", "return", "None", "if", "isinstance", "(", "value", ",", "(", "str", ",", "bytes", ",", "bool", ",", "int", ")", ")", ":", "raise", "ValueError", "(", "'cannot encode objects that are not 2-tuples'", ")", "return", "OrderedDict", "(", "value", ")" ]
take an object and test to see if it can be represented as a dictionary .
train
true
18,445
def floatformat(text, arg=(-1)): try: input_val = force_unicode(text) d = Decimal(input_val) except UnicodeEncodeError: return u'' except InvalidOperation: if (input_val in special_floats): return input_val try: d = Decimal(force_unicode(float(text))) except (ValueError, InvalidOperation, TypeError, UnicodeEncodeError): return u'' try: p = int(arg) except ValueError: return input_val try: m = (int(d) - d) except (ValueError, OverflowError, InvalidOperation): return input_val if ((not m) and (p < 0)): return mark_safe(formats.number_format((u'%d' % int(d)), 0)) if (p == 0): exp = Decimal(1) else: exp = (Decimal('1.0') / (Decimal(10) ** abs(p))) try: return mark_safe(formats.number_format((u'%s' % str(d.quantize(exp, ROUND_HALF_UP))), abs(p))) except InvalidOperation: return input_val
[ "def", "floatformat", "(", "text", ",", "arg", "=", "(", "-", "1", ")", ")", ":", "try", ":", "input_val", "=", "force_unicode", "(", "text", ")", "d", "=", "Decimal", "(", "input_val", ")", "except", "UnicodeEncodeError", ":", "return", "u''", "except", "InvalidOperation", ":", "if", "(", "input_val", "in", "special_floats", ")", ":", "return", "input_val", "try", ":", "d", "=", "Decimal", "(", "force_unicode", "(", "float", "(", "text", ")", ")", ")", "except", "(", "ValueError", ",", "InvalidOperation", ",", "TypeError", ",", "UnicodeEncodeError", ")", ":", "return", "u''", "try", ":", "p", "=", "int", "(", "arg", ")", "except", "ValueError", ":", "return", "input_val", "try", ":", "m", "=", "(", "int", "(", "d", ")", "-", "d", ")", "except", "(", "ValueError", ",", "OverflowError", ",", "InvalidOperation", ")", ":", "return", "input_val", "if", "(", "(", "not", "m", ")", "and", "(", "p", "<", "0", ")", ")", ":", "return", "mark_safe", "(", "formats", ".", "number_format", "(", "(", "u'%d'", "%", "int", "(", "d", ")", ")", ",", "0", ")", ")", "if", "(", "p", "==", "0", ")", ":", "exp", "=", "Decimal", "(", "1", ")", "else", ":", "exp", "=", "(", "Decimal", "(", "'1.0'", ")", "/", "(", "Decimal", "(", "10", ")", "**", "abs", "(", "p", ")", ")", ")", "try", ":", "return", "mark_safe", "(", "formats", ".", "number_format", "(", "(", "u'%s'", "%", "str", "(", "d", ".", "quantize", "(", "exp", ",", "ROUND_HALF_UP", ")", ")", ")", ",", "abs", "(", "p", ")", ")", ")", "except", "InvalidOperation", ":", "return", "input_val" ]
if called without an argument .
train
false
18,446
def parse_ntlm_chal(msg2, ack): global challenge_acks Signature = msg2[0:8] try: msg_type = struct.unpack('<I', msg2[8:12])[0] except Exception: return assert (msg_type == 2) ServerChallenge = msg2[24:32].encode('hex') if (len(challenge_acks) > 50): challenge_acks.popitem(last=False) challenge_acks[ack] = ServerChallenge
[ "def", "parse_ntlm_chal", "(", "msg2", ",", "ack", ")", ":", "global", "challenge_acks", "Signature", "=", "msg2", "[", "0", ":", "8", "]", "try", ":", "msg_type", "=", "struct", ".", "unpack", "(", "'<I'", ",", "msg2", "[", "8", ":", "12", "]", ")", "[", "0", "]", "except", "Exception", ":", "return", "assert", "(", "msg_type", "==", "2", ")", "ServerChallenge", "=", "msg2", "[", "24", ":", "32", "]", ".", "encode", "(", "'hex'", ")", "if", "(", "len", "(", "challenge_acks", ")", ">", "50", ")", ":", "challenge_acks", ".", "popitem", "(", "last", "=", "False", ")", "challenge_acks", "[", "ack", "]", "=", "ServerChallenge" ]
parse server challenge .
train
false
18,447
def default_category_for_module(module): if isinstance(module, str): module = __import__(module, fromlist=['']) name = module.__name__.rsplit('.', 1)[(-1)] qualified_name = module.__name__ return CategoryDescription(name=name, qualified_name=qualified_name)
[ "def", "default_category_for_module", "(", "module", ")", ":", "if", "isinstance", "(", "module", ",", "str", ")", ":", "module", "=", "__import__", "(", "module", ",", "fromlist", "=", "[", "''", "]", ")", "name", "=", "module", ".", "__name__", ".", "rsplit", "(", "'.'", ",", "1", ")", "[", "(", "-", "1", ")", "]", "qualified_name", "=", "module", ".", "__name__", "return", "CategoryDescription", "(", "name", "=", "name", ",", "qualified_name", "=", "qualified_name", ")" ]
return a default constructed :class:categorydescription for a module .
train
false
18,448
def cross_domain(origin_check=is_trusted_origin, **options): def cross_domain_wrap(fn): cors_perms = {'origin_check': origin_check, 'allow_credentials': bool(options.get('allow_credentials'))} @wraps(fn) def cross_domain_handler(self, *args, **kwargs): if (request.params.get('hoist') == 'cookie'): if cors_perms['origin_check'](g.origin): name = request.environ['pylons.routes_dict']['action_name'] resp = fn(self, *args, **kwargs) c.cookies.add(('hoist_%s' % name), ''.join(tup(resp))) response.content_type = 'text/html' return '' else: abort(403) else: self.check_cors() return fn(self, *args, **kwargs) cross_domain_handler.cors_perms = cors_perms return cross_domain_handler return cross_domain_wrap
[ "def", "cross_domain", "(", "origin_check", "=", "is_trusted_origin", ",", "**", "options", ")", ":", "def", "cross_domain_wrap", "(", "fn", ")", ":", "cors_perms", "=", "{", "'origin_check'", ":", "origin_check", ",", "'allow_credentials'", ":", "bool", "(", "options", ".", "get", "(", "'allow_credentials'", ")", ")", "}", "@", "wraps", "(", "fn", ")", "def", "cross_domain_handler", "(", "self", ",", "*", "args", ",", "**", "kwargs", ")", ":", "if", "(", "request", ".", "params", ".", "get", "(", "'hoist'", ")", "==", "'cookie'", ")", ":", "if", "cors_perms", "[", "'origin_check'", "]", "(", "g", ".", "origin", ")", ":", "name", "=", "request", ".", "environ", "[", "'pylons.routes_dict'", "]", "[", "'action_name'", "]", "resp", "=", "fn", "(", "self", ",", "*", "args", ",", "**", "kwargs", ")", "c", ".", "cookies", ".", "add", "(", "(", "'hoist_%s'", "%", "name", ")", ",", "''", ".", "join", "(", "tup", "(", "resp", ")", ")", ")", "response", ".", "content_type", "=", "'text/html'", "return", "''", "else", ":", "abort", "(", "403", ")", "else", ":", "self", ".", "check_cors", "(", ")", "return", "fn", "(", "self", ",", "*", "args", ",", "**", "kwargs", ")", "cross_domain_handler", ".", "cors_perms", "=", "cors_perms", "return", "cross_domain_handler", "return", "cross_domain_wrap" ]
set up cross domain validation and hoisting for a request handler .
train
false
18,449
def IdToCounter(k): if (k > _MAX_SCATTERED_ID): return (0, SCATTERED) elif ((k > _MAX_SEQUENTIAL_ID) and (k <= _MAX_SCATTERED_ID)): return (long((ReverseBitsInt64(k) >> _SCATTER_SHIFT)), SCATTERED) elif (k > 0): return (long(k), SEQUENTIAL) else: raise datastore_errors.BadArgumentError(('invalid id (%d)' % k))
[ "def", "IdToCounter", "(", "k", ")", ":", "if", "(", "k", ">", "_MAX_SCATTERED_ID", ")", ":", "return", "(", "0", ",", "SCATTERED", ")", "elif", "(", "(", "k", ">", "_MAX_SEQUENTIAL_ID", ")", "and", "(", "k", "<=", "_MAX_SCATTERED_ID", ")", ")", ":", "return", "(", "long", "(", "(", "ReverseBitsInt64", "(", "k", ")", ">>", "_SCATTER_SHIFT", ")", ")", ",", "SCATTERED", ")", "elif", "(", "k", ">", "0", ")", ":", "return", "(", "long", "(", "k", ")", ",", "SEQUENTIAL", ")", "else", ":", "raise", "datastore_errors", ".", "BadArgumentError", "(", "(", "'invalid id (%d)'", "%", "k", ")", ")" ]
map id k to the counter value from which it was generated .
train
false
18,450
def get_ofp_module(ofp_version): return get_ofp_modules()[ofp_version]
[ "def", "get_ofp_module", "(", "ofp_version", ")", ":", "return", "get_ofp_modules", "(", ")", "[", "ofp_version", "]" ]
get modules pair for the constants and parser of of-wire of a given of version .
train
false
18,451
def pseudorandom(n, p, random_state=None): import numpy as np p = list(p) cp = np.cumsum(([0] + p)) assert np.allclose(1, cp[(-1)]) assert (len(p) < 256) if (not isinstance(random_state, np.random.RandomState)): random_state = np.random.RandomState(random_state) x = random_state.random_sample(n) out = np.empty(n, dtype='i1') for (i, (low, high)) in enumerate(zip(cp[:(-1)], cp[1:])): out[((x >= low) & (x < high))] = i return out
[ "def", "pseudorandom", "(", "n", ",", "p", ",", "random_state", "=", "None", ")", ":", "import", "numpy", "as", "np", "p", "=", "list", "(", "p", ")", "cp", "=", "np", ".", "cumsum", "(", "(", "[", "0", "]", "+", "p", ")", ")", "assert", "np", ".", "allclose", "(", "1", ",", "cp", "[", "(", "-", "1", ")", "]", ")", "assert", "(", "len", "(", "p", ")", "<", "256", ")", "if", "(", "not", "isinstance", "(", "random_state", ",", "np", ".", "random", ".", "RandomState", ")", ")", ":", "random_state", "=", "np", ".", "random", ".", "RandomState", "(", "random_state", ")", "x", "=", "random_state", ".", "random_sample", "(", "n", ")", "out", "=", "np", ".", "empty", "(", "n", ",", "dtype", "=", "'i1'", ")", "for", "(", "i", ",", "(", "low", ",", "high", ")", ")", "in", "enumerate", "(", "zip", "(", "cp", "[", ":", "(", "-", "1", ")", "]", ",", "cp", "[", "1", ":", "]", ")", ")", ":", "out", "[", "(", "(", "x", ">=", "low", ")", "&", "(", "x", "<", "high", ")", ")", "]", "=", "i", "return", "out" ]
pseudorandom array of integer indexes .
train
false
18,452
def defer_succeed(result): d = defer.Deferred() reactor.callLater(0, d.callback, result) return d
[ "def", "defer_succeed", "(", "result", ")", ":", "d", "=", "defer", ".", "Deferred", "(", ")", "reactor", ".", "callLater", "(", "0", ",", "d", ".", "callback", ",", "result", ")", "return", "d" ]
same as twisted .
train
false
18,453
def pkcs_ilen(n): i = 0 while (n > 0): n >>= 8 i += 1 return i
[ "def", "pkcs_ilen", "(", "n", ")", ":", "i", "=", "0", "while", "(", "n", ">", "0", ")", ":", "n", ">>=", "8", "i", "+=", "1", "return", "i" ]
this is a log base 256 which determines the minimum octet string length for unequivocal representation of integer n by pkcs_i2osp .
train
false
18,454
def transferClosestPaths(oldOrderedLocation, remainingPaths, skein): while (len(remainingPaths) > 0): transferClosestPath(oldOrderedLocation, remainingPaths, skein)
[ "def", "transferClosestPaths", "(", "oldOrderedLocation", ",", "remainingPaths", ",", "skein", ")", ":", "while", "(", "len", "(", "remainingPaths", ")", ">", "0", ")", ":", "transferClosestPath", "(", "oldOrderedLocation", ",", "remainingPaths", ",", "skein", ")" ]
transfer the closest remaining paths .
train
false
18,455
def test_transforms(): xfm = np.random.randn(4, 4).astype(np.float32) new_xfm = xfm.dot(rotate(180, (1, 0, 0)).dot(rotate((-90), (0, 1, 0)))) new_xfm = new_xfm.dot(rotate(90, (0, 0, 1)).dot(rotate(90, (0, 1, 0)))) new_xfm = new_xfm.dot(rotate(90, (1, 0, 0))) assert_allclose(xfm, new_xfm) new_xfm = translate((1, (-1), 1)).dot(translate(((-1), 1, (-1)))).dot(xfm) assert_allclose(xfm, new_xfm) new_xfm = scale((1, 2, 3)).dot(scale((1, (1.0 / 2.0), (1.0 / 3.0)))).dot(xfm) assert_allclose(xfm, new_xfm) xfm = ortho((-1), 1, (-1), 1, (-1), 1) assert_equal(xfm.shape, (4, 4)) xfm = frustum((-1), 1, (-1), 1, (-1), 1) assert_equal(xfm.shape, (4, 4)) xfm = perspective(1, 1, (-1), 1) assert_equal(xfm.shape, (4, 4))
[ "def", "test_transforms", "(", ")", ":", "xfm", "=", "np", ".", "random", ".", "randn", "(", "4", ",", "4", ")", ".", "astype", "(", "np", ".", "float32", ")", "new_xfm", "=", "xfm", ".", "dot", "(", "rotate", "(", "180", ",", "(", "1", ",", "0", ",", "0", ")", ")", ".", "dot", "(", "rotate", "(", "(", "-", "90", ")", ",", "(", "0", ",", "1", ",", "0", ")", ")", ")", ")", "new_xfm", "=", "new_xfm", ".", "dot", "(", "rotate", "(", "90", ",", "(", "0", ",", "0", ",", "1", ")", ")", ".", "dot", "(", "rotate", "(", "90", ",", "(", "0", ",", "1", ",", "0", ")", ")", ")", ")", "new_xfm", "=", "new_xfm", ".", "dot", "(", "rotate", "(", "90", ",", "(", "1", ",", "0", ",", "0", ")", ")", ")", "assert_allclose", "(", "xfm", ",", "new_xfm", ")", "new_xfm", "=", "translate", "(", "(", "1", ",", "(", "-", "1", ")", ",", "1", ")", ")", ".", "dot", "(", "translate", "(", "(", "(", "-", "1", ")", ",", "1", ",", "(", "-", "1", ")", ")", ")", ")", ".", "dot", "(", "xfm", ")", "assert_allclose", "(", "xfm", ",", "new_xfm", ")", "new_xfm", "=", "scale", "(", "(", "1", ",", "2", ",", "3", ")", ")", ".", "dot", "(", "scale", "(", "(", "1", ",", "(", "1.0", "/", "2.0", ")", ",", "(", "1.0", "/", "3.0", ")", ")", ")", ")", ".", "dot", "(", "xfm", ")", "assert_allclose", "(", "xfm", ",", "new_xfm", ")", "xfm", "=", "ortho", "(", "(", "-", "1", ")", ",", "1", ",", "(", "-", "1", ")", ",", "1", ",", "(", "-", "1", ")", ",", "1", ")", "assert_equal", "(", "xfm", ".", "shape", ",", "(", "4", ",", "4", ")", ")", "xfm", "=", "frustum", "(", "(", "-", "1", ")", ",", "1", ",", "(", "-", "1", ")", ",", "1", ",", "(", "-", "1", ")", ",", "1", ")", "assert_equal", "(", "xfm", ".", "shape", ",", "(", "4", ",", "4", ")", ")", "xfm", "=", "perspective", "(", "1", ",", "1", ",", "(", "-", "1", ")", ",", "1", ")", "assert_equal", "(", "xfm", ".", "shape", ",", "(", "4", ",", "4", ")", ")" ]
test transformations .
train
false
18,457
def FormatForAX(text): return ExpandTabs(AddCR(text))
[ "def", "FormatForAX", "(", "text", ")", ":", "return", "ExpandTabs", "(", "AddCR", "(", "text", ")", ")" ]
format a string suitable for an ax host .
train
false
18,460
def interface_ip(iface): (iface_info, error) = _get_iface_info(iface) if (error is False): return iface_info.get(iface, {}).get('inet', {})[0].get('address', '') else: return error
[ "def", "interface_ip", "(", "iface", ")", ":", "(", "iface_info", ",", "error", ")", "=", "_get_iface_info", "(", "iface", ")", "if", "(", "error", "is", "False", ")", ":", "return", "iface_info", ".", "get", "(", "iface", ",", "{", "}", ")", ".", "get", "(", "'inet'", ",", "{", "}", ")", "[", "0", "]", ".", "get", "(", "'address'", ",", "''", ")", "else", ":", "return", "error" ]
return the inet address for a given interface .
train
true
18,461
def importTrialTypes(fileName, returnFieldNames=False): logging.warning('importTrialTypes is DEPRECATED (as of v1.70.00). Please use `importConditions` for identical functionality.') return importConditions(fileName, returnFieldNames)
[ "def", "importTrialTypes", "(", "fileName", ",", "returnFieldNames", "=", "False", ")", ":", "logging", ".", "warning", "(", "'importTrialTypes is DEPRECATED (as of v1.70.00). Please use `importConditions` for identical functionality.'", ")", "return", "importConditions", "(", "fileName", ",", "returnFieldNames", ")" ]
importtrialtypes is deprecated please use importconditions for identical functionality .
train
false
18,464
def AceIterator(handle): for ace_contig in Ace.parse(handle): consensus_seq_str = ace_contig.sequence if ('U' in consensus_seq_str): if ('T' in consensus_seq_str): alpha = generic_nucleotide else: alpha = generic_rna else: alpha = generic_dna if ('*' in consensus_seq_str): assert ('-' not in consensus_seq_str) consensus_seq = Seq(consensus_seq_str.replace('*', '-'), Gapped(alpha, gap_char='-')) else: consensus_seq = Seq(consensus_seq_str, alpha) seq_record = SeqRecord(consensus_seq, id=ace_contig.name, name=ace_contig.name) quals = [] i = 0 for base in consensus_seq: if (base == '-'): quals.append(0) else: quals.append(ace_contig.quality[i]) i += 1 assert (i == len(ace_contig.quality)) seq_record.letter_annotations['phred_quality'] = quals (yield seq_record)
[ "def", "AceIterator", "(", "handle", ")", ":", "for", "ace_contig", "in", "Ace", ".", "parse", "(", "handle", ")", ":", "consensus_seq_str", "=", "ace_contig", ".", "sequence", "if", "(", "'U'", "in", "consensus_seq_str", ")", ":", "if", "(", "'T'", "in", "consensus_seq_str", ")", ":", "alpha", "=", "generic_nucleotide", "else", ":", "alpha", "=", "generic_rna", "else", ":", "alpha", "=", "generic_dna", "if", "(", "'*'", "in", "consensus_seq_str", ")", ":", "assert", "(", "'-'", "not", "in", "consensus_seq_str", ")", "consensus_seq", "=", "Seq", "(", "consensus_seq_str", ".", "replace", "(", "'*'", ",", "'-'", ")", ",", "Gapped", "(", "alpha", ",", "gap_char", "=", "'-'", ")", ")", "else", ":", "consensus_seq", "=", "Seq", "(", "consensus_seq_str", ",", "alpha", ")", "seq_record", "=", "SeqRecord", "(", "consensus_seq", ",", "id", "=", "ace_contig", ".", "name", ",", "name", "=", "ace_contig", ".", "name", ")", "quals", "=", "[", "]", "i", "=", "0", "for", "base", "in", "consensus_seq", ":", "if", "(", "base", "==", "'-'", ")", ":", "quals", ".", "append", "(", "0", ")", "else", ":", "quals", ".", "append", "(", "ace_contig", ".", "quality", "[", "i", "]", ")", "i", "+=", "1", "assert", "(", "i", "==", "len", "(", "ace_contig", ".", "quality", ")", ")", "seq_record", ".", "letter_annotations", "[", "'phred_quality'", "]", "=", "quals", "(", "yield", "seq_record", ")" ]
returns seqrecord objects from an ace file .
train
false
18,466
def getMinimumFromVec3List(vec3List): minimum = complex(999999999.0, 999999999.0) for point in vec3List: minimum = getMinimum(minimum, point.dropAxis(2)) return minimum
[ "def", "getMinimumFromVec3List", "(", "vec3List", ")", ":", "minimum", "=", "complex", "(", "999999999.0", ",", "999999999.0", ")", "for", "point", "in", "vec3List", ":", "minimum", "=", "getMinimum", "(", "minimum", ",", "point", ".", "dropAxis", "(", "2", ")", ")", "return", "minimum" ]
get a complex with each component the minimum of the respective components of a list of vector3s .
train
false
18,468
def get_cpu_info(): try: with open(CPU_INFO_PATH) as fp: content = fp.read() except IOError: return {} lines = content.split('\n') result = [] item = None lines_count = len(lines) for (index, line) in enumerate(lines): line = line.strip() if (not line): if (item and (index != lines_count)): result.append(item) continue split = line.split(':') if (len(split) != 2): continue name = split[0].replace(' DCTB ', '').strip().replace(' ', '_') value = split[1].replace(' DCTB ', '').strip() if (name == 'processor'): item = {} item[name] = value return result
[ "def", "get_cpu_info", "(", ")", ":", "try", ":", "with", "open", "(", "CPU_INFO_PATH", ")", "as", "fp", ":", "content", "=", "fp", ".", "read", "(", ")", "except", "IOError", ":", "return", "{", "}", "lines", "=", "content", ".", "split", "(", "'\\n'", ")", "result", "=", "[", "]", "item", "=", "None", "lines_count", "=", "len", "(", "lines", ")", "for", "(", "index", ",", "line", ")", "in", "enumerate", "(", "lines", ")", ":", "line", "=", "line", ".", "strip", "(", ")", "if", "(", "not", "line", ")", ":", "if", "(", "item", "and", "(", "index", "!=", "lines_count", ")", ")", ":", "result", ".", "append", "(", "item", ")", "continue", "split", "=", "line", ".", "split", "(", "':'", ")", "if", "(", "len", "(", "split", ")", "!=", "2", ")", ":", "continue", "name", "=", "split", "[", "0", "]", ".", "replace", "(", "' DCTB '", ",", "''", ")", ".", "strip", "(", ")", ".", "replace", "(", "' '", ",", "'_'", ")", "value", "=", "split", "[", "1", "]", ".", "replace", "(", "' DCTB '", ",", "''", ")", ".", "strip", "(", ")", "if", "(", "name", "==", "'processor'", ")", ":", "item", "=", "{", "}", "item", "[", "name", "]", "=", "value", "return", "result" ]
retrieve cpu information .
train
false
18,469
def group_status(): return s3_rest_controller()
[ "def", "group_status", "(", ")", ":", "return", "s3_rest_controller", "(", ")" ]
group statuses: restful crud controller .
train
false
18,470
def get_all_alarms(region=None, prefix=None, key=None, keyid=None, profile=None): conn = _get_conn(region=region, key=key, keyid=keyid, profile=profile) alarms = conn.describe_alarms() results = odict.OrderedDict() for alarm in alarms: alarm = _metric_alarm_to_dict(alarm) name = alarm['name'] if prefix: if name.startswith(prefix): continue name = (prefix + alarm['name']) del alarm['name'] alarm_sls = [] alarm_sls.append({'name': name}) alarm_sls.append({'attributes': alarm}) results[('manage alarm ' + name)] = {'boto_cloudwatch_alarm.present': alarm_sls} return _safe_dump(results)
[ "def", "get_all_alarms", "(", "region", "=", "None", ",", "prefix", "=", "None", ",", "key", "=", "None", ",", "keyid", "=", "None", ",", "profile", "=", "None", ")", ":", "conn", "=", "_get_conn", "(", "region", "=", "region", ",", "key", "=", "key", ",", "keyid", "=", "keyid", ",", "profile", "=", "profile", ")", "alarms", "=", "conn", ".", "describe_alarms", "(", ")", "results", "=", "odict", ".", "OrderedDict", "(", ")", "for", "alarm", "in", "alarms", ":", "alarm", "=", "_metric_alarm_to_dict", "(", "alarm", ")", "name", "=", "alarm", "[", "'name'", "]", "if", "prefix", ":", "if", "name", ".", "startswith", "(", "prefix", ")", ":", "continue", "name", "=", "(", "prefix", "+", "alarm", "[", "'name'", "]", ")", "del", "alarm", "[", "'name'", "]", "alarm_sls", "=", "[", "]", "alarm_sls", ".", "append", "(", "{", "'name'", ":", "name", "}", ")", "alarm_sls", ".", "append", "(", "{", "'attributes'", ":", "alarm", "}", ")", "results", "[", "(", "'manage alarm '", "+", "name", ")", "]", "=", "{", "'boto_cloudwatch_alarm.present'", ":", "alarm_sls", "}", "return", "_safe_dump", "(", "results", ")" ]
get all alarm details .
train
true
18,471
@register.filter(name='widget_type') def widget_type(field): try: return field.field.widget.__class__.__name__.lower() except AttributeError: return None
[ "@", "register", ".", "filter", "(", "name", "=", "'widget_type'", ")", "def", "widget_type", "(", "field", ")", ":", "try", ":", "return", "field", ".", "field", ".", "widget", ".", "__class__", ".", "__name__", ".", "lower", "(", ")", "except", "AttributeError", ":", "return", "None" ]
return the widget type .
train
false
18,473
def str_endswith(arr, pat, na=np.nan): f = (lambda x: x.endswith(pat)) return _na_map(f, arr, na, dtype=bool)
[ "def", "str_endswith", "(", "arr", ",", "pat", ",", "na", "=", "np", ".", "nan", ")", ":", "f", "=", "(", "lambda", "x", ":", "x", ".", "endswith", "(", "pat", ")", ")", "return", "_na_map", "(", "f", ",", "arr", ",", "na", ",", "dtype", "=", "bool", ")" ]
return boolean series indicating whether each string in the series/index ends with passed pattern .
train
true
18,475
def message_get(context, message_id): return IMPL.message_get(context, message_id)
[ "def", "message_get", "(", "context", ",", "message_id", ")", ":", "return", "IMPL", ".", "message_get", "(", "context", ",", "message_id", ")" ]
return a message with the specified id .
train
false
18,476
def single_source_shortest_path(G, source, cutoff=None): if (source not in G): raise nx.NodeNotFound('Source {} not in G'.format(source)) level = 0 nextlevel = {source: 1} paths = {source: [source]} if (cutoff == 0): return paths while nextlevel: thislevel = nextlevel nextlevel = {} for v in thislevel: for w in G[v]: if (w not in paths): paths[w] = (paths[v] + [w]) nextlevel[w] = 1 level = (level + 1) if ((cutoff is not None) and (cutoff <= level)): break return paths
[ "def", "single_source_shortest_path", "(", "G", ",", "source", ",", "cutoff", "=", "None", ")", ":", "if", "(", "source", "not", "in", "G", ")", ":", "raise", "nx", ".", "NodeNotFound", "(", "'Source {} not in G'", ".", "format", "(", "source", ")", ")", "level", "=", "0", "nextlevel", "=", "{", "source", ":", "1", "}", "paths", "=", "{", "source", ":", "[", "source", "]", "}", "if", "(", "cutoff", "==", "0", ")", ":", "return", "paths", "while", "nextlevel", ":", "thislevel", "=", "nextlevel", "nextlevel", "=", "{", "}", "for", "v", "in", "thislevel", ":", "for", "w", "in", "G", "[", "v", "]", ":", "if", "(", "w", "not", "in", "paths", ")", ":", "paths", "[", "w", "]", "=", "(", "paths", "[", "v", "]", "+", "[", "w", "]", ")", "nextlevel", "[", "w", "]", "=", "1", "level", "=", "(", "level", "+", "1", ")", "if", "(", "(", "cutoff", "is", "not", "None", ")", "and", "(", "cutoff", "<=", "level", ")", ")", ":", "break", "return", "paths" ]
compute shortest path between source and all other nodes reachable from source .
train
false
18,478
def check_cdf_slope(cdf): norm_intensity = np.linspace(0, 1, len(cdf)) (slope, intercept) = np.polyfit(norm_intensity, cdf, 1) assert (0.9 < slope < 1.1)
[ "def", "check_cdf_slope", "(", "cdf", ")", ":", "norm_intensity", "=", "np", ".", "linspace", "(", "0", ",", "1", ",", "len", "(", "cdf", ")", ")", "(", "slope", ",", "intercept", ")", "=", "np", ".", "polyfit", "(", "norm_intensity", ",", "cdf", ",", "1", ")", "assert", "(", "0.9", "<", "slope", "<", "1.1", ")" ]
slope of cdf which should equal 1 for an equalized histogram .
train
false
18,479
def _ExpandDirectories(filenames): expanded = set() for filename in filenames: if (not os.path.isdir(filename)): expanded.add(filename) continue for (root, _, files) in os.walk(filename): for loopfile in files: fullname = os.path.join(root, loopfile) if fullname.startswith(('.' + os.path.sep)): fullname = fullname[len(('.' + os.path.sep)):] expanded.add(fullname) filtered = [] for filename in expanded: if (os.path.splitext(filename)[1][1:] in GetAllExtensions()): filtered.append(filename) return filtered
[ "def", "_ExpandDirectories", "(", "filenames", ")", ":", "expanded", "=", "set", "(", ")", "for", "filename", "in", "filenames", ":", "if", "(", "not", "os", ".", "path", ".", "isdir", "(", "filename", ")", ")", ":", "expanded", ".", "add", "(", "filename", ")", "continue", "for", "(", "root", ",", "_", ",", "files", ")", "in", "os", ".", "walk", "(", "filename", ")", ":", "for", "loopfile", "in", "files", ":", "fullname", "=", "os", ".", "path", ".", "join", "(", "root", ",", "loopfile", ")", "if", "fullname", ".", "startswith", "(", "(", "'.'", "+", "os", ".", "path", ".", "sep", ")", ")", ":", "fullname", "=", "fullname", "[", "len", "(", "(", "'.'", "+", "os", ".", "path", ".", "sep", ")", ")", ":", "]", "expanded", ".", "add", "(", "fullname", ")", "filtered", "=", "[", "]", "for", "filename", "in", "expanded", ":", "if", "(", "os", ".", "path", ".", "splitext", "(", "filename", ")", "[", "1", "]", "[", "1", ":", "]", "in", "GetAllExtensions", "(", ")", ")", ":", "filtered", ".", "append", "(", "filename", ")", "return", "filtered" ]
searches a list of filenames and replaces directories in the list with all files descending from those directories .
train
true
18,480
def strip_whitespace(tokens): for (i, token) in enumerate(tokens): if (token.type != u'S'): break else: return [] tokens = tokens[i:] while (tokens and (tokens[(-1)].type == u'S')): tokens.pop() return tokens
[ "def", "strip_whitespace", "(", "tokens", ")", ":", "for", "(", "i", ",", "token", ")", "in", "enumerate", "(", "tokens", ")", ":", "if", "(", "token", ".", "type", "!=", "u'S'", ")", ":", "break", "else", ":", "return", "[", "]", "tokens", "=", "tokens", "[", "i", ":", "]", "while", "(", "tokens", "and", "(", "tokens", "[", "(", "-", "1", ")", "]", ".", "type", "==", "u'S'", ")", ")", ":", "tokens", ".", "pop", "(", ")", "return", "tokens" ]
remove whitespace from a text or byte string .
train
false
18,481
def convert_missing_indexer(indexer): if isinstance(indexer, dict): indexer = indexer['key'] if isinstance(indexer, bool): raise KeyError('cannot use a single bool to index into setitem') return (indexer, True) return (indexer, False)
[ "def", "convert_missing_indexer", "(", "indexer", ")", ":", "if", "isinstance", "(", "indexer", ",", "dict", ")", ":", "indexer", "=", "indexer", "[", "'key'", "]", "if", "isinstance", "(", "indexer", ",", "bool", ")", ":", "raise", "KeyError", "(", "'cannot use a single bool to index into setitem'", ")", "return", "(", "indexer", ",", "True", ")", "return", "(", "indexer", ",", "False", ")" ]
reverse convert a missing indexer .
train
true