id_within_dataset
int64
1
55.5k
snippet
stringlengths
19
14.2k
tokens
listlengths
6
1.63k
nl
stringlengths
6
352
split_within_dataset
stringclasses
1 value
is_duplicated
bool
2 classes
50,895
def floor(x): return Floor()(x)
[ "def", "floor", "(", "x", ")", ":", "return", "Floor", "(", ")", "(", "x", ")" ]
elementwise floor function .
train
false
50,898
def seed_permissions_roles(course_key): administrator_role = _save_forum_role(course_key, FORUM_ROLE_ADMINISTRATOR) moderator_role = _save_forum_role(course_key, FORUM_ROLE_MODERATOR) community_ta_role = _save_forum_role(course_key, FORUM_ROLE_COMMUNITY_TA) student_role = _save_forum_role(course_key, FORUM_ROLE_STUDENT) for per in STUDENT_ROLE_PERMISSIONS: student_role.add_permission(per) for per in MODERATOR_ROLE_PERMISSIONS: moderator_role.add_permission(per) for per in ADMINISTRATOR_ROLE_PERMISSIONS: administrator_role.add_permission(per) moderator_role.inherit_permissions(student_role) community_ta_role.inherit_permissions(moderator_role) administrator_role.inherit_permissions(moderator_role)
[ "def", "seed_permissions_roles", "(", "course_key", ")", ":", "administrator_role", "=", "_save_forum_role", "(", "course_key", ",", "FORUM_ROLE_ADMINISTRATOR", ")", "moderator_role", "=", "_save_forum_role", "(", "course_key", ",", "FORUM_ROLE_MODERATOR", ")", "community_ta_role", "=", "_save_forum_role", "(", "course_key", ",", "FORUM_ROLE_COMMUNITY_TA", ")", "student_role", "=", "_save_forum_role", "(", "course_key", ",", "FORUM_ROLE_STUDENT", ")", "for", "per", "in", "STUDENT_ROLE_PERMISSIONS", ":", "student_role", ".", "add_permission", "(", "per", ")", "for", "per", "in", "MODERATOR_ROLE_PERMISSIONS", ":", "moderator_role", ".", "add_permission", "(", "per", ")", "for", "per", "in", "ADMINISTRATOR_ROLE_PERMISSIONS", ":", "administrator_role", ".", "add_permission", "(", "per", ")", "moderator_role", ".", "inherit_permissions", "(", "student_role", ")", "community_ta_role", ".", "inherit_permissions", "(", "moderator_role", ")", "administrator_role", ".", "inherit_permissions", "(", "moderator_role", ")" ]
create and assign permissions for forum roles .
train
false
50,899
@app.teardown_appcontext def close_database(exception): top = _app_ctx_stack.top if hasattr(top, 'sqlite_db'): top.sqlite_db.close()
[ "@", "app", ".", "teardown_appcontext", "def", "close_database", "(", "exception", ")", ":", "top", "=", "_app_ctx_stack", ".", "top", "if", "hasattr", "(", "top", ",", "'sqlite_db'", ")", ":", "top", ".", "sqlite_db", ".", "close", "(", ")" ]
closes the database again at the end of the request .
train
false
50,900
@requires_application() def test_arrow_attributes(): with TestingCanvas() as c: arrow = visuals.Arrow(pos=vertices, arrow_type='stealth', arrows=arrows, arrow_size=10, color='red', connect='segments', parent=c.scene) def size_test(): arrow.arrow_size = 0.0 def type_test(): arrow.arrow_type = 'random_non_existent' assert_raises(ValueError, size_test) assert_raises(ValueError, type_test)
[ "@", "requires_application", "(", ")", "def", "test_arrow_attributes", "(", ")", ":", "with", "TestingCanvas", "(", ")", "as", "c", ":", "arrow", "=", "visuals", ".", "Arrow", "(", "pos", "=", "vertices", ",", "arrow_type", "=", "'stealth'", ",", "arrows", "=", "arrows", ",", "arrow_size", "=", "10", ",", "color", "=", "'red'", ",", "connect", "=", "'segments'", ",", "parent", "=", "c", ".", "scene", ")", "def", "size_test", "(", ")", ":", "arrow", ".", "arrow_size", "=", "0.0", "def", "type_test", "(", ")", ":", "arrow", ".", "arrow_type", "=", "'random_non_existent'", "assert_raises", "(", "ValueError", ",", "size_test", ")", "assert_raises", "(", "ValueError", ",", "type_test", ")" ]
tests if the arrowvisual performs the required checks for the attributes .
train
false
50,901
def bugs_to_json_response(data, bunch_of_bugs, callback_function_name=''): obj_serializer = serializers.get_serializer('python')() bugs = obj_serializer.serialize(bunch_of_bugs) for bug in bugs: project = Project.objects.get(pk=int(bug['fields']['project'])) bug['fields']['project'] = project.display_name data_list = [{'bugs': bugs}] json_as_string = json.dumps(data_list, default=encode_datetime) json_string_with_callback = (((callback_function_name + '(') + json_as_string) + ')') return HttpResponse(json_string_with_callback)
[ "def", "bugs_to_json_response", "(", "data", ",", "bunch_of_bugs", ",", "callback_function_name", "=", "''", ")", ":", "obj_serializer", "=", "serializers", ".", "get_serializer", "(", "'python'", ")", "(", ")", "bugs", "=", "obj_serializer", ".", "serialize", "(", "bunch_of_bugs", ")", "for", "bug", "in", "bugs", ":", "project", "=", "Project", ".", "objects", ".", "get", "(", "pk", "=", "int", "(", "bug", "[", "'fields'", "]", "[", "'project'", "]", ")", ")", "bug", "[", "'fields'", "]", "[", "'project'", "]", "=", "project", ".", "display_name", "data_list", "=", "[", "{", "'bugs'", ":", "bugs", "}", "]", "json_as_string", "=", "json", ".", "dumps", "(", "data_list", ",", "default", "=", "encode_datetime", ")", "json_string_with_callback", "=", "(", "(", "(", "callback_function_name", "+", "'('", ")", "+", "json_as_string", ")", "+", "')'", ")", "return", "HttpResponse", "(", "json_string_with_callback", ")" ]
the search results page accesses this view via jquerys getjson method .
train
false
50,902
def notTorNZBFile(filename): return (not (filename.endswith(u'.torrent') or filename.endswith(u'.nzb')))
[ "def", "notTorNZBFile", "(", "filename", ")", ":", "return", "(", "not", "(", "filename", ".", "endswith", "(", "u'.torrent'", ")", "or", "filename", ".", "endswith", "(", "u'.nzb'", ")", ")", ")" ]
returns true if filename is not a nzb nor torrent file .
train
false
50,903
def Hypergeometric(name, N, m, n): return rv(name, HypergeometricDistribution, N, m, n)
[ "def", "Hypergeometric", "(", "name", ",", "N", ",", "m", ",", "n", ")", ":", "return", "rv", "(", "name", ",", "HypergeometricDistribution", ",", "N", ",", "m", ",", "n", ")" ]
create a finite random variable representing a hypergeometric distribution .
train
false
50,904
def enable_share(cookie, tokens, fid_list): url = ''.join([const.PAN_URL, 'share/set?channel=chunlei&clienttype=0&web=1', '&bdstoken=', tokens['bdstoken']]) data = encoder.encode_uri('fid_list={0}&schannel=0&channel_list=[]'.format(fid_list)) req = net.urlopen(url, headers={'Cookie': cookie.header_output(), 'Content-type': const.CONTENT_FORM_UTF8}, data=data.encode()) if req: content = req.data return json.loads(content.decode()) else: return None
[ "def", "enable_share", "(", "cookie", ",", "tokens", ",", "fid_list", ")", ":", "url", "=", "''", ".", "join", "(", "[", "const", ".", "PAN_URL", ",", "'share/set?channel=chunlei&clienttype=0&web=1'", ",", "'&bdstoken='", ",", "tokens", "[", "'bdstoken'", "]", "]", ")", "data", "=", "encoder", ".", "encode_uri", "(", "'fid_list={0}&schannel=0&channel_list=[]'", ".", "format", "(", "fid_list", ")", ")", "req", "=", "net", ".", "urlopen", "(", "url", ",", "headers", "=", "{", "'Cookie'", ":", "cookie", ".", "header_output", "(", ")", ",", "'Content-type'", ":", "const", ".", "CONTENT_FORM_UTF8", "}", ",", "data", "=", "data", ".", "encode", "(", ")", ")", "if", "req", ":", "content", "=", "req", ".", "data", "return", "json", ".", "loads", "(", "content", ".", "decode", "(", ")", ")", "else", ":", "return", "None" ]
fid_list - 是一个list .
train
true
50,906
def cache_feed(): resourcename = 'cache' output = s3_rest_controller(module, resourcename) return output
[ "def", "cache_feed", "(", ")", ":", "resourcename", "=", "'cache'", "output", "=", "s3_rest_controller", "(", "module", ",", "resourcename", ")", "return", "output" ]
restful crud controller - cache georss/kml feeds & make them available to the map viewing client as geojson the create .
train
false
50,907
def set_game_score(token, user_id, score, force=None, disable_edit_message=None, chat_id=None, message_id=None, inline_message_id=None): method_url = 'setGameScore' payload = {'user_id': user_id, 'score': score} if force: payload['force'] = force if chat_id: payload['chat_id'] = chat_id if message_id: payload['message_id'] = message_id if inline_message_id: payload['inline_message_id'] = inline_message_id if disable_edit_message: payload['disable_edit_message'] = disable_edit_message return _make_request(token, method_url, params=payload)
[ "def", "set_game_score", "(", "token", ",", "user_id", ",", "score", ",", "force", "=", "None", ",", "disable_edit_message", "=", "None", ",", "chat_id", "=", "None", ",", "message_id", "=", "None", ",", "inline_message_id", "=", "None", ")", ":", "method_url", "=", "'setGameScore'", "payload", "=", "{", "'user_id'", ":", "user_id", ",", "'score'", ":", "score", "}", "if", "force", ":", "payload", "[", "'force'", "]", "=", "force", "if", "chat_id", ":", "payload", "[", "'chat_id'", "]", "=", "chat_id", "if", "message_id", ":", "payload", "[", "'message_id'", "]", "=", "message_id", "if", "inline_message_id", ":", "payload", "[", "'inline_message_id'", "]", "=", "inline_message_id", "if", "disable_edit_message", ":", "payload", "[", "'disable_edit_message'", "]", "=", "disable_edit_message", "return", "_make_request", "(", "token", ",", "method_url", ",", "params", "=", "payload", ")" ]
use this method to set the score of the specified user in a game .
train
true
50,908
def __pack_message(operation, data): request_id = _randint() message = struct.pack('<i', (16 + len(data))) message += struct.pack('<i', request_id) message += _ZERO_32 message += struct.pack('<i', operation) return (request_id, (message + data))
[ "def", "__pack_message", "(", "operation", ",", "data", ")", ":", "request_id", "=", "_randint", "(", ")", "message", "=", "struct", ".", "pack", "(", "'<i'", ",", "(", "16", "+", "len", "(", "data", ")", ")", ")", "message", "+=", "struct", ".", "pack", "(", "'<i'", ",", "request_id", ")", "message", "+=", "_ZERO_32", "message", "+=", "struct", ".", "pack", "(", "'<i'", ",", "operation", ")", "return", "(", "request_id", ",", "(", "message", "+", "data", ")", ")" ]
takes message data and adds a message header based on the operation .
train
true
50,909
def strip_prefix(device_name): device_name = strip_dev(device_name) return (_pref.sub('', device_name) if device_name else device_name)
[ "def", "strip_prefix", "(", "device_name", ")", ":", "device_name", "=", "strip_dev", "(", "device_name", ")", "return", "(", "_pref", ".", "sub", "(", "''", ",", "device_name", ")", "if", "device_name", "else", "device_name", ")" ]
remove both leading /dev/ and xvd or sd or vd or hd .
train
false
50,910
@register('http') def _check_http(brain, match_kind, match, target_dict, cred_dict): url = ('http:' + (match % target_dict)) data = {'target': jsonutils.dumps(target_dict), 'credentials': jsonutils.dumps(cred_dict)} post_data = urllib.urlencode(data) f = urllib2.urlopen(url, post_data) return (f.read() == 'True')
[ "@", "register", "(", "'http'", ")", "def", "_check_http", "(", "brain", ",", "match_kind", ",", "match", ",", "target_dict", ",", "cred_dict", ")", ":", "url", "=", "(", "'http:'", "+", "(", "match", "%", "target_dict", ")", ")", "data", "=", "{", "'target'", ":", "jsonutils", ".", "dumps", "(", "target_dict", ")", ",", "'credentials'", ":", "jsonutils", ".", "dumps", "(", "cred_dict", ")", "}", "post_data", "=", "urllib", ".", "urlencode", "(", "data", ")", "f", "=", "urllib2", ".", "urlopen", "(", "url", ",", "post_data", ")", "return", "(", "f", ".", "read", "(", ")", "==", "'True'", ")" ]
check http: rules by calling to a remote server .
train
false
50,911
@with_setup(prepare_stdout) def test_output_with_success_colorless2(): runner = Runner(join(abspath(dirname(__file__)), 'output_features', 'runner_features'), verbosity=3, no_color=True) runner.run() assert_stdout_lines('\nFeature: Dumb feature # tests/functional/output_features/runner_features/first.feature:1\n In order to test success # tests/functional/output_features/runner_features/first.feature:2\n As a programmer # tests/functional/output_features/runner_features/first.feature:3\n I want to see that the output is green # tests/functional/output_features/runner_features/first.feature:4\n\n Scenario: Do nothing # tests/functional/output_features/runner_features/first.feature:6\n Given I do nothing # tests/functional/output_features/runner_features/dumb_steps.py:6\n\n1 feature (1 passed)\n1 scenario (1 passed)\n1 step (1 passed)\n')
[ "@", "with_setup", "(", "prepare_stdout", ")", "def", "test_output_with_success_colorless2", "(", ")", ":", "runner", "=", "Runner", "(", "join", "(", "abspath", "(", "dirname", "(", "__file__", ")", ")", ",", "'output_features'", ",", "'runner_features'", ")", ",", "verbosity", "=", "3", ",", "no_color", "=", "True", ")", "runner", ".", "run", "(", ")", "assert_stdout_lines", "(", "'\\nFeature: Dumb feature # tests/functional/output_features/runner_features/first.feature:1\\n In order to test success # tests/functional/output_features/runner_features/first.feature:2\\n As a programmer # tests/functional/output_features/runner_features/first.feature:3\\n I want to see that the output is green # tests/functional/output_features/runner_features/first.feature:4\\n\\n Scenario: Do nothing # tests/functional/output_features/runner_features/first.feature:6\\n Given I do nothing # tests/functional/output_features/runner_features/dumb_steps.py:6\\n\\n1 feature (1 passed)\\n1 scenario (1 passed)\\n1 step (1 passed)\\n'", ")" ]
testing the colorless output of a successful feature .
train
false
50,912
def get_opts(): return __opts__
[ "def", "get_opts", "(", ")", ":", "return", "__opts__" ]
return the configuration options passed to this minion cli example: .
train
false
50,915
def isfinal(elt): return (type(elt) in [str, int, float, unicode, datetime.datetime, REGEXP_T])
[ "def", "isfinal", "(", "elt", ")", ":", "return", "(", "type", "(", "elt", ")", "in", "[", "str", ",", "int", ",", "float", ",", "unicode", ",", "datetime", ".", "datetime", ",", "REGEXP_T", "]", ")" ]
decides whether or not elt is a final element .
train
false
50,916
def _get_links(sr_id, sort, time): q = Link._query((Link.c.sr_id == sr_id), sort=db_sort(sort), data=True) if (time != 'all'): q._filter(db_times[time]) res = make_results(q) return res
[ "def", "_get_links", "(", "sr_id", ",", "sort", ",", "time", ")", ":", "q", "=", "Link", ".", "_query", "(", "(", "Link", ".", "c", ".", "sr_id", "==", "sr_id", ")", ",", "sort", "=", "db_sort", "(", "sort", ")", ",", "data", "=", "True", ")", "if", "(", "time", "!=", "'all'", ")", ":", "q", ".", "_filter", "(", "db_times", "[", "time", "]", ")", "res", "=", "make_results", "(", "q", ")", "return", "res" ]
general link query for a subreddit .
train
false
50,917
def get_nav_history(code, start=None, end=None, retry_count=3, pause=0.001, timeout=10): start = (du.today_last_year() if (start is None) else start) end = (du.today() if (end is None) else end) ismonetary = False df_fund = get_fund_info(code) fund_type = df_fund.ix[0]['Type2Name'] if ((fund_type.find(u'\u503a\u5238\u578b') != (-1)) or (fund_type.find(u'\u8d27\u5e01\u578b') != (-1))): ismonetary = True ct._write_head() nums = _get_nav_histroy_num(code, start, end, ismonetary) data = _parse_nav_history_data(code, start, end, nums, ismonetary, retry_count, pause, timeout) return data
[ "def", "get_nav_history", "(", "code", ",", "start", "=", "None", ",", "end", "=", "None", ",", "retry_count", "=", "3", ",", "pause", "=", "0.001", ",", "timeout", "=", "10", ")", ":", "start", "=", "(", "du", ".", "today_last_year", "(", ")", "if", "(", "start", "is", "None", ")", "else", "start", ")", "end", "=", "(", "du", ".", "today", "(", ")", "if", "(", "end", "is", "None", ")", "else", "end", ")", "ismonetary", "=", "False", "df_fund", "=", "get_fund_info", "(", "code", ")", "fund_type", "=", "df_fund", ".", "ix", "[", "0", "]", "[", "'Type2Name'", "]", "if", "(", "(", "fund_type", ".", "find", "(", "u'\\u503a\\u5238\\u578b'", ")", "!=", "(", "-", "1", ")", ")", "or", "(", "fund_type", ".", "find", "(", "u'\\u8d27\\u5e01\\u578b'", ")", "!=", "(", "-", "1", ")", ")", ")", ":", "ismonetary", "=", "True", "ct", ".", "_write_head", "(", ")", "nums", "=", "_get_nav_histroy_num", "(", "code", ",", "start", ",", "end", ",", "ismonetary", ")", "data", "=", "_parse_nav_history_data", "(", "code", ",", "start", ",", "end", ",", "nums", ",", "ismonetary", ",", "retry_count", ",", "pause", ",", "timeout", ")", "return", "data" ]
parameters code:string 基金代码 e .
train
false
50,918
def requestPdpContextActivation(AccessPointName_presence=0): a = TpPd(pd=8) b = MessageType(mesType=68) c = PacketDataProtocolAddress() packet = ((a / b) / c) if (AccessPointName_presence is 1): d = AccessPointName(ieiAPN=40) packet = (packet / d) return packet
[ "def", "requestPdpContextActivation", "(", "AccessPointName_presence", "=", "0", ")", ":", "a", "=", "TpPd", "(", "pd", "=", "8", ")", "b", "=", "MessageType", "(", "mesType", "=", "68", ")", "c", "=", "PacketDataProtocolAddress", "(", ")", "packet", "=", "(", "(", "a", "/", "b", ")", "/", "c", ")", "if", "(", "AccessPointName_presence", "is", "1", ")", ":", "d", "=", "AccessPointName", "(", "ieiAPN", "=", "40", ")", "packet", "=", "(", "packet", "/", "d", ")", "return", "packet" ]
request pdp context activation section 9 .
train
true
50,919
def _commoncrypto_pbkdf2(data, salt, iterations, digest, keylen): c_hashfunc = ctypes.c_uint32(_commoncrypto_hashlib_to_crypto_map_get(digest)) c_pass = ctypes.c_char_p(data) c_passlen = ctypes.c_size_t(len(data)) c_salt = ctypes.c_char_p(salt) c_saltlen = ctypes.c_size_t(len(salt)) c_iter = ctypes.c_uint(iterations) c_keylen = ctypes.c_size_t(keylen) c_buff = ctypes.create_string_buffer(keylen) crypto.CCKeyDerivationPBKDF.restype = ctypes.c_int crypto.CCKeyDerivationPBKDF.argtypes = [ctypes.c_uint32, ctypes.c_char_p, ctypes.c_size_t, ctypes.c_char_p, ctypes.c_size_t, ctypes.c_uint32, ctypes.c_uint, ctypes.c_char_p, ctypes.c_size_t] ret = crypto.CCKeyDerivationPBKDF(2, c_pass, c_passlen, c_salt, c_saltlen, c_hashfunc, c_iter, c_buff, c_keylen) return ((1 - ret), c_buff)
[ "def", "_commoncrypto_pbkdf2", "(", "data", ",", "salt", ",", "iterations", ",", "digest", ",", "keylen", ")", ":", "c_hashfunc", "=", "ctypes", ".", "c_uint32", "(", "_commoncrypto_hashlib_to_crypto_map_get", "(", "digest", ")", ")", "c_pass", "=", "ctypes", ".", "c_char_p", "(", "data", ")", "c_passlen", "=", "ctypes", ".", "c_size_t", "(", "len", "(", "data", ")", ")", "c_salt", "=", "ctypes", ".", "c_char_p", "(", "salt", ")", "c_saltlen", "=", "ctypes", ".", "c_size_t", "(", "len", "(", "salt", ")", ")", "c_iter", "=", "ctypes", ".", "c_uint", "(", "iterations", ")", "c_keylen", "=", "ctypes", ".", "c_size_t", "(", "keylen", ")", "c_buff", "=", "ctypes", ".", "create_string_buffer", "(", "keylen", ")", "crypto", ".", "CCKeyDerivationPBKDF", ".", "restype", "=", "ctypes", ".", "c_int", "crypto", ".", "CCKeyDerivationPBKDF", ".", "argtypes", "=", "[", "ctypes", ".", "c_uint32", ",", "ctypes", ".", "c_char_p", ",", "ctypes", ".", "c_size_t", ",", "ctypes", ".", "c_char_p", ",", "ctypes", ".", "c_size_t", ",", "ctypes", ".", "c_uint32", ",", "ctypes", ".", "c_uint", ",", "ctypes", ".", "c_char_p", ",", "ctypes", ".", "c_size_t", "]", "ret", "=", "crypto", ".", "CCKeyDerivationPBKDF", "(", "2", ",", "c_pass", ",", "c_passlen", ",", "c_salt", ",", "c_saltlen", ",", "c_hashfunc", ",", "c_iter", ",", "c_buff", ",", "c_keylen", ")", "return", "(", "(", "1", "-", "ret", ")", ",", "c_buff", ")" ]
common crypto compatibile wrapper .
train
false
50,920
@step(u'a file named "{filename}" exists') def step_file_named_filename_exists(context, filename): step_file_named_filename_should_exist(context, filename)
[ "@", "step", "(", "u'a file named \"{filename}\" exists'", ")", "def", "step_file_named_filename_exists", "(", "context", ",", "filename", ")", ":", "step_file_named_filename_should_exist", "(", "context", ",", "filename", ")" ]
verifies that a file with this filename exists .
train
false
50,922
def app_factory(global_conf, **kwargs): kwargs = load_app_properties(kwds=kwargs, config_prefix='TOOL_SHED_CONFIG_') if ('app' in kwargs): app = kwargs.pop('app') else: try: from galaxy.webapps.tool_shed.app import UniverseApplication app = UniverseApplication(global_conf=global_conf, **kwargs) except: import traceback import sys traceback.print_exc() sys.exit(1) atexit.register(app.shutdown) webapp = CommunityWebApplication(app, session_cookie='galaxycommunitysession', name='tool_shed') add_ui_controllers(webapp, app) webapp.add_route('/view/{owner}', controller='repository', action='sharable_owner') webapp.add_route('/view/{owner}/{name}', controller='repository', action='sharable_repository') webapp.add_route('/view/{owner}/{name}/{changeset_revision}', controller='repository', action='sharable_repository_revision') webapp.add_route('/repository/static/images/{repository_id}/{image_file:.+?}', controller='repository', action='display_image_in_repository', repository_id=None, image_file=None) webapp.add_route('/{controller}/{action}', action='index') webapp.add_route('/{action}', controller='repository', action='index') webapp.add_route('/repos/*path_info', controller='hg', action='handle_request', path_info='/') webapp.add_api_controllers('galaxy.webapps.tool_shed.api', app) webapp.mapper.connect('api_key_retrieval', '/api/authenticate/baseauth/', controller='authenticate', action='get_tool_shed_api_key', conditions=dict(method=['GET'])) webapp.mapper.connect('group', '/api/groups/', controller='groups', action='index', conditions=dict(method=['GET'])) webapp.mapper.connect('group', '/api/groups/', controller='groups', action='create', conditions=dict(method=['POST'])) webapp.mapper.connect('group', '/api/groups/{encoded_id}', controller='groups', action='show', conditions=dict(method=['GET'])) webapp.mapper.resource('category', 'categories', controller='categories', name_prefix='category_', path_prefix='/api', parent_resources=dict(member_name='category', collection_name='categories')) webapp.mapper.connect('repositories_in_category', '/api/categories/{category_id}/repositories', controller='categories', action='get_repositories', conditions=dict(method=['GET'])) webapp.mapper.connect('show_updates_for_repository', '/api/repositories/updates', controller='repositories', action='updates', conditions=dict(method=['GET'])) webapp.mapper.resource('repository', 'repositories', controller='repositories', collection={'add_repository_registry_entry': 'POST', 'get_repository_revision_install_info': 'GET', 'get_ordered_installable_revisions': 'GET', 'get_installable_revisions': 'GET', 'remove_repository_registry_entry': 'POST', 'repository_ids_for_setting_metadata': 'GET', 'reset_metadata_on_repositories': 'POST', 'reset_metadata_on_repository': 'POST'}, name_prefix='repository_', path_prefix='/api', new={'import_capsule': 'POST'}, parent_resources=dict(member_name='repository', collection_name='repositories')) webapp.mapper.resource('repository_revision', 'repository_revisions', member={'repository_dependencies': 'GET', 'export': 'POST'}, controller='repository_revisions', name_prefix='repository_revision_', path_prefix='/api', parent_resources=dict(member_name='repository_revision', collection_name='repository_revisions')) webapp.mapper.resource('user', 'users', controller='users', name_prefix='user_', path_prefix='/api', parent_resources=dict(member_name='user', collection_name='users')) webapp.mapper.connect('update_repository', '/api/repositories/{id}', controller='repositories', action='update', conditions=dict(method=['PATCH', 'PUT'])) webapp.mapper.connect('repository_create_changeset_revision', '/api/repositories/{id}/changeset_revision', controller='repositories', action='create_changeset_revision', conditions=dict(method=['POST'])) webapp.mapper.connect('repository_get_metadata', '/api/repositories/{id}/metadata', controller='repositories', action='metadata', conditions=dict(method=['GET'])) webapp.mapper.connect('repository_show_tools', '/api/repositories/{id}/{changeset}/show_tools', controller='repositories', action='show_tools', conditions=dict(method=['GET'])) webapp.mapper.connect('create_repository', '/api/repositories', controller='repositories', action='create', conditions=dict(method=['POST'])) webapp.mapper.connect('tools', '/api/tools', controller='tools', action='index', conditions=dict(method=['GET'])) webapp.mapper.connect('version', '/api/version', controller='configuration', action='version', conditions=dict(method=['GET'])) webapp.finalize_config() if kwargs.get('middleware', True): webapp = wrap_in_middleware(webapp, global_conf, **kwargs) if asbool(kwargs.get('static_enabled', True)): if process_is_uwsgi: log.error('Static middleware is enabled in your configuration but this is a uwsgi process. Refusing to wrap in static middleware.') else: webapp = wrap_in_static(webapp, global_conf, **kwargs) try: galaxy.webapps.tool_shed.model.mapping.metadata.bind.dispose() except: log.exception('Unable to dispose of pooled tool_shed model database connections.') return webapp
[ "def", "app_factory", "(", "global_conf", ",", "**", "kwargs", ")", ":", "kwargs", "=", "load_app_properties", "(", "kwds", "=", "kwargs", ",", "config_prefix", "=", "'TOOL_SHED_CONFIG_'", ")", "if", "(", "'app'", "in", "kwargs", ")", ":", "app", "=", "kwargs", ".", "pop", "(", "'app'", ")", "else", ":", "try", ":", "from", "galaxy", ".", "webapps", ".", "tool_shed", ".", "app", "import", "UniverseApplication", "app", "=", "UniverseApplication", "(", "global_conf", "=", "global_conf", ",", "**", "kwargs", ")", "except", ":", "import", "traceback", "import", "sys", "traceback", ".", "print_exc", "(", ")", "sys", ".", "exit", "(", "1", ")", "atexit", ".", "register", "(", "app", ".", "shutdown", ")", "webapp", "=", "CommunityWebApplication", "(", "app", ",", "session_cookie", "=", "'galaxycommunitysession'", ",", "name", "=", "'tool_shed'", ")", "add_ui_controllers", "(", "webapp", ",", "app", ")", "webapp", ".", "add_route", "(", "'/view/{owner}'", ",", "controller", "=", "'repository'", ",", "action", "=", "'sharable_owner'", ")", "webapp", ".", "add_route", "(", "'/view/{owner}/{name}'", ",", "controller", "=", "'repository'", ",", "action", "=", "'sharable_repository'", ")", "webapp", ".", "add_route", "(", "'/view/{owner}/{name}/{changeset_revision}'", ",", "controller", "=", "'repository'", ",", "action", "=", "'sharable_repository_revision'", ")", "webapp", ".", "add_route", "(", "'/repository/static/images/{repository_id}/{image_file:.+?}'", ",", "controller", "=", "'repository'", ",", "action", "=", "'display_image_in_repository'", ",", "repository_id", "=", "None", ",", "image_file", "=", "None", ")", "webapp", ".", "add_route", "(", "'/{controller}/{action}'", ",", "action", "=", "'index'", ")", "webapp", ".", "add_route", "(", "'/{action}'", ",", "controller", "=", "'repository'", ",", "action", "=", "'index'", ")", "webapp", ".", "add_route", "(", "'/repos/*path_info'", ",", "controller", "=", "'hg'", ",", "action", "=", "'handle_request'", ",", "path_info", "=", "'/'", ")", "webapp", ".", "add_api_controllers", "(", "'galaxy.webapps.tool_shed.api'", ",", "app", ")", "webapp", ".", "mapper", ".", "connect", "(", "'api_key_retrieval'", ",", "'/api/authenticate/baseauth/'", ",", "controller", "=", "'authenticate'", ",", "action", "=", "'get_tool_shed_api_key'", ",", "conditions", "=", "dict", "(", "method", "=", "[", "'GET'", "]", ")", ")", "webapp", ".", "mapper", ".", "connect", "(", "'group'", ",", "'/api/groups/'", ",", "controller", "=", "'groups'", ",", "action", "=", "'index'", ",", "conditions", "=", "dict", "(", "method", "=", "[", "'GET'", "]", ")", ")", "webapp", ".", "mapper", ".", "connect", "(", "'group'", ",", "'/api/groups/'", ",", "controller", "=", "'groups'", ",", "action", "=", "'create'", ",", "conditions", "=", "dict", "(", "method", "=", "[", "'POST'", "]", ")", ")", "webapp", ".", "mapper", ".", "connect", "(", "'group'", ",", "'/api/groups/{encoded_id}'", ",", "controller", "=", "'groups'", ",", "action", "=", "'show'", ",", "conditions", "=", "dict", "(", "method", "=", "[", "'GET'", "]", ")", ")", "webapp", ".", "mapper", ".", "resource", "(", "'category'", ",", "'categories'", ",", "controller", "=", "'categories'", ",", "name_prefix", "=", "'category_'", ",", "path_prefix", "=", "'/api'", ",", "parent_resources", "=", "dict", "(", "member_name", "=", "'category'", ",", "collection_name", "=", "'categories'", ")", ")", "webapp", ".", "mapper", ".", "connect", "(", "'repositories_in_category'", ",", "'/api/categories/{category_id}/repositories'", ",", "controller", "=", "'categories'", ",", "action", "=", "'get_repositories'", ",", "conditions", "=", "dict", "(", "method", "=", "[", "'GET'", "]", ")", ")", "webapp", ".", "mapper", ".", "connect", "(", "'show_updates_for_repository'", ",", "'/api/repositories/updates'", ",", "controller", "=", "'repositories'", ",", "action", "=", "'updates'", ",", "conditions", "=", "dict", "(", "method", "=", "[", "'GET'", "]", ")", ")", "webapp", ".", "mapper", ".", "resource", "(", "'repository'", ",", "'repositories'", ",", "controller", "=", "'repositories'", ",", "collection", "=", "{", "'add_repository_registry_entry'", ":", "'POST'", ",", "'get_repository_revision_install_info'", ":", "'GET'", ",", "'get_ordered_installable_revisions'", ":", "'GET'", ",", "'get_installable_revisions'", ":", "'GET'", ",", "'remove_repository_registry_entry'", ":", "'POST'", ",", "'repository_ids_for_setting_metadata'", ":", "'GET'", ",", "'reset_metadata_on_repositories'", ":", "'POST'", ",", "'reset_metadata_on_repository'", ":", "'POST'", "}", ",", "name_prefix", "=", "'repository_'", ",", "path_prefix", "=", "'/api'", ",", "new", "=", "{", "'import_capsule'", ":", "'POST'", "}", ",", "parent_resources", "=", "dict", "(", "member_name", "=", "'repository'", ",", "collection_name", "=", "'repositories'", ")", ")", "webapp", ".", "mapper", ".", "resource", "(", "'repository_revision'", ",", "'repository_revisions'", ",", "member", "=", "{", "'repository_dependencies'", ":", "'GET'", ",", "'export'", ":", "'POST'", "}", ",", "controller", "=", "'repository_revisions'", ",", "name_prefix", "=", "'repository_revision_'", ",", "path_prefix", "=", "'/api'", ",", "parent_resources", "=", "dict", "(", "member_name", "=", "'repository_revision'", ",", "collection_name", "=", "'repository_revisions'", ")", ")", "webapp", ".", "mapper", ".", "resource", "(", "'user'", ",", "'users'", ",", "controller", "=", "'users'", ",", "name_prefix", "=", "'user_'", ",", "path_prefix", "=", "'/api'", ",", "parent_resources", "=", "dict", "(", "member_name", "=", "'user'", ",", "collection_name", "=", "'users'", ")", ")", "webapp", ".", "mapper", ".", "connect", "(", "'update_repository'", ",", "'/api/repositories/{id}'", ",", "controller", "=", "'repositories'", ",", "action", "=", "'update'", ",", "conditions", "=", "dict", "(", "method", "=", "[", "'PATCH'", ",", "'PUT'", "]", ")", ")", "webapp", ".", "mapper", ".", "connect", "(", "'repository_create_changeset_revision'", ",", "'/api/repositories/{id}/changeset_revision'", ",", "controller", "=", "'repositories'", ",", "action", "=", "'create_changeset_revision'", ",", "conditions", "=", "dict", "(", "method", "=", "[", "'POST'", "]", ")", ")", "webapp", ".", "mapper", ".", "connect", "(", "'repository_get_metadata'", ",", "'/api/repositories/{id}/metadata'", ",", "controller", "=", "'repositories'", ",", "action", "=", "'metadata'", ",", "conditions", "=", "dict", "(", "method", "=", "[", "'GET'", "]", ")", ")", "webapp", ".", "mapper", ".", "connect", "(", "'repository_show_tools'", ",", "'/api/repositories/{id}/{changeset}/show_tools'", ",", "controller", "=", "'repositories'", ",", "action", "=", "'show_tools'", ",", "conditions", "=", "dict", "(", "method", "=", "[", "'GET'", "]", ")", ")", "webapp", ".", "mapper", ".", "connect", "(", "'create_repository'", ",", "'/api/repositories'", ",", "controller", "=", "'repositories'", ",", "action", "=", "'create'", ",", "conditions", "=", "dict", "(", "method", "=", "[", "'POST'", "]", ")", ")", "webapp", ".", "mapper", ".", "connect", "(", "'tools'", ",", "'/api/tools'", ",", "controller", "=", "'tools'", ",", "action", "=", "'index'", ",", "conditions", "=", "dict", "(", "method", "=", "[", "'GET'", "]", ")", ")", "webapp", ".", "mapper", ".", "connect", "(", "'version'", ",", "'/api/version'", ",", "controller", "=", "'configuration'", ",", "action", "=", "'version'", ",", "conditions", "=", "dict", "(", "method", "=", "[", "'GET'", "]", ")", ")", "webapp", ".", "finalize_config", "(", ")", "if", "kwargs", ".", "get", "(", "'middleware'", ",", "True", ")", ":", "webapp", "=", "wrap_in_middleware", "(", "webapp", ",", "global_conf", ",", "**", "kwargs", ")", "if", "asbool", "(", "kwargs", ".", "get", "(", "'static_enabled'", ",", "True", ")", ")", ":", "if", "process_is_uwsgi", ":", "log", ".", "error", "(", "'Static middleware is enabled in your configuration but this is a uwsgi process. Refusing to wrap in static middleware.'", ")", "else", ":", "webapp", "=", "wrap_in_static", "(", "webapp", ",", "global_conf", ",", "**", "kwargs", ")", "try", ":", "galaxy", ".", "webapps", ".", "tool_shed", ".", "model", ".", "mapping", ".", "metadata", ".", "bind", ".", "dispose", "(", ")", "except", ":", "log", ".", "exception", "(", "'Unable to dispose of pooled tool_shed model database connections.'", ")", "return", "webapp" ]
return a wsgi application serving the root object .
train
false
50,923
def scale_timedelta(context, builder, val, srcty, destty): factor = npdatetime.get_timedelta_conversion_factor(srcty.unit, destty.unit) if (factor is None): raise NotImplementedError(('cannot convert timedelta64 from %r to %r' % (srcty.unit, destty.unit))) return scale_by_constant(builder, val, factor)
[ "def", "scale_timedelta", "(", "context", ",", "builder", ",", "val", ",", "srcty", ",", "destty", ")", ":", "factor", "=", "npdatetime", ".", "get_timedelta_conversion_factor", "(", "srcty", ".", "unit", ",", "destty", ".", "unit", ")", "if", "(", "factor", "is", "None", ")", ":", "raise", "NotImplementedError", "(", "(", "'cannot convert timedelta64 from %r to %r'", "%", "(", "srcty", ".", "unit", ",", "destty", ".", "unit", ")", ")", ")", "return", "scale_by_constant", "(", "builder", ",", "val", ",", "factor", ")" ]
scale the timedelta64 *val* from *srcty* to *destty* .
train
false
50,925
def encode_hex_to_base32(hex_string): bin_form = binascii.unhexlify(hex_string) return base64.b32encode(bin_form)
[ "def", "encode_hex_to_base32", "(", "hex_string", ")", ":", "bin_form", "=", "binascii", ".", "unhexlify", "(", "hex_string", ")", "return", "base64", ".", "b32encode", "(", "bin_form", ")" ]
encodes hex to base32 bit as per rfc4648 .
train
false
50,926
def getStepKeyFromPoint(point): return (int(round(point.real)), int(round(point.imag)))
[ "def", "getStepKeyFromPoint", "(", "point", ")", ":", "return", "(", "int", "(", "round", "(", "point", ".", "real", ")", ")", ",", "int", "(", "round", "(", "point", ".", "imag", ")", ")", ")" ]
get step key for the point .
train
false
50,927
def test_issue_3825(): x = Symbol('x') y = Symbol('y') a1 = (x + y) a2 = (y + x) a2.is_comparable h1 = hash(a1) h2 = hash(a2) assert (h1 == h2)
[ "def", "test_issue_3825", "(", ")", ":", "x", "=", "Symbol", "(", "'x'", ")", "y", "=", "Symbol", "(", "'y'", ")", "a1", "=", "(", "x", "+", "y", ")", "a2", "=", "(", "y", "+", "x", ")", "a2", ".", "is_comparable", "h1", "=", "hash", "(", "a1", ")", "h2", "=", "hash", "(", "a2", ")", "assert", "(", "h1", "==", "h2", ")" ]
catch: hash instability .
train
false
50,928
def send_login(): form_class = _security.passwordless_login_form if request.json: form = form_class(MultiDict(request.json)) else: form = form_class() if form.validate_on_submit(): send_login_instructions(form.user) if (request.json is None): do_flash(*get_message('LOGIN_EMAIL_SENT', email=form.user.email)) if request.json: return _render_json(form) return _security.render_template(config_value('SEND_LOGIN_TEMPLATE'), send_login_form=form, **_ctx('send_login'))
[ "def", "send_login", "(", ")", ":", "form_class", "=", "_security", ".", "passwordless_login_form", "if", "request", ".", "json", ":", "form", "=", "form_class", "(", "MultiDict", "(", "request", ".", "json", ")", ")", "else", ":", "form", "=", "form_class", "(", ")", "if", "form", ".", "validate_on_submit", "(", ")", ":", "send_login_instructions", "(", "form", ".", "user", ")", "if", "(", "request", ".", "json", "is", "None", ")", ":", "do_flash", "(", "*", "get_message", "(", "'LOGIN_EMAIL_SENT'", ",", "email", "=", "form", ".", "user", ".", "email", ")", ")", "if", "request", ".", "json", ":", "return", "_render_json", "(", "form", ")", "return", "_security", ".", "render_template", "(", "config_value", "(", "'SEND_LOGIN_TEMPLATE'", ")", ",", "send_login_form", "=", "form", ",", "**", "_ctx", "(", "'send_login'", ")", ")" ]
view function that sends login instructions for passwordless login .
train
true
50,929
def type_or_constraint_repr(constraint): if isinstance(constraint, type): return constraint.__name__ elif isinstance(constraint, Exactly): return repr(constraint)
[ "def", "type_or_constraint_repr", "(", "constraint", ")", ":", "if", "isinstance", "(", "constraint", ",", "type", ")", ":", "return", "constraint", ".", "__name__", "elif", "isinstance", "(", "constraint", ",", "Exactly", ")", ":", "return", "repr", "(", "constraint", ")" ]
generate correct repr for types and typeconstraints .
train
false
50,931
def _expand_cookie_path(protocolinfo_response, pid_resolver, pid_resolution_arg): cookie_path = protocolinfo_response.cookie_path if (cookie_path and (not os.path.isabs(cookie_path))): try: tor_pid = pid_resolver(pid_resolution_arg) if (not tor_pid): raise IOError('pid lookup failed') tor_cwd = stem.util.system.cwd(tor_pid) if (not tor_cwd): raise IOError('cwd lookup failed') cookie_path = stem.util.system.expand_path(cookie_path, tor_cwd) except IOError as exc: resolver_labels = {stem.util.system.pid_by_name: ' by name', stem.util.system.pid_by_port: ' by port', stem.util.system.pid_by_open_file: ' by socket file'} pid_resolver_label = resolver_labels.get(pid_resolver, '') log.debug(('unable to expand relative tor cookie path%s: %s' % (pid_resolver_label, exc))) protocolinfo_response.cookie_path = cookie_path
[ "def", "_expand_cookie_path", "(", "protocolinfo_response", ",", "pid_resolver", ",", "pid_resolution_arg", ")", ":", "cookie_path", "=", "protocolinfo_response", ".", "cookie_path", "if", "(", "cookie_path", "and", "(", "not", "os", ".", "path", ".", "isabs", "(", "cookie_path", ")", ")", ")", ":", "try", ":", "tor_pid", "=", "pid_resolver", "(", "pid_resolution_arg", ")", "if", "(", "not", "tor_pid", ")", ":", "raise", "IOError", "(", "'pid lookup failed'", ")", "tor_cwd", "=", "stem", ".", "util", ".", "system", ".", "cwd", "(", "tor_pid", ")", "if", "(", "not", "tor_cwd", ")", ":", "raise", "IOError", "(", "'cwd lookup failed'", ")", "cookie_path", "=", "stem", ".", "util", ".", "system", ".", "expand_path", "(", "cookie_path", ",", "tor_cwd", ")", "except", "IOError", "as", "exc", ":", "resolver_labels", "=", "{", "stem", ".", "util", ".", "system", ".", "pid_by_name", ":", "' by name'", ",", "stem", ".", "util", ".", "system", ".", "pid_by_port", ":", "' by port'", ",", "stem", ".", "util", ".", "system", ".", "pid_by_open_file", ":", "' by socket file'", "}", "pid_resolver_label", "=", "resolver_labels", ".", "get", "(", "pid_resolver", ",", "''", ")", "log", ".", "debug", "(", "(", "'unable to expand relative tor cookie path%s: %s'", "%", "(", "pid_resolver_label", ",", "exc", ")", ")", ")", "protocolinfo_response", ".", "cookie_path", "=", "cookie_path" ]
attempts to expand a relative cookie path with the given pid resolver .
train
false
50,932
def set2rel(s): new = set() for elem in s: if isinstance(elem, string_types): new.add((elem,)) elif isinstance(elem, int): new.add(str(elem)) else: new.add(elem) return new
[ "def", "set2rel", "(", "s", ")", ":", "new", "=", "set", "(", ")", "for", "elem", "in", "s", ":", "if", "isinstance", "(", "elem", ",", "string_types", ")", ":", "new", ".", "add", "(", "(", "elem", ",", ")", ")", "elif", "isinstance", "(", "elem", ",", "int", ")", ":", "new", ".", "add", "(", "str", "(", "elem", ")", ")", "else", ":", "new", ".", "add", "(", "elem", ")", "return", "new" ]
convert a set containing individuals into a set of unary tuples .
train
false
50,933
def test_gloo_without_app(): class DummyParser(gloo.glir.BaseGlirParser, ): def __init__(self): self.commands = [] def parse(self, commands): self.commands.extend(commands) p = DummyParser() c = gloo.context.FakeCanvas() c.context.shared.parser = p gloo.clear() c.flush() gloo.clear() c.flush() assert (len(p.commands) in (2, 3)) assert (p.commands[(-1)][1] == 'glClear')
[ "def", "test_gloo_without_app", "(", ")", ":", "class", "DummyParser", "(", "gloo", ".", "glir", ".", "BaseGlirParser", ",", ")", ":", "def", "__init__", "(", "self", ")", ":", "self", ".", "commands", "=", "[", "]", "def", "parse", "(", "self", ",", "commands", ")", ":", "self", ".", "commands", ".", "extend", "(", "commands", ")", "p", "=", "DummyParser", "(", ")", "c", "=", "gloo", ".", "context", ".", "FakeCanvas", "(", ")", "c", ".", "context", ".", "shared", ".", "parser", "=", "p", "gloo", ".", "clear", "(", ")", "c", ".", "flush", "(", ")", "gloo", ".", "clear", "(", ")", "c", ".", "flush", "(", ")", "assert", "(", "len", "(", "p", ".", "commands", ")", "in", "(", "2", ",", "3", ")", ")", "assert", "(", "p", ".", "commands", "[", "(", "-", "1", ")", "]", "[", "1", "]", "==", "'glClear'", ")" ]
test gloo without vispy .
train
false
50,935
def handle_extendsnode(extendsnode, context): if (BLOCK_CONTEXT_KEY not in context.render_context): context.render_context[BLOCK_CONTEXT_KEY] = BlockContext() block_context = context.render_context[BLOCK_CONTEXT_KEY] blocks = dict(((n.name, n) for n in extendsnode.nodelist.get_nodes_by_type(BlockNode))) block_context.add_blocks(blocks) compiled_parent = extendsnode.get_parent(context) parent_nodelist = compiled_parent.nodelist for node in parent_nodelist: if (not isinstance(node, TextNode)): if isinstance(node, ExtendsNode): return handle_extendsnode(node, context) break blocks = dict(((n.name, n) for n in parent_nodelist.get_nodes_by_type(BlockNode))) block_context.add_blocks(blocks) block_stack = [] new_nodelist = remove_block_nodes(parent_nodelist, block_stack, block_context) return new_nodelist
[ "def", "handle_extendsnode", "(", "extendsnode", ",", "context", ")", ":", "if", "(", "BLOCK_CONTEXT_KEY", "not", "in", "context", ".", "render_context", ")", ":", "context", ".", "render_context", "[", "BLOCK_CONTEXT_KEY", "]", "=", "BlockContext", "(", ")", "block_context", "=", "context", ".", "render_context", "[", "BLOCK_CONTEXT_KEY", "]", "blocks", "=", "dict", "(", "(", "(", "n", ".", "name", ",", "n", ")", "for", "n", "in", "extendsnode", ".", "nodelist", ".", "get_nodes_by_type", "(", "BlockNode", ")", ")", ")", "block_context", ".", "add_blocks", "(", "blocks", ")", "compiled_parent", "=", "extendsnode", ".", "get_parent", "(", "context", ")", "parent_nodelist", "=", "compiled_parent", ".", "nodelist", "for", "node", "in", "parent_nodelist", ":", "if", "(", "not", "isinstance", "(", "node", ",", "TextNode", ")", ")", ":", "if", "isinstance", "(", "node", ",", "ExtendsNode", ")", ":", "return", "handle_extendsnode", "(", "node", ",", "context", ")", "break", "blocks", "=", "dict", "(", "(", "(", "n", ".", "name", ",", "n", ")", "for", "n", "in", "parent_nodelist", ".", "get_nodes_by_type", "(", "BlockNode", ")", ")", ")", "block_context", ".", "add_blocks", "(", "blocks", ")", "block_stack", "=", "[", "]", "new_nodelist", "=", "remove_block_nodes", "(", "parent_nodelist", ",", "block_stack", ",", "block_context", ")", "return", "new_nodelist" ]
create a copy of node tree of a derived template replacing all blocks tags with the nodes of appropriate blocks .
train
false
50,936
def backward_eye(n): M = eye(n) for i in range(int((M.rows / 2))): M.row_swap((0 + i), ((M.rows - 1) - i)) return M
[ "def", "backward_eye", "(", "n", ")", ":", "M", "=", "eye", "(", "n", ")", "for", "i", "in", "range", "(", "int", "(", "(", "M", ".", "rows", "/", "2", ")", ")", ")", ":", "M", ".", "row_swap", "(", "(", "0", "+", "i", ")", ",", "(", "(", "M", ".", "rows", "-", "1", ")", "-", "i", ")", ")", "return", "M" ]
returns the backward identity matrix of dimensions n x n .
train
false
50,937
def iterateLineGenerator(proto, gen): coll = _IteratorBuffer(proto.transport.writeSequence, gen) return proto.schedule(coll)
[ "def", "iterateLineGenerator", "(", "proto", ",", "gen", ")", ":", "coll", "=", "_IteratorBuffer", "(", "proto", ".", "transport", ".", "writeSequence", ",", "gen", ")", "return", "proto", ".", "schedule", "(", "coll", ")" ]
hook the given protocol instance up to the given iterator with an _iteratorbuffer and schedule the result to be exhausted via the protocol .
train
false
50,939
def _is_valid_shell(shell): if salt.utils.is_windows(): return True shells = '/etc/shells' available_shells = [] if os.path.exists(shells): try: with salt.utils.fopen(shells, 'r') as shell_fp: lines = shell_fp.read().splitlines() for line in lines: if line.startswith('#'): continue else: available_shells.append(line) except OSError: return True else: return None if (shell in available_shells): return True else: return False
[ "def", "_is_valid_shell", "(", "shell", ")", ":", "if", "salt", ".", "utils", ".", "is_windows", "(", ")", ":", "return", "True", "shells", "=", "'/etc/shells'", "available_shells", "=", "[", "]", "if", "os", ".", "path", ".", "exists", "(", "shells", ")", ":", "try", ":", "with", "salt", ".", "utils", ".", "fopen", "(", "shells", ",", "'r'", ")", "as", "shell_fp", ":", "lines", "=", "shell_fp", ".", "read", "(", ")", ".", "splitlines", "(", ")", "for", "line", "in", "lines", ":", "if", "line", ".", "startswith", "(", "'#'", ")", ":", "continue", "else", ":", "available_shells", ".", "append", "(", "line", ")", "except", "OSError", ":", "return", "True", "else", ":", "return", "None", "if", "(", "shell", "in", "available_shells", ")", ":", "return", "True", "else", ":", "return", "False" ]
attempts to search for valid shells on a system and see if a given shell is in the list .
train
false
50,940
def iterate_attributes(cls): keys = dir(cls) for key in keys: for c in cls.__mro__: if (key in c.__dict__): (yield (key, c.__dict__[key])) break
[ "def", "iterate_attributes", "(", "cls", ")", ":", "keys", "=", "dir", "(", "cls", ")", "for", "key", "in", "keys", ":", "for", "c", "in", "cls", ".", "__mro__", ":", "if", "(", "key", "in", "c", ".", "__dict__", ")", ":", "(", "yield", "(", "key", ",", "c", ".", "__dict__", "[", "key", "]", ")", ")", "break" ]
iterate all the keys and attributes associated with a class .
train
false
50,941
def windowview(folder, view=None): fsr = Carbon.File.FSRef(folder) folder_alias = fsr.FSNewAliasMinimal() if (view is None): return _getwindowview(folder_alias) return _setwindowview(folder_alias, view)
[ "def", "windowview", "(", "folder", ",", "view", "=", "None", ")", ":", "fsr", "=", "Carbon", ".", "File", ".", "FSRef", "(", "folder", ")", "folder_alias", "=", "fsr", ".", "FSNewAliasMinimal", "(", ")", "if", "(", "view", "is", "None", ")", ":", "return", "_getwindowview", "(", "folder_alias", ")", "return", "_setwindowview", "(", "folder_alias", ",", "view", ")" ]
windowview: set the view of the window for the folder .
train
false
50,942
def relative_recursive_glob(dirname, pattern): assert (pattern == '**') if dirname: (yield pattern[:0]) for relative_dir in _iter_relative_dirs(dirname): (yield relative_dir)
[ "def", "relative_recursive_glob", "(", "dirname", ",", "pattern", ")", ":", "assert", "(", "pattern", "==", "'**'", ")", "if", "dirname", ":", "(", "yield", "pattern", "[", ":", "0", "]", ")", "for", "relative_dir", "in", "_iter_relative_dirs", "(", "dirname", ")", ":", "(", "yield", "relative_dir", ")" ]
recursive glob for one directory and all its subdirectories .
train
false
50,943
def suppress(action='ignore', **kwarg): return ((action,), kwarg)
[ "def", "suppress", "(", "action", "=", "'ignore'", ",", "**", "kwarg", ")", ":", "return", "(", "(", "action", ",", ")", ",", "kwarg", ")" ]
sets up the .
train
false
50,944
@skip('netstandard') def test_cp_19510(): import clr clr.AddReference('System.Xml') import System.Xml doc = System.Xml.XmlDocument() doc.LoadXml('<tag attr="value">Data</tag>') root = doc.SelectSingleNode('tag') AreEqual(root.Attributes['attr'].Name, 'attr')
[ "@", "skip", "(", "'netstandard'", ")", "def", "test_cp_19510", "(", ")", ":", "import", "clr", "clr", ".", "AddReference", "(", "'System.Xml'", ")", "import", "System", ".", "Xml", "doc", "=", "System", ".", "Xml", ".", "XmlDocument", "(", ")", "doc", ".", "LoadXml", "(", "'<tag attr=\"value\">Data</tag>'", ")", "root", "=", "doc", ".", "SelectSingleNode", "(", "'tag'", ")", "AreEqual", "(", "root", ".", "Attributes", "[", "'attr'", "]", ".", "Name", ",", "'attr'", ")" ]
test indexing on .
train
false
50,945
def normcase(s): return s
[ "def", "normcase", "(", "s", ")", ":", "return", "s" ]
normalize case of pathname .
train
false
50,946
@decorators.memoize def _check_mdata_put(): return salt.utils.which('mdata-put')
[ "@", "decorators", ".", "memoize", "def", "_check_mdata_put", "(", ")", ":", "return", "salt", ".", "utils", ".", "which", "(", "'mdata-put'", ")" ]
looks to see if mdata-put is present on the system .
train
false
50,947
def HostNameValid(host): valid = False for attr in ['AF_INET', 'AF_INET6']: try: socket.inet_pton(socket.__getattribute__(attr), host) valid = True break except socket.error: pass if (not valid): if ((not host) or ((not DnsNameValid(host)) and (not ALPHA_RE.match(host)))): return False else: return True else: return True
[ "def", "HostNameValid", "(", "host", ")", ":", "valid", "=", "False", "for", "attr", "in", "[", "'AF_INET'", ",", "'AF_INET6'", "]", ":", "try", ":", "socket", ".", "inet_pton", "(", "socket", ".", "__getattribute__", "(", "attr", ")", ",", "host", ")", "valid", "=", "True", "break", "except", "socket", ".", "error", ":", "pass", "if", "(", "not", "valid", ")", ":", "if", "(", "(", "not", "host", ")", "or", "(", "(", "not", "DnsNameValid", "(", "host", ")", ")", "and", "(", "not", "ALPHA_RE", ".", "match", "(", "host", ")", ")", ")", ")", ":", "return", "False", "else", ":", "return", "True", "else", ":", "return", "True" ]
tests whether a string is a valid host-name .
train
false
50,948
def comparePosition(firstElement, secondElement): return cmp(firstElement._markpos, secondElement._markpos)
[ "def", "comparePosition", "(", "firstElement", ",", "secondElement", ")", ":", "return", "cmp", "(", "firstElement", ".", "_markpos", ",", "secondElement", ".", "_markpos", ")" ]
compare the two elements given by their position in the document or documents they were parsed from .
train
false
50,949
def is_ignorable_404(uri): return any((pattern.search(uri) for pattern in getattr(settings, 'IGNORABLE_404_URLS', ())))
[ "def", "is_ignorable_404", "(", "uri", ")", ":", "return", "any", "(", "(", "pattern", ".", "search", "(", "uri", ")", "for", "pattern", "in", "getattr", "(", "settings", ",", "'IGNORABLE_404_URLS'", ",", "(", ")", ")", ")", ")" ]
returns true if a 404 at the given url *shouldnt* notify the site managers .
train
false
50,950
def numericise(value, empty2zero=False, default_blank=''): if (value is not None): try: value = int(value) except ValueError: try: value = float(value) except ValueError: if (value == ''): if empty2zero: value = 0 else: value = default_blank return value
[ "def", "numericise", "(", "value", ",", "empty2zero", "=", "False", ",", "default_blank", "=", "''", ")", ":", "if", "(", "value", "is", "not", "None", ")", ":", "try", ":", "value", "=", "int", "(", "value", ")", "except", "ValueError", ":", "try", ":", "value", "=", "float", "(", "value", ")", "except", "ValueError", ":", "if", "(", "value", "==", "''", ")", ":", "if", "empty2zero", ":", "value", "=", "0", "else", ":", "value", "=", "default_blank", "return", "value" ]
returns a value that depends on the input string: - float if input can be converted to float - integer if input can be converted to integer - zero if the input string is empty and empty2zero flag is set - the same input string .
train
true
50,951
def readBody(response): def cancel(deferred): '\n Cancel a L{readBody} call, close the connection to the HTTP server\n immediately, if it is still open.\n\n @param deferred: The cancelled L{defer.Deferred}.\n ' abort = getAbort() if (abort is not None): abort() d = defer.Deferred(cancel) protocol = _ReadBodyProtocol(response.code, response.phrase, d) def getAbort(): return getattr(protocol.transport, 'abortConnection', None) response.deliverBody(protocol) if ((protocol.transport is not None) and (getAbort() is None)): warnings.warn('Using readBody with a transport that does not have an abortConnection method', category=DeprecationWarning, stacklevel=2) return d
[ "def", "readBody", "(", "response", ")", ":", "def", "cancel", "(", "deferred", ")", ":", "abort", "=", "getAbort", "(", ")", "if", "(", "abort", "is", "not", "None", ")", ":", "abort", "(", ")", "d", "=", "defer", ".", "Deferred", "(", "cancel", ")", "protocol", "=", "_ReadBodyProtocol", "(", "response", ".", "code", ",", "response", ".", "phrase", ",", "d", ")", "def", "getAbort", "(", ")", ":", "return", "getattr", "(", "protocol", ".", "transport", ",", "'abortConnection'", ",", "None", ")", "response", ".", "deliverBody", "(", "protocol", ")", "if", "(", "(", "protocol", ".", "transport", "is", "not", "None", ")", "and", "(", "getAbort", "(", ")", "is", "None", ")", ")", ":", "warnings", ".", "warn", "(", "'Using readBody with a transport that does not have an abortConnection method'", ",", "category", "=", "DeprecationWarning", ",", "stacklevel", "=", "2", ")", "return", "d" ]
get the body of an l{iresponse} and return it as a byte string .
train
false
50,954
def _CheckFieldName(name): _ValidateString(name, 'name', MAXIMUM_FIELD_NAME_LENGTH) if (not re.match(_FIELD_NAME_PATTERN, name)): raise ValueError(('field name "%s" should match pattern: %s' % (name, _FIELD_NAME_PATTERN))) return name
[ "def", "_CheckFieldName", "(", "name", ")", ":", "_ValidateString", "(", "name", ",", "'name'", ",", "MAXIMUM_FIELD_NAME_LENGTH", ")", "if", "(", "not", "re", ".", "match", "(", "_FIELD_NAME_PATTERN", ",", "name", ")", ")", ":", "raise", "ValueError", "(", "(", "'field name \"%s\" should match pattern: %s'", "%", "(", "name", ",", "_FIELD_NAME_PATTERN", ")", ")", ")", "return", "name" ]
checks field name is not too long and matches field name pattern .
train
false
50,957
def tstd(a, limits=None, inclusive=(True, True), axis=0, ddof=1): return np.sqrt(tvar(a, limits, inclusive, axis, ddof))
[ "def", "tstd", "(", "a", ",", "limits", "=", "None", ",", "inclusive", "=", "(", "True", ",", "True", ")", ",", "axis", "=", "0", ",", "ddof", "=", "1", ")", ":", "return", "np", ".", "sqrt", "(", "tvar", "(", "a", ",", "limits", ",", "inclusive", ",", "axis", ",", "ddof", ")", ")" ]
compute the trimmed sample standard deviation this function finds the sample standard deviation of given values .
train
false
50,959
def _encode_sk_name_entry(backend, attributes): stack = backend._lib.sk_X509_NAME_ENTRY_new_null() for attribute in attributes: name_entry = _encode_name_entry(backend, attribute) res = backend._lib.sk_X509_NAME_ENTRY_push(stack, name_entry) backend.openssl_assert((res == 1)) return stack
[ "def", "_encode_sk_name_entry", "(", "backend", ",", "attributes", ")", ":", "stack", "=", "backend", ".", "_lib", ".", "sk_X509_NAME_ENTRY_new_null", "(", ")", "for", "attribute", "in", "attributes", ":", "name_entry", "=", "_encode_name_entry", "(", "backend", ",", "attribute", ")", "res", "=", "backend", ".", "_lib", ".", "sk_X509_NAME_ENTRY_push", "(", "stack", ",", "name_entry", ")", "backend", ".", "openssl_assert", "(", "(", "res", "==", "1", ")", ")", "return", "stack" ]
the sk_x50_name_entry created will not be gcd .
train
false
50,961
def test_ternary(method, prec, exp_range, restricted_range, itr, stat): if (method in TernaryRestricted): exp_range = restricted_range for op in all_ternary(prec, exp_range, itr): t = TestSet(method, op) try: if (not convert(t)): continue callfuncs(t) verify(t, stat) except VerifyError as err: log(err) if (not method.startswith('__')): for op in ternary_optarg(prec, exp_range, itr): t = TestSet(method, op) try: if (not convert(t)): continue callfuncs(t) verify(t, stat) except VerifyError as err: log(err)
[ "def", "test_ternary", "(", "method", ",", "prec", ",", "exp_range", ",", "restricted_range", ",", "itr", ",", "stat", ")", ":", "if", "(", "method", "in", "TernaryRestricted", ")", ":", "exp_range", "=", "restricted_range", "for", "op", "in", "all_ternary", "(", "prec", ",", "exp_range", ",", "itr", ")", ":", "t", "=", "TestSet", "(", "method", ",", "op", ")", "try", ":", "if", "(", "not", "convert", "(", "t", ")", ")", ":", "continue", "callfuncs", "(", "t", ")", "verify", "(", "t", ",", "stat", ")", "except", "VerifyError", "as", "err", ":", "log", "(", "err", ")", "if", "(", "not", "method", ".", "startswith", "(", "'__'", ")", ")", ":", "for", "op", "in", "ternary_optarg", "(", "prec", ",", "exp_range", ",", "itr", ")", ":", "t", "=", "TestSet", "(", "method", ",", "op", ")", "try", ":", "if", "(", "not", "convert", "(", "t", ")", ")", ":", "continue", "callfuncs", "(", "t", ")", "verify", "(", "t", ",", "stat", ")", "except", "VerifyError", "as", "err", ":", "log", "(", "err", ")" ]
iterate a ternary function through many test cases .
train
false
50,962
@click.command('nginx') @click.option('--yes', help='Yes to regeneration of nginx config file', default=False, is_flag=True) def setup_nginx(yes=False): from bench.config.nginx import make_nginx_conf make_nginx_conf(bench_path='.', yes=yes)
[ "@", "click", ".", "command", "(", "'nginx'", ")", "@", "click", ".", "option", "(", "'--yes'", ",", "help", "=", "'Yes to regeneration of nginx config file'", ",", "default", "=", "False", ",", "is_flag", "=", "True", ")", "def", "setup_nginx", "(", "yes", "=", "False", ")", ":", "from", "bench", ".", "config", ".", "nginx", "import", "make_nginx_conf", "make_nginx_conf", "(", "bench_path", "=", "'.'", ",", "yes", "=", "yes", ")" ]
generate config for nginx .
train
false
50,963
def getEngineFilename(): base_search_path = os.path.dirname(inspect.getfile(getEngineFilename)) search_filename = 'CuraEngine' if (platform.system() == 'Windows'): search_filename += '.exe' if (version.isDevVersion() and os.path.exists('C:/Software/Cura_SteamEngine/_bin/Release/Cura_SteamEngine.exe')): return 'C:/Software/Cura_SteamEngine/_bin/Release/Cura_SteamEngine.exe' for n in xrange(0, 10): full_filename = os.path.abspath(os.path.join(base_search_path, '/'.join((['..'] * n)), search_filename)) if os.path.isfile(full_filename): return full_filename full_filename = os.path.abspath(os.path.join(base_search_path, '/'.join((['..'] * n)), 'CuraEngine', search_filename)) if os.path.isfile(full_filename): return full_filename if os.path.isfile('/usr/bin/CuraEngine'): return '/usr/bin/CuraEngine' if os.path.isfile('/usr/local/bin/CuraEngine'): return '/usr/local/bin/CuraEngine' return ''
[ "def", "getEngineFilename", "(", ")", ":", "base_search_path", "=", "os", ".", "path", ".", "dirname", "(", "inspect", ".", "getfile", "(", "getEngineFilename", ")", ")", "search_filename", "=", "'CuraEngine'", "if", "(", "platform", ".", "system", "(", ")", "==", "'Windows'", ")", ":", "search_filename", "+=", "'.exe'", "if", "(", "version", ".", "isDevVersion", "(", ")", "and", "os", ".", "path", ".", "exists", "(", "'C:/Software/Cura_SteamEngine/_bin/Release/Cura_SteamEngine.exe'", ")", ")", ":", "return", "'C:/Software/Cura_SteamEngine/_bin/Release/Cura_SteamEngine.exe'", "for", "n", "in", "xrange", "(", "0", ",", "10", ")", ":", "full_filename", "=", "os", ".", "path", ".", "abspath", "(", "os", ".", "path", ".", "join", "(", "base_search_path", ",", "'/'", ".", "join", "(", "(", "[", "'..'", "]", "*", "n", ")", ")", ",", "search_filename", ")", ")", "if", "os", ".", "path", ".", "isfile", "(", "full_filename", ")", ":", "return", "full_filename", "full_filename", "=", "os", ".", "path", ".", "abspath", "(", "os", ".", "path", ".", "join", "(", "base_search_path", ",", "'/'", ".", "join", "(", "(", "[", "'..'", "]", "*", "n", ")", ")", ",", "'CuraEngine'", ",", "search_filename", ")", ")", "if", "os", ".", "path", ".", "isfile", "(", "full_filename", ")", ":", "return", "full_filename", "if", "os", ".", "path", ".", "isfile", "(", "'/usr/bin/CuraEngine'", ")", ":", "return", "'/usr/bin/CuraEngine'", "if", "os", ".", "path", ".", "isfile", "(", "'/usr/local/bin/CuraEngine'", ")", ":", "return", "'/usr/local/bin/CuraEngine'", "return", "''" ]
finds and returns the path to the current engine executable .
train
false
50,965
def get_username(user_id): if (user_id == feconf.MIGRATION_BOT_USER_ID): return feconf.MIGRATION_BOT_USERNAME else: return get_user_settings(user_id, strict=True).username
[ "def", "get_username", "(", "user_id", ")", ":", "if", "(", "user_id", "==", "feconf", ".", "MIGRATION_BOT_USER_ID", ")", ":", "return", "feconf", ".", "MIGRATION_BOT_USERNAME", "else", ":", "return", "get_user_settings", "(", "user_id", ",", "strict", "=", "True", ")", ".", "username" ]
gets username corresponding to the given user_id .
train
false
50,966
def _relevant(option): from certbot import renewal from certbot.plugins import disco as plugins_disco plugins = list(plugins_disco.PluginsRegistry.find_all()) return ((option in renewal.CONFIG_ITEMS) or any((option.startswith((x + '_')) for x in plugins)))
[ "def", "_relevant", "(", "option", ")", ":", "from", "certbot", "import", "renewal", "from", "certbot", ".", "plugins", "import", "disco", "as", "plugins_disco", "plugins", "=", "list", "(", "plugins_disco", ".", "PluginsRegistry", ".", "find_all", "(", ")", ")", "return", "(", "(", "option", "in", "renewal", ".", "CONFIG_ITEMS", ")", "or", "any", "(", "(", "option", ".", "startswith", "(", "(", "x", "+", "'_'", ")", ")", "for", "x", "in", "plugins", ")", ")", ")" ]
is this option one that could be restored for future renewal purposes? .
train
false
50,967
def _ifconfig_getnode(): for args in ('', '-a', '-av'): mac = _find_mac('ifconfig', args, ['hwaddr', 'ether'], (lambda i: (i + 1))) if mac: return mac import socket ip_addr = socket.gethostbyname(socket.gethostname()) mac = _find_mac('arp', '-an', [ip_addr], (lambda i: (-1))) if mac: return mac mac = _find_mac('lanscan', '-ai', ['lan0'], (lambda i: 0)) if mac: return mac return None
[ "def", "_ifconfig_getnode", "(", ")", ":", "for", "args", "in", "(", "''", ",", "'-a'", ",", "'-av'", ")", ":", "mac", "=", "_find_mac", "(", "'ifconfig'", ",", "args", ",", "[", "'hwaddr'", ",", "'ether'", "]", ",", "(", "lambda", "i", ":", "(", "i", "+", "1", ")", ")", ")", "if", "mac", ":", "return", "mac", "import", "socket", "ip_addr", "=", "socket", ".", "gethostbyname", "(", "socket", ".", "gethostname", "(", ")", ")", "mac", "=", "_find_mac", "(", "'arp'", ",", "'-an'", ",", "[", "ip_addr", "]", ",", "(", "lambda", "i", ":", "(", "-", "1", ")", ")", ")", "if", "mac", ":", "return", "mac", "mac", "=", "_find_mac", "(", "'lanscan'", ",", "'-ai'", ",", "[", "'lan0'", "]", ",", "(", "lambda", "i", ":", "0", ")", ")", "if", "mac", ":", "return", "mac", "return", "None" ]
get the hardware address on unix by running ifconfig .
train
true
50,970
def isnull(obj): return _isnull(obj)
[ "def", "isnull", "(", "obj", ")", ":", "return", "_isnull", "(", "obj", ")" ]
detect missing values parameters arr : ndarray or object value object to check for null-ness returns isnulled : array-like of bool or bool array or bool indicating whether an object is null or if an array is given which of the element is null .
train
false
50,972
def is_port_trusted(port): return port['device_owner'].startswith(n_const.DEVICE_OWNER_NETWORK_PREFIX)
[ "def", "is_port_trusted", "(", "port", ")", ":", "return", "port", "[", "'device_owner'", "]", ".", "startswith", "(", "n_const", ".", "DEVICE_OWNER_NETWORK_PREFIX", ")" ]
used to determine if port can be trusted not to attack network .
train
false
50,973
def _guess_type(field): data_types = set([int, float]) if isinstance(field, (dict, list)): return 'nested' if isinstance(field, int): return 'int' if isinstance(field, float): return 'float' for data_type in list(data_types): try: data_type(field) except (TypeError, ValueError): data_types.discard(data_type) if (not data_types): break if (int in data_types): return 'integer' elif (float in data_types): return 'numeric' for format in _DATE_FORMATS: try: datetime.datetime.strptime(field, format) return 'timestamp' except (ValueError, TypeError): continue return 'text'
[ "def", "_guess_type", "(", "field", ")", ":", "data_types", "=", "set", "(", "[", "int", ",", "float", "]", ")", "if", "isinstance", "(", "field", ",", "(", "dict", ",", "list", ")", ")", ":", "return", "'nested'", "if", "isinstance", "(", "field", ",", "int", ")", ":", "return", "'int'", "if", "isinstance", "(", "field", ",", "float", ")", ":", "return", "'float'", "for", "data_type", "in", "list", "(", "data_types", ")", ":", "try", ":", "data_type", "(", "field", ")", "except", "(", "TypeError", ",", "ValueError", ")", ":", "data_types", ".", "discard", "(", "data_type", ")", "if", "(", "not", "data_types", ")", ":", "break", "if", "(", "int", "in", "data_types", ")", ":", "return", "'integer'", "elif", "(", "float", "in", "data_types", ")", ":", "return", "'numeric'", "for", "format", "in", "_DATE_FORMATS", ":", "try", ":", "datetime", ".", "datetime", ".", "strptime", "(", "field", ",", "format", ")", "return", "'timestamp'", "except", "(", "ValueError", ",", "TypeError", ")", ":", "continue", "return", "'text'" ]
simple guess type of field .
train
false
50,974
def post_clear_cache(func): @wraps(func) def post_clear_cache_if_not_raised(self, *args, **kwargs): rval = func(self, *args, **kwargs) self._delete_entries_cache() return rval return post_clear_cache_if_not_raised
[ "def", "post_clear_cache", "(", "func", ")", ":", "@", "wraps", "(", "func", ")", "def", "post_clear_cache_if_not_raised", "(", "self", ",", "*", "args", ",", "**", "kwargs", ")", ":", "rval", "=", "func", "(", "self", ",", "*", "args", ",", "**", "kwargs", ")", "self", ".", "_delete_entries_cache", "(", ")", "return", "rval", "return", "post_clear_cache_if_not_raised" ]
decorator for functions that alter the index using the git command .
train
true
50,976
@register.simple_tag def check_severity(check): try: return escape(CHECKS[check].severity) except KeyError: return u'info'
[ "@", "register", ".", "simple_tag", "def", "check_severity", "(", "check", ")", ":", "try", ":", "return", "escape", "(", "CHECKS", "[", "check", "]", ".", "severity", ")", "except", "KeyError", ":", "return", "u'info'" ]
returns check severity .
train
false
50,977
def _get_s3transfer_performance_script(script_name): s3transfer_directory = os.path.dirname(s3transfer.__file__) s3transfer_directory = os.path.dirname(s3transfer_directory) scripts_directory = 'scripts/performance' scripts_directory = os.path.join(s3transfer_directory, scripts_directory) script = os.path.join(scripts_directory, script_name) if os.path.isfile(script): return script else: return None
[ "def", "_get_s3transfer_performance_script", "(", "script_name", ")", ":", "s3transfer_directory", "=", "os", ".", "path", ".", "dirname", "(", "s3transfer", ".", "__file__", ")", "s3transfer_directory", "=", "os", ".", "path", ".", "dirname", "(", "s3transfer_directory", ")", "scripts_directory", "=", "'scripts/performance'", "scripts_directory", "=", "os", ".", "path", ".", "join", "(", "s3transfer_directory", ",", "scripts_directory", ")", "script", "=", "os", ".", "path", ".", "join", "(", "scripts_directory", ",", "script_name", ")", "if", "os", ".", "path", ".", "isfile", "(", "script", ")", ":", "return", "script", "else", ":", "return", "None" ]
retrieves an s3transfer performance script if available .
train
false
50,978
def bool_int(value): if isinstance(value, basestring): if (value.lower() in ('', '0', 'false', 'f', 'no', 'n', 'off')): value = 0 return int(bool(value))
[ "def", "bool_int", "(", "value", ")", ":", "if", "isinstance", "(", "value", ",", "basestring", ")", ":", "if", "(", "value", ".", "lower", "(", ")", "in", "(", "''", ",", "'0'", ",", "'false'", ",", "'f'", ",", "'no'", ",", "'n'", ",", "'off'", ")", ")", ":", "value", "=", "0", "return", "int", "(", "bool", "(", "value", ")", ")" ]
casts a config value into a 0 or 1 .
train
false
50,981
@pytest.mark.parametrize('specialchars', [' ', ' abcde ', ' ab cd', ' abcde', 'abcde ', ' a b c d e ', ' a b c d e ']) @pytest.mark.django_db def test_clean_specialchars_whitespace(specialchars): form_data = {'code': 'foo', 'fullname': 'Foo', 'checkstyle': 'foo', 'nplurals': '2', 'specialchars': specialchars} form = LanguageForm(form_data) assert form.is_valid() assert (' ' in form.cleaned_data['specialchars'])
[ "@", "pytest", ".", "mark", ".", "parametrize", "(", "'specialchars'", ",", "[", "' '", ",", "' abcde '", ",", "' ab cd'", ",", "' abcde'", ",", "'abcde '", ",", "' a b c d e '", ",", "' a b c d e '", "]", ")", "@", "pytest", ".", "mark", ".", "django_db", "def", "test_clean_specialchars_whitespace", "(", "specialchars", ")", ":", "form_data", "=", "{", "'code'", ":", "'foo'", ",", "'fullname'", ":", "'Foo'", ",", "'checkstyle'", ":", "'foo'", ",", "'nplurals'", ":", "'2'", ",", "'specialchars'", ":", "specialchars", "}", "form", "=", "LanguageForm", "(", "form_data", ")", "assert", "form", ".", "is_valid", "(", ")", "assert", "(", "' '", "in", "form", ".", "cleaned_data", "[", "'specialchars'", "]", ")" ]
tests whitespace is accepted in special characters .
train
false
50,982
def relu__(x): return T.switch((x < 0.0), 0.0, x)
[ "def", "relu__", "(", "x", ")", ":", "return", "T", ".", "switch", "(", "(", "x", "<", "0.0", ")", ",", "0.0", ",", "x", ")" ]
alternative relu implementation .
train
false
50,983
@app.route('/auth/info/googlejwt', methods=['GET']) def auth_info_google_jwt(): return auth_info()
[ "@", "app", ".", "route", "(", "'/auth/info/googlejwt'", ",", "methods", "=", "[", "'GET'", "]", ")", "def", "auth_info_google_jwt", "(", ")", ":", "return", "auth_info", "(", ")" ]
auth info with google signed jwt .
train
false
50,984
def time_until(d, now=None, count=2, accuracy=6, simple=False): if (not now): now = datetime.datetime.now() return time_since(now, d, count, accuracy, simple)
[ "def", "time_until", "(", "d", ",", "now", "=", "None", ",", "count", "=", "2", ",", "accuracy", "=", "6", ",", "simple", "=", "False", ")", ":", "if", "(", "not", "now", ")", ":", "now", "=", "datetime", ".", "datetime", ".", "now", "(", ")", "return", "time_since", "(", "now", ",", "d", ",", "count", ",", "accuracy", ",", "simple", ")" ]
like timesince .
train
false
50,986
def copy_weights(teacher_model, student_model, layer_names): for name in layer_names: weights = teacher_model.get_layer(name=name).get_weights() student_model.get_layer(name=name).set_weights(weights)
[ "def", "copy_weights", "(", "teacher_model", ",", "student_model", ",", "layer_names", ")", ":", "for", "name", "in", "layer_names", ":", "weights", "=", "teacher_model", ".", "get_layer", "(", "name", "=", "name", ")", ".", "get_weights", "(", ")", "student_model", ".", "get_layer", "(", "name", "=", "name", ")", ".", "set_weights", "(", "weights", ")" ]
copy weights from teacher_model to student_model .
train
false
50,987
def logged_run_process(reactor, command): d = Deferred() action = TWISTED_CHILD_PROCESS_ACTION(command=command) with action.context(): d2 = DeferredContext(d) protocol = _LoggingProcessProtocol(d, action) reactor.spawnProcess(protocol, command[0], command) def process_ended((reason, output)): status = reason.value.status if status: raise _CalledProcessError(returncode=status, cmd=command, output=output) return _ProcessResult(command=command, status=status, output=output) d2.addCallback(process_ended) d2.addActionFinish() return d2.result
[ "def", "logged_run_process", "(", "reactor", ",", "command", ")", ":", "d", "=", "Deferred", "(", ")", "action", "=", "TWISTED_CHILD_PROCESS_ACTION", "(", "command", "=", "command", ")", "with", "action", ".", "context", "(", ")", ":", "d2", "=", "DeferredContext", "(", "d", ")", "protocol", "=", "_LoggingProcessProtocol", "(", "d", ",", "action", ")", "reactor", ".", "spawnProcess", "(", "protocol", ",", "command", "[", "0", "]", ",", "command", ")", "def", "process_ended", "(", "(", "reason", ",", "output", ")", ")", ":", "status", "=", "reason", ".", "value", ".", "status", "if", "status", ":", "raise", "_CalledProcessError", "(", "returncode", "=", "status", ",", "cmd", "=", "command", ",", "output", "=", "output", ")", "return", "_ProcessResult", "(", "command", "=", "command", ",", "status", "=", "status", ",", "output", "=", "output", ")", "d2", ".", "addCallback", "(", "process_ended", ")", "d2", ".", "addActionFinish", "(", ")", "return", "d2", ".", "result" ]
run a child process .
train
false
50,988
def _check_pil_jpeg_bytes(): from PIL import Image buf = BytesIO() img = Image.new('RGB', (4, 4)) try: img.save(buf, 'jpeg') except Exception as e: ename = e.__class__.__name__ raise SkipTest(("PIL can't write JPEG to BytesIO: %s: %s" % (ename, e)))
[ "def", "_check_pil_jpeg_bytes", "(", ")", ":", "from", "PIL", "import", "Image", "buf", "=", "BytesIO", "(", ")", "img", "=", "Image", ".", "new", "(", "'RGB'", ",", "(", "4", ",", "4", ")", ")", "try", ":", "img", ".", "save", "(", "buf", ",", "'jpeg'", ")", "except", "Exception", "as", "e", ":", "ename", "=", "e", ".", "__class__", ".", "__name__", "raise", "SkipTest", "(", "(", "\"PIL can't write JPEG to BytesIO: %s: %s\"", "%", "(", "ename", ",", "e", ")", ")", ")" ]
skip if pil cant write jpegs to bytesio objects .
train
false
50,989
def get_sd_configcheck(agentConfig, configs): print '\nSource of the configuration objects built by the agent:\n' for (check_name, config) in configs.iteritems(): print ('Check "%s":\n source --> %s\n config --> %s\n' % (check_name, config[0], config[1])) try: print_containers() except Exception: print 'Failed to collect containers info.' try: print_templates(agentConfig) except Exception: print 'Failed to collect configuration templates.'
[ "def", "get_sd_configcheck", "(", "agentConfig", ",", "configs", ")", ":", "print", "'\\nSource of the configuration objects built by the agent:\\n'", "for", "(", "check_name", ",", "config", ")", "in", "configs", ".", "iteritems", "(", ")", ":", "print", "(", "'Check \"%s\":\\n source --> %s\\n config --> %s\\n'", "%", "(", "check_name", ",", "config", "[", "0", "]", ",", "config", "[", "1", "]", ")", ")", "try", ":", "print_containers", "(", ")", "except", "Exception", ":", "print", "'Failed to collect containers info.'", "try", ":", "print_templates", "(", "agentConfig", ")", "except", "Exception", ":", "print", "'Failed to collect configuration templates.'" ]
trace how the configuration objects are loaded and from where .
train
false
50,990
@login_required def account_redirect(request): return redirect(u'profile_update')
[ "@", "login_required", "def", "account_redirect", "(", "request", ")", ":", "return", "redirect", "(", "u'profile_update'", ")" ]
just gives the url prefix for accounts an action - redirect to the profile update form .
train
false
50,992
def assert_type(name, obj, types, none_ok=True): if (obj is None): if none_ok: return True else: raise AssertionError(('%s may not be None' % name)) if (not isinstance(types, (tuple, list))): types = [types] for cls in types: if isinstance(obj, cls): return True allowed_types = '|'.join(map((lambda x: str(x)), types)) stack = traceback.extract_stack() stack_msg = ('Function call %s() in %s:%d' % (stack[(-2)][2], stack[(-3)][0], stack[(-3)][1])) type_msg = ('%s must be instance of %s (but is %s)' % (name, allowed_types, str(type(obj)))) raise AssertionError(((stack_msg + ': ') + type_msg))
[ "def", "assert_type", "(", "name", ",", "obj", ",", "types", ",", "none_ok", "=", "True", ")", ":", "if", "(", "obj", "is", "None", ")", ":", "if", "none_ok", ":", "return", "True", "else", ":", "raise", "AssertionError", "(", "(", "'%s may not be None'", "%", "name", ")", ")", "if", "(", "not", "isinstance", "(", "types", ",", "(", "tuple", ",", "list", ")", ")", ")", ":", "types", "=", "[", "types", "]", "for", "cls", "in", "types", ":", "if", "isinstance", "(", "obj", ",", "cls", ")", ":", "return", "True", "allowed_types", "=", "'|'", ".", "join", "(", "map", "(", "(", "lambda", "x", ":", "str", "(", "x", ")", ")", ",", "types", ")", ")", "stack", "=", "traceback", ".", "extract_stack", "(", ")", "stack_msg", "=", "(", "'Function call %s() in %s:%d'", "%", "(", "stack", "[", "(", "-", "2", ")", "]", "[", "2", "]", ",", "stack", "[", "(", "-", "3", ")", "]", "[", "0", "]", ",", "stack", "[", "(", "-", "3", ")", "]", "[", "1", "]", ")", ")", "type_msg", "=", "(", "'%s must be instance of %s (but is %s)'", "%", "(", "name", ",", "allowed_types", ",", "str", "(", "type", "(", "obj", ")", ")", ")", ")", "raise", "AssertionError", "(", "(", "(", "stack_msg", "+", "': '", ")", "+", "type_msg", ")", ")" ]
assert that a parameter is of a given type .
train
false
50,994
def find_initrd(path): if (path is None): return None if os.path.isfile(path): return path elif os.path.isdir(path): return find_highest_files(path, 'initrd.img', _re_initrd) elif (file_is_remote(path) and remote_file_exists(path)): return path return None
[ "def", "find_initrd", "(", "path", ")", ":", "if", "(", "path", "is", "None", ")", ":", "return", "None", "if", "os", ".", "path", ".", "isfile", "(", "path", ")", ":", "return", "path", "elif", "os", ".", "path", ".", "isdir", "(", "path", ")", ":", "return", "find_highest_files", "(", "path", ",", "'initrd.img'", ",", "_re_initrd", ")", "elif", "(", "file_is_remote", "(", "path", ")", "and", "remote_file_exists", "(", "path", ")", ")", ":", "return", "path", "return", "None" ]
given a directory or a filename .
train
false
50,995
def _maybe_upcast(values, fill_value=np.nan, dtype=None, copy=False): if is_extension_type(values): if copy: values = values.copy() else: if (dtype is None): dtype = values.dtype (new_dtype, fill_value) = _maybe_promote(dtype, fill_value) if (new_dtype != values.dtype): values = values.astype(new_dtype) elif copy: values = values.copy() return (values, fill_value)
[ "def", "_maybe_upcast", "(", "values", ",", "fill_value", "=", "np", ".", "nan", ",", "dtype", "=", "None", ",", "copy", "=", "False", ")", ":", "if", "is_extension_type", "(", "values", ")", ":", "if", "copy", ":", "values", "=", "values", ".", "copy", "(", ")", "else", ":", "if", "(", "dtype", "is", "None", ")", ":", "dtype", "=", "values", ".", "dtype", "(", "new_dtype", ",", "fill_value", ")", "=", "_maybe_promote", "(", "dtype", ",", "fill_value", ")", "if", "(", "new_dtype", "!=", "values", ".", "dtype", ")", ":", "values", "=", "values", ".", "astype", "(", "new_dtype", ")", "elif", "copy", ":", "values", "=", "values", ".", "copy", "(", ")", "return", "(", "values", ",", "fill_value", ")" ]
provide explict type promotion and coercion parameters values : the ndarray that we want to maybe upcast fill_value : what we want to fill with dtype : if none .
train
false
50,996
def read(results_file): results = {} if (not os.path.exists(results_file)): raise IOError('Results file does not exist.') with open(results_file) as handle: lines = handle.readlines() (results, num_params) = _parse_baseml.parse_basics(lines, results) results = _parse_baseml.parse_parameters(lines, results, num_params) if (results.get('version') is None): raise ValueError('Invalid results file') return results
[ "def", "read", "(", "results_file", ")", ":", "results", "=", "{", "}", "if", "(", "not", "os", ".", "path", ".", "exists", "(", "results_file", ")", ")", ":", "raise", "IOError", "(", "'Results file does not exist.'", ")", "with", "open", "(", "results_file", ")", "as", "handle", ":", "lines", "=", "handle", ".", "readlines", "(", ")", "(", "results", ",", "num_params", ")", "=", "_parse_baseml", ".", "parse_basics", "(", "lines", ",", "results", ")", "results", "=", "_parse_baseml", ".", "parse_parameters", "(", "lines", ",", "results", ",", "num_params", ")", "if", "(", "results", ".", "get", "(", "'version'", ")", "is", "None", ")", ":", "raise", "ValueError", "(", "'Invalid results file'", ")", "return", "results" ]
parses an xml file from the ncbi entrez utilities into python objects .
train
false
50,997
def all_subindices(index): return (index[start:stop] for (start, stop) in product_upper_triangle(range((len(index) + 1))))
[ "def", "all_subindices", "(", "index", ")", ":", "return", "(", "index", "[", "start", ":", "stop", "]", "for", "(", "start", ",", "stop", ")", "in", "product_upper_triangle", "(", "range", "(", "(", "len", "(", "index", ")", "+", "1", ")", ")", ")", ")" ]
return all valid sub-indices of a pandas index .
train
false
51,001
@contextlib.contextmanager def with_comprehensive_theme_context(theme=None): if theme: domain = '{theme}.org'.format(theme=re.sub('\\.org$', '', theme)) (site, __) = Site.objects.get_or_create(domain=domain, name=theme) (site_theme, __) = SiteTheme.objects.get_or_create(site=site, theme_dir_name=theme) with patch('openedx.core.djangoapps.theming.helpers.get_current_site_theme', return_value=site_theme): with patch('openedx.core.djangoapps.theming.helpers.get_current_site', return_value=site): (yield) else: (yield)
[ "@", "contextlib", ".", "contextmanager", "def", "with_comprehensive_theme_context", "(", "theme", "=", "None", ")", ":", "if", "theme", ":", "domain", "=", "'{theme}.org'", ".", "format", "(", "theme", "=", "re", ".", "sub", "(", "'\\\\.org$'", ",", "''", ",", "theme", ")", ")", "(", "site", ",", "__", ")", "=", "Site", ".", "objects", ".", "get_or_create", "(", "domain", "=", "domain", ",", "name", "=", "theme", ")", "(", "site_theme", ",", "__", ")", "=", "SiteTheme", ".", "objects", ".", "get_or_create", "(", "site", "=", "site", ",", "theme_dir_name", "=", "theme", ")", "with", "patch", "(", "'openedx.core.djangoapps.theming.helpers.get_current_site_theme'", ",", "return_value", "=", "site_theme", ")", ":", "with", "patch", "(", "'openedx.core.djangoapps.theming.helpers.get_current_site'", ",", "return_value", "=", "site", ")", ":", "(", "yield", ")", "else", ":", "(", "yield", ")" ]
a function to run a test as if request was made to the given theme .
train
false
51,002
def generate_tool_guid(repository_clone_url, tool): tmp_url = common_util.remove_protocol_and_user_from_clone_url(repository_clone_url) return ('%s/%s/%s' % (tmp_url, tool.id, tool.version))
[ "def", "generate_tool_guid", "(", "repository_clone_url", ",", "tool", ")", ":", "tmp_url", "=", "common_util", ".", "remove_protocol_and_user_from_clone_url", "(", "repository_clone_url", ")", "return", "(", "'%s/%s/%s'", "%", "(", "tmp_url", ",", "tool", ".", "id", ",", "tool", ".", "version", ")", ")" ]
generate a guid for the installed tool .
train
false
51,003
def vectorize_if_needed(func, *x): if any(map(isiterable, x)): return np.vectorize(func)(*x) else: return func(*x)
[ "def", "vectorize_if_needed", "(", "func", ",", "*", "x", ")", ":", "if", "any", "(", "map", "(", "isiterable", ",", "x", ")", ")", ":", "return", "np", ".", "vectorize", "(", "func", ")", "(", "*", "x", ")", "else", ":", "return", "func", "(", "*", "x", ")" ]
helper function to vectorize functions on array inputs .
train
false
51,007
def create_hash_map(): hashmap = {} from base64 import encodestring as base64 import pwd login_name = pwd.getpwuid(os.geteuid())[0] conn = http.client.HTTPSConnection(u'api.github.com') conn.request(u'GET', u'/repos/nipy/nipype', headers={u'Authorization': (u'Basic %s' % base64(login_name))}) try: conn.request(u'GET', u'/repos/nipy/nipype/git/trees/master?recursive=1') except: pass else: r1 = conn.getresponse() if (r1.reason != u'OK'): raise Exception((u'HTTP Response %s:%s' % (r1.status, r1.reason))) payload = simplejson.loads(r1.read()) for infodict in payload[u'tree']: if (infodict[u'type'] == u'blob'): hashmap[infodict[u'sha']] = infodict[u'path'] return hashmap
[ "def", "create_hash_map", "(", ")", ":", "hashmap", "=", "{", "}", "from", "base64", "import", "encodestring", "as", "base64", "import", "pwd", "login_name", "=", "pwd", ".", "getpwuid", "(", "os", ".", "geteuid", "(", ")", ")", "[", "0", "]", "conn", "=", "http", ".", "client", ".", "HTTPSConnection", "(", "u'api.github.com'", ")", "conn", ".", "request", "(", "u'GET'", ",", "u'/repos/nipy/nipype'", ",", "headers", "=", "{", "u'Authorization'", ":", "(", "u'Basic %s'", "%", "base64", "(", "login_name", ")", ")", "}", ")", "try", ":", "conn", ".", "request", "(", "u'GET'", ",", "u'/repos/nipy/nipype/git/trees/master?recursive=1'", ")", "except", ":", "pass", "else", ":", "r1", "=", "conn", ".", "getresponse", "(", ")", "if", "(", "r1", ".", "reason", "!=", "u'OK'", ")", ":", "raise", "Exception", "(", "(", "u'HTTP Response %s:%s'", "%", "(", "r1", ".", "status", ",", "r1", ".", "reason", ")", ")", ")", "payload", "=", "simplejson", ".", "loads", "(", "r1", ".", "read", "(", ")", ")", "for", "infodict", "in", "payload", "[", "u'tree'", "]", ":", "if", "(", "infodict", "[", "u'type'", "]", "==", "u'blob'", ")", ":", "hashmap", "[", "infodict", "[", "u'sha'", "]", "]", "=", "infodict", "[", "u'path'", "]", "return", "hashmap" ]
create a hash map for all objects .
train
false
51,008
def getVector3ByMultiplierPrefix(elementNode, multiplier, prefix, vector3): if (multiplier == 0.0): return vector3 oldMultipliedValueVector3 = (vector3 * multiplier) vector3ByPrefix = getVector3ByPrefix(oldMultipliedValueVector3.copy(), elementNode, prefix) if (vector3ByPrefix == oldMultipliedValueVector3): return vector3 return (vector3ByPrefix / multiplier)
[ "def", "getVector3ByMultiplierPrefix", "(", "elementNode", ",", "multiplier", ",", "prefix", ",", "vector3", ")", ":", "if", "(", "multiplier", "==", "0.0", ")", ":", "return", "vector3", "oldMultipliedValueVector3", "=", "(", "vector3", "*", "multiplier", ")", "vector3ByPrefix", "=", "getVector3ByPrefix", "(", "oldMultipliedValueVector3", ".", "copy", "(", ")", ",", "elementNode", ",", "prefix", ")", "if", "(", "vector3ByPrefix", "==", "oldMultipliedValueVector3", ")", ":", "return", "vector3", "return", "(", "vector3ByPrefix", "/", "multiplier", ")" ]
get vector3 from multiplier .
train
false
51,009
@box(types.RawPointer) def box_raw_pointer(typ, val, c): ll_intp = c.context.get_value_type(types.uintp) addr = c.builder.ptrtoint(val, ll_intp) return c.box(types.uintp, addr)
[ "@", "box", "(", "types", ".", "RawPointer", ")", "def", "box_raw_pointer", "(", "typ", ",", "val", ",", "c", ")", ":", "ll_intp", "=", "c", ".", "context", ".", "get_value_type", "(", "types", ".", "uintp", ")", "addr", "=", "c", ".", "builder", ".", "ptrtoint", "(", "val", ",", "ll_intp", ")", "return", "c", ".", "box", "(", "types", ".", "uintp", ",", "addr", ")" ]
convert a raw pointer to a python int .
train
false
51,010
def add_imap_status_info_rows(folder_id, account_id, db_session): if (not db_session.query(ImapFolderSyncStatus).filter_by(account_id=account_id, folder_id=folder_id).all()): db_session.add(ImapFolderSyncStatus(account_id=account_id, folder_id=folder_id, state='initial')) if (not db_session.query(ImapFolderInfo).filter_by(account_id=account_id, folder_id=folder_id).all()): db_session.add(ImapFolderInfo(account_id=account_id, folder_id=folder_id, uidvalidity=1, highestmodseq=22))
[ "def", "add_imap_status_info_rows", "(", "folder_id", ",", "account_id", ",", "db_session", ")", ":", "if", "(", "not", "db_session", ".", "query", "(", "ImapFolderSyncStatus", ")", ".", "filter_by", "(", "account_id", "=", "account_id", ",", "folder_id", "=", "folder_id", ")", ".", "all", "(", ")", ")", ":", "db_session", ".", "add", "(", "ImapFolderSyncStatus", "(", "account_id", "=", "account_id", ",", "folder_id", "=", "folder_id", ",", "state", "=", "'initial'", ")", ")", "if", "(", "not", "db_session", ".", "query", "(", "ImapFolderInfo", ")", ".", "filter_by", "(", "account_id", "=", "account_id", ",", "folder_id", "=", "folder_id", ")", ".", "all", "(", ")", ")", ":", "db_session", ".", "add", "(", "ImapFolderInfo", "(", "account_id", "=", "account_id", ",", "folder_id", "=", "folder_id", ",", "uidvalidity", "=", "1", ",", "highestmodseq", "=", "22", ")", ")" ]
add placeholder imapfoldersyncstatus and imapfolderinfo rows for this folder_id if none exist .
train
false
51,011
def do_get_current_language_bidi(parser, token): args = token.contents.split() if ((len(args) != 3) or (args[1] != 'as')): raise TemplateSyntaxError(("'get_current_language_bidi' requires 'as variable' (got %r)" % args)) return GetCurrentLanguageBidiNode(args[2])
[ "def", "do_get_current_language_bidi", "(", "parser", ",", "token", ")", ":", "args", "=", "token", ".", "contents", ".", "split", "(", ")", "if", "(", "(", "len", "(", "args", ")", "!=", "3", ")", "or", "(", "args", "[", "1", "]", "!=", "'as'", ")", ")", ":", "raise", "TemplateSyntaxError", "(", "(", "\"'get_current_language_bidi' requires 'as variable' (got %r)\"", "%", "args", ")", ")", "return", "GetCurrentLanguageBidiNode", "(", "args", "[", "2", "]", ")" ]
this will store the current language layout in the context .
train
false
51,012
def RunTest(name, handler): test_path = os.path.join(_module_dir, (name + '.py')) try: x = __LoadModule(test_path, name) for test in x.List(): test_object = eval(('x.%s(handler)' % test)) if (test_object.Enabled() and (not test_object.Test())): return False except BaseException as e: print(("Couldn't load test %s: %s" % (name, str(e)))) return True
[ "def", "RunTest", "(", "name", ",", "handler", ")", ":", "test_path", "=", "os", ".", "path", ".", "join", "(", "_module_dir", ",", "(", "name", "+", "'.py'", ")", ")", "try", ":", "x", "=", "__LoadModule", "(", "test_path", ",", "name", ")", "for", "test", "in", "x", ".", "List", "(", ")", ":", "test_object", "=", "eval", "(", "(", "'x.%s(handler)'", "%", "test", ")", ")", "if", "(", "test_object", ".", "Enabled", "(", ")", "and", "(", "not", "test_object", ".", "Test", "(", ")", ")", ")", ":", "return", "False", "except", "BaseException", "as", "e", ":", "print", "(", "(", "\"Couldn't load test %s: %s\"", "%", "(", "name", ",", "str", "(", "e", ")", ")", ")", ")", "return", "True" ]
run a single test .
train
false
51,013
def wait_for_read(socks, timeout=None): return _wait_for_io_events(socks, EVENT_READ, timeout)
[ "def", "wait_for_read", "(", "socks", ",", "timeout", "=", "None", ")", ":", "return", "_wait_for_io_events", "(", "socks", ",", "EVENT_READ", ",", "timeout", ")" ]
waits for reading to be available from a list of sockets or optionally a single socket if passed in .
train
false
51,014
@FileSystem.in_directory(current_directory, 'django', 'brocolis') def test_harvest_with_debug_mode_enabled(): for option in ['-d', '--debug-mode']: (status, out) = run_scenario('leaves', 'enabled', **{option: None}) assert_equals(status, 0, out)
[ "@", "FileSystem", ".", "in_directory", "(", "current_directory", ",", "'django'", ",", "'brocolis'", ")", "def", "test_harvest_with_debug_mode_enabled", "(", ")", ":", "for", "option", "in", "[", "'-d'", ",", "'--debug-mode'", "]", ":", "(", "status", ",", "out", ")", "=", "run_scenario", "(", "'leaves'", ",", "'enabled'", ",", "**", "{", "option", ":", "None", "}", ")", "assert_equals", "(", "status", ",", "0", ",", "out", ")" ]
python manage .
train
false
51,015
def _set_concurrent_future_state(concurr, source): assert source.done() if source.cancelled(): concurr.cancel() if (not concurr.set_running_or_notify_cancel()): return exception = source.exception() if (exception is not None): concurr.set_exception(exception) else: result = source.result() concurr.set_result(result)
[ "def", "_set_concurrent_future_state", "(", "concurr", ",", "source", ")", ":", "assert", "source", ".", "done", "(", ")", "if", "source", ".", "cancelled", "(", ")", ":", "concurr", ".", "cancel", "(", ")", "if", "(", "not", "concurr", ".", "set_running_or_notify_cancel", "(", ")", ")", ":", "return", "exception", "=", "source", ".", "exception", "(", ")", "if", "(", "exception", "is", "not", "None", ")", ":", "concurr", ".", "set_exception", "(", "exception", ")", "else", ":", "result", "=", "source", ".", "result", "(", ")", "concurr", ".", "set_result", "(", "result", ")" ]
copy state from a future to a concurrent .
train
false
51,016
def create_trigger_type_db(trigger_type): trigger_type_api = TriggerTypeAPI(**trigger_type) trigger_type_api.validate() ref = ResourceReference.to_string_reference(name=trigger_type_api.name, pack=trigger_type_api.pack) trigger_type_db = get_trigger_type_db(ref) if (not trigger_type_db): trigger_type_db = TriggerTypeAPI.to_model(trigger_type_api) LOG.debug('verified trigger and formulated TriggerDB=%s', trigger_type_db) trigger_type_db = TriggerType.add_or_update(trigger_type_db) return trigger_type_db
[ "def", "create_trigger_type_db", "(", "trigger_type", ")", ":", "trigger_type_api", "=", "TriggerTypeAPI", "(", "**", "trigger_type", ")", "trigger_type_api", ".", "validate", "(", ")", "ref", "=", "ResourceReference", ".", "to_string_reference", "(", "name", "=", "trigger_type_api", ".", "name", ",", "pack", "=", "trigger_type_api", ".", "pack", ")", "trigger_type_db", "=", "get_trigger_type_db", "(", "ref", ")", "if", "(", "not", "trigger_type_db", ")", ":", "trigger_type_db", "=", "TriggerTypeAPI", ".", "to_model", "(", "trigger_type_api", ")", "LOG", ".", "debug", "(", "'verified trigger and formulated TriggerDB=%s'", ",", "trigger_type_db", ")", "trigger_type_db", "=", "TriggerType", ".", "add_or_update", "(", "trigger_type_db", ")", "return", "trigger_type_db" ]
creates a trigger type db object in the db given trigger_type definition as dict .
train
false
51,018
def urljoin_bytes(*atoms): url = ntob('/').join([x for x in atoms if x]) while (ntob('//') in url): url = url.replace(ntob('//'), ntob('/')) return (url or ntob('/'))
[ "def", "urljoin_bytes", "(", "*", "atoms", ")", ":", "url", "=", "ntob", "(", "'/'", ")", ".", "join", "(", "[", "x", "for", "x", "in", "atoms", "if", "x", "]", ")", "while", "(", "ntob", "(", "'//'", ")", "in", "url", ")", ":", "url", "=", "url", ".", "replace", "(", "ntob", "(", "'//'", ")", ",", "ntob", "(", "'/'", ")", ")", "return", "(", "url", "or", "ntob", "(", "'/'", ")", ")" ]
return the given path *atoms .
train
false
51,020
def is_linux(): return sys.platform.startswith(u'linux')
[ "def", "is_linux", "(", ")", ":", "return", "sys", ".", "platform", ".", "startswith", "(", "u'linux'", ")" ]
is this a linux machine? .
train
false
51,022
def pportInAcknowledge(): if (port.DlPortReadPortUchar(statusRegAdrs) & 64): return 1 else: return 0
[ "def", "pportInAcknowledge", "(", ")", ":", "if", "(", "port", ".", "DlPortReadPortUchar", "(", "statusRegAdrs", ")", "&", "64", ")", ":", "return", "1", "else", ":", "return", "0" ]
input from acknowledge pin .
train
false
51,023
def pointbiserialr(x, y): (rpb, prob) = pearsonr(x, y) return PointbiserialrResult(rpb, prob)
[ "def", "pointbiserialr", "(", "x", ",", "y", ")", ":", "(", "rpb", ",", "prob", ")", "=", "pearsonr", "(", "x", ",", "y", ")", "return", "PointbiserialrResult", "(", "rpb", ",", "prob", ")" ]
calculates a point biserial correlation coefficient and its p-value .
train
false
51,024
def floating_ip_get_by_fixed_ip_id(context, fixed_ip_id): return IMPL.floating_ip_get_by_fixed_ip_id(context, fixed_ip_id)
[ "def", "floating_ip_get_by_fixed_ip_id", "(", "context", ",", "fixed_ip_id", ")", ":", "return", "IMPL", ".", "floating_ip_get_by_fixed_ip_id", "(", "context", ",", "fixed_ip_id", ")" ]
get a floating ips by fixed address .
train
false
51,025
def DisableInterfaces(interface): set_tested_versions = ['vista', '2008'] set_args = ['/c', 'netsh', 'set', 'interface', interface, 'DISABLED'] host_version = platform.platform().lower() for version in set_tested_versions: if (host_version.find(version) != (-1)): res = client_utils_common.Execute('cmd', set_args, time_limit=(-1), bypass_whitelist=True) return res return ('', 'Command not available for this version.', 99, '')
[ "def", "DisableInterfaces", "(", "interface", ")", ":", "set_tested_versions", "=", "[", "'vista'", ",", "'2008'", "]", "set_args", "=", "[", "'/c'", ",", "'netsh'", ",", "'set'", ",", "'interface'", ",", "interface", ",", "'DISABLED'", "]", "host_version", "=", "platform", ".", "platform", "(", ")", ".", "lower", "(", ")", "for", "version", "in", "set_tested_versions", ":", "if", "(", "host_version", ".", "find", "(", "version", ")", "!=", "(", "-", "1", ")", ")", ":", "res", "=", "client_utils_common", ".", "Execute", "(", "'cmd'", ",", "set_args", ",", "time_limit", "=", "(", "-", "1", ")", ",", "bypass_whitelist", "=", "True", ")", "return", "res", "return", "(", "''", ",", "'Command not available for this version.'", ",", "99", ",", "''", ")" ]
tries to disable an interface .
train
true
51,026
def versionToUsefulObject(version): from incremental import Version return Version(*[x.value for x in version.asList()[1:] if x])
[ "def", "versionToUsefulObject", "(", "version", ")", ":", "from", "incremental", "import", "Version", "return", "Version", "(", "*", "[", "x", ".", "value", "for", "x", "in", "version", ".", "asList", "(", ")", "[", "1", ":", "]", "if", "x", "]", ")" ]
change an ast c{version()} to a real one .
train
false
51,027
def _get_pkg_license(pkg): licenses = set() cpr = '/usr/share/doc/{0}/copyright'.format(pkg) if os.path.exists(cpr): with salt.utils.fopen(cpr) as fp_: for line in fp_.read().split(os.linesep): if line.startswith('License:'): licenses.add(line.split(':', 1)[1].strip()) return ', '.join(sorted(licenses))
[ "def", "_get_pkg_license", "(", "pkg", ")", ":", "licenses", "=", "set", "(", ")", "cpr", "=", "'/usr/share/doc/{0}/copyright'", ".", "format", "(", "pkg", ")", "if", "os", ".", "path", ".", "exists", "(", "cpr", ")", ":", "with", "salt", ".", "utils", ".", "fopen", "(", "cpr", ")", "as", "fp_", ":", "for", "line", "in", "fp_", ".", "read", "(", ")", ".", "split", "(", "os", ".", "linesep", ")", ":", "if", "line", ".", "startswith", "(", "'License:'", ")", ":", "licenses", ".", "add", "(", "line", ".", "split", "(", "':'", ",", "1", ")", "[", "1", "]", ".", "strip", "(", ")", ")", "return", "', '", ".", "join", "(", "sorted", "(", "licenses", ")", ")" ]
try to get a license from the package .
train
true
51,028
def api_payload_to_create_params(payload): required_parameters = ['collection_type', 'element_identifiers'] missing_parameters = [p for p in required_parameters if (p not in payload)] if missing_parameters: message = ('Missing required parameters %s' % missing_parameters) raise exceptions.ObjectAttributeMissingException(message) params = dict(collection_type=payload.get('collection_type'), element_identifiers=payload.get('element_identifiers'), name=payload.get('name', None)) return params
[ "def", "api_payload_to_create_params", "(", "payload", ")", ":", "required_parameters", "=", "[", "'collection_type'", ",", "'element_identifiers'", "]", "missing_parameters", "=", "[", "p", "for", "p", "in", "required_parameters", "if", "(", "p", "not", "in", "payload", ")", "]", "if", "missing_parameters", ":", "message", "=", "(", "'Missing required parameters %s'", "%", "missing_parameters", ")", "raise", "exceptions", ".", "ObjectAttributeMissingException", "(", "message", ")", "params", "=", "dict", "(", "collection_type", "=", "payload", ".", "get", "(", "'collection_type'", ")", ",", "element_identifiers", "=", "payload", ".", "get", "(", "'element_identifiers'", ")", ",", "name", "=", "payload", ".", "get", "(", "'name'", ",", "None", ")", ")", "return", "params" ]
cleanup api payload to pass into dataset_collections .
train
false
51,029
def test_lazy_process(): t = tempfile.NamedTemporaryFile() os.remove(t.name) lazy_process = LazyProcess(['bash', '-c', ('touch %s; sleep 100' % t.name)]) assert (not os.path.exists(t.name)) lazy_process.start_process() while (not os.path.exists(t.name)): time.sleep(0.01) assert (lazy_process.process.poll() is None) lazy_process.shutdown() ret_val = None for i in range(10): ret_val = lazy_process.process.poll() if (ret_val is not None): break time.sleep(0.01) assert (ret_val is not None)
[ "def", "test_lazy_process", "(", ")", ":", "t", "=", "tempfile", ".", "NamedTemporaryFile", "(", ")", "os", ".", "remove", "(", "t", ".", "name", ")", "lazy_process", "=", "LazyProcess", "(", "[", "'bash'", ",", "'-c'", ",", "(", "'touch %s; sleep 100'", "%", "t", ".", "name", ")", "]", ")", "assert", "(", "not", "os", ".", "path", ".", "exists", "(", "t", ".", "name", ")", ")", "lazy_process", ".", "start_process", "(", ")", "while", "(", "not", "os", ".", "path", ".", "exists", "(", "t", ".", "name", ")", ")", ":", "time", ".", "sleep", "(", "0.01", ")", "assert", "(", "lazy_process", ".", "process", ".", "poll", "(", ")", "is", "None", ")", "lazy_process", ".", "shutdown", "(", ")", "ret_val", "=", "None", "for", "i", "in", "range", "(", "10", ")", ":", "ret_val", "=", "lazy_process", ".", "process", ".", "poll", "(", ")", "if", "(", "ret_val", "is", "not", "None", ")", ":", "break", "time", ".", "sleep", "(", "0.01", ")", "assert", "(", "ret_val", "is", "not", "None", ")" ]
create process .
train
false